mirror of
https://www.modelscope.cn/Qwen/QwQ-32B.git
synced 2025-12-08 21:34:21 +08:00
update README
This commit is contained in:
parent
debcd63c0c
commit
887ddbd72b
@ -100,11 +100,13 @@ To achieve optimal performance, we recommend the following settings:
|
|||||||
- Use Temperature=0.6 and TopP=0.95 instead of Greedy decoding to avoid endless repetitions.
|
- Use Temperature=0.6 and TopP=0.95 instead of Greedy decoding to avoid endless repetitions.
|
||||||
- Use TopK between 20 and 40 to filter out rare token occurrences while maintaining the diversity of the generated output.
|
- Use TopK between 20 and 40 to filter out rare token occurrences while maintaining the diversity of the generated output.
|
||||||
|
|
||||||
3. **Standardize Output Format**: We recommend using prompts to standardize model outputs when benchmarking.
|
3. **No Thinking Content in History**: In multi-turn conversations, the historical model output should only include the final output part and does not need to include the thinking content. This feature is already implemented in `apply_chat_template`.
|
||||||
|
|
||||||
|
4. **Standardize Output Format**: We recommend using prompts to standardize model outputs when benchmarking.
|
||||||
- **Math Problems**: Include "Please reason step by step, and put your final answer within \boxed{}." in the prompt.
|
- **Math Problems**: Include "Please reason step by step, and put your final answer within \boxed{}." in the prompt.
|
||||||
- **Multiple-Choice Questions**: Add the following JSON structure to the prompt to standardize responses: "Please show your choice in the `answer` field with only the choice letter, e.g.,`\"answer\": \"C\"`." in the prompt.
|
- **Multiple-Choice Questions**: Add the following JSON structure to the prompt to standardize responses: "Please show your choice in the `answer` field with only the choice letter, e.g.,`\"answer\": \"C\"`." in the prompt.
|
||||||
|
|
||||||
4. **Handle Long Inputs**: For inputs exceeding 32,768 tokens, enable [YaRN](https://arxiv.org/abs/2309.00071) to improve the model's ability to capture long-sequence information effectively.
|
5. **Handle Long Inputs**: For inputs exceeding 32,768 tokens, enable [YaRN](https://arxiv.org/abs/2309.00071) to improve the model's ability to capture long-sequence information effectively.
|
||||||
|
|
||||||
For supported frameworks, you could add the following to `config.json` to enable YaRN:
|
For supported frameworks, you could add the following to `config.json` to enable YaRN:
|
||||||
```json
|
```json
|
||||||
|
|||||||
Loading…
x
Reference in New Issue
Block a user