Update README.md
Browse files
README.md
CHANGED
|
@@ -29,6 +29,9 @@ pip install torch transformers accelerate
|
|
| 29 |
|
| 30 |
## Quick Start
|
| 31 |
|
|
|
|
|
|
|
|
|
|
| 32 |
```python
|
| 33 |
import math
|
| 34 |
import copy
|
|
@@ -232,7 +235,7 @@ prompts = [
|
|
| 232 |
],
|
| 233 |
]
|
| 234 |
|
| 235 |
-
encoded = [tokenizer.apply_chat_template(m, add_generation_prompt=True, tokenize=True) for m in prompts]
|
| 236 |
prompt_lens = [len(e) for e in encoded]
|
| 237 |
max_len = max(prompt_lens)
|
| 238 |
pad_id = tokenizer.pad_token_id
|
|
|
|
| 29 |
|
| 30 |
## Quick Start
|
| 31 |
|
| 32 |
+
> [!NOTE]
|
| 33 |
+
> We recommend setting `enable_thinking=False` when using the model to ensure stable behavior and reproducible results.
|
| 34 |
+
|
| 35 |
```python
|
| 36 |
import math
|
| 37 |
import copy
|
|
|
|
| 235 |
],
|
| 236 |
]
|
| 237 |
|
| 238 |
+
encoded = [tokenizer.apply_chat_template(m, add_generation_prompt=True, tokenize=True, enable_thinking=False) for m in prompts]
|
| 239 |
prompt_lens = [len(e) for e in encoded]
|
| 240 |
max_len = max(prompt_lens)
|
| 241 |
pad_id = tokenizer.pad_token_id
|