OnAnOrange commited on
Commit
96ddee5
·
verified ·
1 Parent(s): aeef665

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -29,6 +29,9 @@ pip install torch transformers accelerate
29
 
30
  ## Quick Start
31
 
 
 
 
32
  ```python
33
  import math
34
  import copy
@@ -232,7 +235,7 @@ prompts = [
232
  ],
233
  ]
234
 
235
- encoded = [tokenizer.apply_chat_template(m, add_generation_prompt=True, tokenize=True) for m in prompts]
236
  prompt_lens = [len(e) for e in encoded]
237
  max_len = max(prompt_lens)
238
  pad_id = tokenizer.pad_token_id
 
29
 
30
  ## Quick Start
31
 
32
+ > [!NOTE]
33
+ > We recommend setting `enable_thinking=False` when using the model to ensure stable behavior and reproducible results.
34
+
35
  ```python
36
  import math
37
  import copy
 
235
  ],
236
  ]
237
 
238
+ encoded = [tokenizer.apply_chat_template(m, add_generation_prompt=True, tokenize=True, enable_thinking=False) for m in prompts]
239
  prompt_lens = [len(e) for e in encoded]
240
  max_len = max(prompt_lens)
241
  pad_id = tokenizer.pad_token_id