Update README.md
Browse files
README.md
CHANGED
|
@@ -55,9 +55,8 @@ library_name: transformers
|
|
| 55 |
|
| 56 |
## Usage
|
| 57 |
|
| 58 |
-
|
| 59 |
-
|
| 60 |
-
Inference is better with `temperature=0.8`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set the maximum number of tokens to 32k to allow for longer chain-of-thought (CoT).
|
| 61 |
|
| 62 |
### Input Formats
|
| 63 |
|
|
|
|
| 55 |
|
| 56 |
## Usage
|
| 57 |
|
| 58 |
+
> [!IMPORTANT]
|
| 59 |
+
> To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
|
|
|
|
| 60 |
|
| 61 |
### Input Formats
|
| 62 |
|