Update README.md
Browse files
README.md
CHANGED
|
@@ -49,7 +49,7 @@ We use a simple template to generate the solution for evalplus:
|
|
| 49 |
|
| 50 |
## Quickstart
|
| 51 |
|
| 52 |
-
Here provides a code snippet with `apply_chat_template` to show you how to load the tokenizer and model and how to generate contents. You should
|
| 53 |
```python
|
| 54 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 55 |
device = "cuda" # the device to load the model onto
|
|
|
|
| 49 |
|
| 50 |
## Quickstart
|
| 51 |
|
| 52 |
+
Here provides a code snippet with `apply_chat_template` to show you how to load the tokenizer and model and how to generate contents. You should upgrade the transformers if you receive an error when loading the tokenizer
|
| 53 |
```python
|
| 54 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 55 |
device = "cuda" # the device to load the model onto
|