use repo model
#6
by
Realisticly
- opened
README.md
CHANGED
|
@@ -65,7 +65,7 @@ You can use transformers to run inference with Yi-Coder models (both chat and ba
|
|
| 65 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 66 |
|
| 67 |
device = "cuda" # the device to load the model onto
|
| 68 |
-
model_path = "01-ai/Yi-Coder-
|
| 69 |
|
| 70 |
tokenizer = AutoTokenizer.from_pretrained(model_path)
|
| 71 |
model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto").eval()
|
|
|
|
| 65 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 66 |
|
| 67 |
device = "cuda" # the device to load the model onto
|
| 68 |
+
model_path = "01-ai/Yi-Coder-1.5B-Chat"
|
| 69 |
|
| 70 |
tokenizer = AutoTokenizer.from_pretrained(model_path)
|
| 71 |
model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto").eval()
|