Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -65,7 +65,7 @@ You can use transformers to run inference with Yi-Coder models (both chat and ba
65
  from transformers import AutoTokenizer, AutoModelForCausalLM
66
 
67
  device = "cuda" # the device to load the model onto
68
- model_path = "01-ai/Yi-Coder-9B-Chat"
69
 
70
  tokenizer = AutoTokenizer.from_pretrained(model_path)
71
  model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto").eval()
 
65
  from transformers import AutoTokenizer, AutoModelForCausalLM
66
 
67
  device = "cuda" # the device to load the model onto
68
+ model_path = "01-ai/Yi-Coder-1.5B-Chat"
69
 
70
  tokenizer = AutoTokenizer.from_pretrained(model_path)
71
  model = AutoModelForCausalLM.from_pretrained(model_path, device_map="auto").eval()