Fix model path in the transformers example code

#1
by phh - opened
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -32,8 +32,8 @@ Bonsai can be easily used through the Huggingface Transformers library. However,
32
  ```{python}
33
  from transformers import AutoTokenizer, AutoModelForCausalLM
34
 
35
- tokenizer = AutoTokenizer.from_pretrained("hespere-ai/Bonsai", trust_remote_code=True)
36
- model = AutoModelForCausalLM.from_pretrained("hespere-ai/Bonsai", trust_remote_code=True)
37
  text = "What is the capital of France?"
38
  inputs = tokenizer(text, return_tensors="pt")
39
  outputs = model.generate(**inputs, max_length=100)
 
32
  ```{python}
33
  from transformers import AutoTokenizer, AutoModelForCausalLM
34
 
35
+ tokenizer = AutoTokenizer.from_pretrained("deepgrove/Bonsai", trust_remote_code=True)
36
+ model = AutoModelForCausalLM.from_pretrained("deepgrove/Bonsai", trust_remote_code=True)
37
  text = "What is the capital of France?"
38
  inputs = tokenizer(text, return_tensors="pt")
39
  outputs = model.generate(**inputs, max_length=100)