Update README.md
Browse files
README.md
CHANGED
|
@@ -121,7 +121,7 @@ TeleChat模型相比同规模模型在评测效果方面也有较好的表现,
|
|
| 121 |
>>> os.environ["CUDA_VISIBLE_DEVICES"] = '0'
|
| 122 |
>>> tokenizer = AutoTokenizer.from_pretrained('../models/1B')
|
| 123 |
>>> model = AutoModelForCausalLM.from_pretrained('../models/1B', trust_remote_code=True, device_map="auto", torch_dtype=torch.float16)
|
| 124 |
-
>>> generate_config = GenerationConfig.from_pretrained('../models/
|
| 125 |
>>> question="生抽与老抽的区别?"
|
| 126 |
>>> answer, history = model.chat(tokenizer = tokenizer, question=question, history=[], generation_config=generate_config, stream=False)
|
| 127 |
>>> print(answer)
|
|
|
|
| 121 |
>>> os.environ["CUDA_VISIBLE_DEVICES"] = '0'
|
| 122 |
>>> tokenizer = AutoTokenizer.from_pretrained('../models/1B')
|
| 123 |
>>> model = AutoModelForCausalLM.from_pretrained('../models/1B', trust_remote_code=True, device_map="auto", torch_dtype=torch.float16)
|
| 124 |
+
>>> generate_config = GenerationConfig.from_pretrained('../models/1B')
|
| 125 |
>>> question="生抽与老抽的区别?"
|
| 126 |
>>> answer, history = model.chat(tokenizer = tokenizer, question=question, history=[], generation_config=generate_config, stream=False)
|
| 127 |
>>> print(answer)
|