Update README.md
Browse files
README.md
CHANGED
|
@@ -7,7 +7,9 @@ My model is a state-of-the-art language processing AI designed to understand and
|
|
| 7 |
How to use?
|
| 8 |
|
| 9 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
|
|
| 10 |
from transformers.generation import GenerationConfig
|
|
|
|
| 11 |
import torch
|
| 12 |
|
| 13 |
|
|
@@ -15,12 +17,17 @@ model = AutoModelForCausalLM.from_pretrained(
|
|
| 15 |
'TwT-6/open_llm_leaderboard_demo',
|
| 16 |
attn_implementation="flash_attention_2",
|
| 17 |
trust_remote_code=True, torch_dtype=torch.bfloat16, device_map="auto").eval()
|
|
|
|
| 18 |
tokenizer = AutoTokenizer.from_pretrained('TwT-6/open_llm_leaderboard_demo', trust_remote_code=True)
|
|
|
|
| 19 |
inputs = '你好'
|
|
|
|
| 20 |
inputs = f'<|omni_start|>### User:\n{inputs}\n\n### Assistant:\n'
|
|
|
|
| 21 |
inputs = tokenizer(inputs, return_tensors="pt").to('cuda')
|
| 22 |
|
| 23 |
output_ids = model.generate(**inputs)[0].cpu()
|
|
|
|
| 24 |
output = tokenizer.decode(output_ids[inputs.input_ids.shape[-1]:])
|
| 25 |
|
| 26 |
print(output)
|
|
|
|
| 7 |
How to use?
|
| 8 |
|
| 9 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
| 10 |
+
|
| 11 |
from transformers.generation import GenerationConfig
|
| 12 |
+
|
| 13 |
import torch
|
| 14 |
|
| 15 |
|
|
|
|
| 17 |
'TwT-6/open_llm_leaderboard_demo',
|
| 18 |
attn_implementation="flash_attention_2",
|
| 19 |
trust_remote_code=True, torch_dtype=torch.bfloat16, device_map="auto").eval()
|
| 20 |
+
|
| 21 |
tokenizer = AutoTokenizer.from_pretrained('TwT-6/open_llm_leaderboard_demo', trust_remote_code=True)
|
| 22 |
+
|
| 23 |
inputs = '你好'
|
| 24 |
+
|
| 25 |
inputs = f'<|omni_start|>### User:\n{inputs}\n\n### Assistant:\n'
|
| 26 |
+
|
| 27 |
inputs = tokenizer(inputs, return_tensors="pt").to('cuda')
|
| 28 |
|
| 29 |
output_ids = model.generate(**inputs)[0].cpu()
|
| 30 |
+
|
| 31 |
output = tokenizer.decode(output_ids[inputs.input_ids.shape[-1]:])
|
| 32 |
|
| 33 |
print(output)
|