TwT-6 commited on
Commit
88679c6
·
verified ·
1 Parent(s): 35cd376

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -0
README.md CHANGED
@@ -7,7 +7,9 @@ My model is a state-of-the-art language processing AI designed to understand and
7
  How to use?
8
 
9
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
10
  from transformers.generation import GenerationConfig
 
11
  import torch
12
 
13
 
@@ -15,12 +17,17 @@ model = AutoModelForCausalLM.from_pretrained(
15
  'TwT-6/open_llm_leaderboard_demo',
16
  attn_implementation="flash_attention_2",
17
  trust_remote_code=True, torch_dtype=torch.bfloat16, device_map="auto").eval()
 
18
  tokenizer = AutoTokenizer.from_pretrained('TwT-6/open_llm_leaderboard_demo', trust_remote_code=True)
 
19
  inputs = '你好'
 
20
  inputs = f'<|omni_start|>### User:\n{inputs}\n\n### Assistant:\n'
 
21
  inputs = tokenizer(inputs, return_tensors="pt").to('cuda')
22
 
23
  output_ids = model.generate(**inputs)[0].cpu()
 
24
  output = tokenizer.decode(output_ids[inputs.input_ids.shape[-1]:])
25
 
26
  print(output)
 
7
  How to use?
8
 
9
  from transformers import AutoModelForCausalLM, AutoTokenizer
10
+
11
  from transformers.generation import GenerationConfig
12
+
13
  import torch
14
 
15
 
 
17
  'TwT-6/open_llm_leaderboard_demo',
18
  attn_implementation="flash_attention_2",
19
  trust_remote_code=True, torch_dtype=torch.bfloat16, device_map="auto").eval()
20
+
21
  tokenizer = AutoTokenizer.from_pretrained('TwT-6/open_llm_leaderboard_demo', trust_remote_code=True)
22
+
23
  inputs = '你好'
24
+
25
  inputs = f'<|omni_start|>### User:\n{inputs}\n\n### Assistant:\n'
26
+
27
  inputs = tokenizer(inputs, return_tensors="pt").to('cuda')
28
 
29
  output_ids = model.generate(**inputs)[0].cpu()
30
+
31
  output = tokenizer.decode(output_ids[inputs.input_ids.shape[-1]:])
32
 
33
  print(output)