omitakahiro commited on
Commit
dca4e8a
·
verified ·
1 Parent(s): 6330e4a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -4
README.md CHANGED
@@ -27,15 +27,19 @@ import torch
27
  from transformers import AutoModelForCausalLM, AutoTokenizer
28
 
29
  tokenizer = AutoTokenizer.from_pretrained("stockmark/Stockmark-2-100B-Instruct-beta")
30
- model = AutoModelForCausalLM.from_pretrained("stockmark/Stockmark-2-100B-Instruct-beta", device_map="auto", torch_dtype=torch.bfloat16)
 
 
31
 
32
- instruction = ""
33
- input_ids = tokenizer(instruction, add_generation_prompt=True).input_ids.to(model.device)
 
 
34
 
35
  with torch.inference_mode():
36
  tokens = model.generate(
37
  input_ids,
38
- max_new_tokens = 1024,
39
  do_sample = True,
40
  temperature = 0.7,
41
  top_p = 0.95,
 
27
  from transformers import AutoModelForCausalLM, AutoTokenizer
28
 
29
  tokenizer = AutoTokenizer.from_pretrained("stockmark/Stockmark-2-100B-Instruct-beta")
30
+ model = AutoModelForCausalLM.from_pretrained(
31
+ "stockmark/Stockmark-2-100B-Instruct-beta", device_map="auto", torch_dtype=torch.bfloat16
32
+ )
33
 
34
+ instruction = "自然言語処理とは?"
35
+ input_ids = tokenizer.apply_chat_template(
36
+ [{"role": "user", "content": instruction}], add_generation_prompt=True, return_tensors="pt"
37
+ ).to(model.device)
38
 
39
  with torch.inference_mode():
40
  tokens = model.generate(
41
  input_ids,
42
+ max_new_tokens = 256,
43
  do_sample = True,
44
  temperature = 0.7,
45
  top_p = 0.95,