Bochkov commited on
Commit
fcf9573
·
verified ·
1 Parent(s): a3bb9c4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -1
README.md CHANGED
@@ -97,7 +97,7 @@ import torch
97
  from transformers import AutoTokenizer, AutoModelForCausalLM
98
 
99
  tokenizer = AutoTokenizer.from_pretrained("Bochkov/emergent-semantics-model-1024-float-335m")
100
- model = AutoModelForCausalLM.from_pretrained("Bochkov/emergent-semantics-model-1024-float-335m", trust_remote_code=True)
101
 
102
  inputs = torch.tensor([tokenizer.encode("Question: What is the capital of Japan?\nAnswer:")], dtype=torch.long, device='cuda')
103
 
@@ -108,6 +108,9 @@ outputs = model.generate(
108
  )
109
  print(tokenizer.decode(outputs[0].tolist()))
110
 
 
 
 
111
  ```
112
 
113
  ---
 
97
  from transformers import AutoTokenizer, AutoModelForCausalLM
98
 
99
  tokenizer = AutoTokenizer.from_pretrained("Bochkov/emergent-semantics-model-1024-float-335m")
100
+ model = AutoModelForCausalLM.from_pretrained("Bochkov/emergent-semantics-model-1024-float-335m", trust_remote_code=True).to('cuda')
101
 
102
  inputs = torch.tensor([tokenizer.encode("Question: What is the capital of Japan?\nAnswer:")], dtype=torch.long, device='cuda')
103
 
 
108
  )
109
  print(tokenizer.decode(outputs[0].tolist()))
110
 
111
+ #Question: What is the capital of Japan?
112
+ #Answer:Tokyo Metropolitan
113
+
114
  ```
115
 
116
  ---