benjamin commited on
Commit
f037ba0
·
verified ·
1 Parent(s): 2761bdc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -3
README.md CHANGED
@@ -30,16 +30,21 @@ Llama3-2-3B-IT-Byte performs competitively although it has been trained only on
30
  from transformers import AutoTokenizer, AutoModelForCausalLM
31
 
32
  tokenizer = AutoTokenizer.from_pretrained("benjamin/Llama3-2-3B-IT-Byte")
33
- print("Vocab Size:", len(tokenizer)) # 256 bytes + some special tokens
34
 
35
  device = "cuda"
36
- model = AutoModelForCausalLM.from_pretrained("benjamin/Llama3-2-3B-IT-Byte", trust_remote_code=True)
 
 
37
  model = model.to(device)
38
 
39
- tokens = tokenizer.apply_chat_template([{"role": "user", "content": "Hello, how are you doing?"}], return_tensors="pt")
 
 
40
  eot_id = tokenizer.convert_tokens_to_ids("<|eot_id|>")
41
  out = model.generate(tokens.to(model.device), eos_token_id=eot_id)
42
  print(tokenizer.decode(out[0]))
 
43
  ```
44
 
45
  ## Training
 
30
  from transformers import AutoTokenizer, AutoModelForCausalLM
31
 
32
  tokenizer = AutoTokenizer.from_pretrained("benjamin/Llama3-2-3B-IT-Byte")
33
+ print("Vocab Size:", len(tokenizer)) # 256 bytes + some special tokens
34
 
35
  device = "cuda"
36
+ model = AutoModelForCausalLM.from_pretrained(
37
+ "benjamin/Llama3-2-3B-IT-Byte", trust_remote_code=True
38
+ )
39
  model = model.to(device)
40
 
41
+ tokens = tokenizer.apply_chat_template(
42
+ [{"role": "user", "content": "Hello, how are you doing?"}], return_tensors="pt"
43
+ )
44
  eot_id = tokenizer.convert_tokens_to_ids("<|eot_id|>")
45
  out = model.generate(tokens.to(model.device), eos_token_id=eot_id)
46
  print(tokenizer.decode(out[0]))
47
+
48
  ```
49
 
50
  ## Training