Update README.md
Browse files
README.md
CHANGED
|
@@ -50,11 +50,11 @@ device = "cuda"
|
|
| 50 |
bolmo = AutoModelForCausalLM.from_pretrained("allenai/Bolmo-7B", trust_remote_code=True).to(device)
|
| 51 |
tokenizer = AutoTokenizer.from_pretrained("allenai/Bolmo-7B", trust_remote_code=True)
|
| 52 |
|
| 53 |
-
message = ["
|
| 54 |
input_ids = tokenizer(message, return_tensors="pt")["input_ids"].to(device)
|
| 55 |
|
| 56 |
# `max_new_tokens` is the amuont of bytes to generate
|
| 57 |
-
response = bolmo.generate(input_ids, max_new_tokens=
|
| 58 |
print(tokenizer.decode(response[0], skip_special_tokens=True))
|
| 59 |
```
|
| 60 |
|
|
|
|
| 50 |
bolmo = AutoModelForCausalLM.from_pretrained("allenai/Bolmo-7B", trust_remote_code=True).to(device)
|
| 51 |
tokenizer = AutoTokenizer.from_pretrained("allenai/Bolmo-7B", trust_remote_code=True)
|
| 52 |
|
| 53 |
+
message = ["Language modeling is "]
|
| 54 |
input_ids = tokenizer(message, return_tensors="pt")["input_ids"].to(device)
|
| 55 |
|
| 56 |
# `max_new_tokens` is the amuont of bytes to generate
|
| 57 |
+
response = bolmo.generate(input_ids, max_new_tokens=256, do_sample=True, temperature=0.1)
|
| 58 |
print(tokenizer.decode(response[0], skip_special_tokens=True))
|
| 59 |
```
|
| 60 |
|