Update README.md
Browse files
README.md
CHANGED
|
@@ -13,7 +13,7 @@ widget:
|
|
| 13 |
inference:
|
| 14 |
parameters:
|
| 15 |
max_new_tokens: 100
|
| 16 |
-
do_sample:
|
| 17 |
pipeline_tag: text2text-generation
|
| 18 |
---
|
| 19 |
# Gemma-2B Fine-Tuned Python Model
|
|
@@ -65,7 +65,7 @@ inputs = encodeds.to(device)
|
|
| 65 |
|
| 66 |
|
| 67 |
# Increase max_new_tokens if needed
|
| 68 |
-
generated_ids = model.generate(inputs, max_new_tokens=1000, do_sample=
|
| 69 |
ans = ''
|
| 70 |
for i in tokenizer.decode(generated_ids[0], skip_special_tokens=True).split('<end_of_turn>')[:2]:
|
| 71 |
ans += i
|
|
|
|
| 13 |
inference:
|
| 14 |
parameters:
|
| 15 |
max_new_tokens: 100
|
| 16 |
+
do_sample: True
|
| 17 |
pipeline_tag: text2text-generation
|
| 18 |
---
|
| 19 |
# Gemma-2B Fine-Tuned Python Model
|
|
|
|
| 65 |
|
| 66 |
|
| 67 |
# Increase max_new_tokens if needed
|
| 68 |
+
generated_ids = model.generate(inputs, max_new_tokens=1000, do_sample=True, pad_token_id=tokenizer.eos_token_id)
|
| 69 |
ans = ''
|
| 70 |
for i in tokenizer.decode(generated_ids[0], skip_special_tokens=True).split('<end_of_turn>')[:2]:
|
| 71 |
ans += i
|