reduced top_p to 0.8
Browse files
app.py
CHANGED
|
@@ -45,7 +45,7 @@ def llama_generation(input_text: str,
|
|
| 45 |
output_ids = llama_model.generate(input_ids=input_ids,
|
| 46 |
max_new_tokens=256,
|
| 47 |
temperature=0.5,
|
| 48 |
-
top_p=0.
|
| 49 |
repetition_penalty=2.0)
|
| 50 |
|
| 51 |
# Decode
|
|
|
|
| 45 |
output_ids = llama_model.generate(input_ids=input_ids,
|
| 46 |
max_new_tokens=256,
|
| 47 |
temperature=0.5,
|
| 48 |
+
top_p=0.8,
|
| 49 |
repetition_penalty=2.0)
|
| 50 |
|
| 51 |
# Decode
|