debug
Browse files
data/models/llama3-1-70b.py
CHANGED
|
@@ -38,7 +38,6 @@ def chat_completion():
|
|
| 38 |
generate_kwargs = dict(
|
| 39 |
inputs,
|
| 40 |
top_p=top_p,
|
| 41 |
-
repetition_penalty=penalty,
|
| 42 |
max_new_tokens=max_tokens,
|
| 43 |
do_sample=True,
|
| 44 |
temperature=temperature,
|
|
|
|
| 38 |
generate_kwargs = dict(
|
| 39 |
inputs,
|
| 40 |
top_p=top_p,
|
|
|
|
| 41 |
max_new_tokens=max_tokens,
|
| 42 |
do_sample=True,
|
| 43 |
temperature=temperature,
|