Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -679,7 +679,7 @@ Now, write your summary. Do not add preceding text before the abstraction, and d
|
|
| 679 |
SamplingParams(
|
| 680 |
temperature=0.0,
|
| 681 |
top_k=1,
|
| 682 |
-
max_tokens=
|
| 683 |
repetition_penalty=1.2
|
| 684 |
)
|
| 685 |
)
|
|
@@ -696,7 +696,7 @@ Now, write your summary. Do not add preceding text before the abstraction, and d
|
|
| 696 |
with torch.no_grad():
|
| 697 |
outputs = state.llm_model.generate(
|
| 698 |
input_ids,
|
| 699 |
-
max_new_tokens=
|
| 700 |
temperature=0.00,
|
| 701 |
do_sample=True,
|
| 702 |
repetition_penalty=1.2
|
|
@@ -774,7 +774,7 @@ def extract_trial_spaces(trial_text: str) -> str:
|
|
| 774 |
SamplingParams(
|
| 775 |
temperature=0.0,
|
| 776 |
top_k=1,
|
| 777 |
-
max_tokens=
|
| 778 |
repetition_penalty=1.3
|
| 779 |
)
|
| 780 |
)
|
|
|
|
| 679 |
SamplingParams(
|
| 680 |
temperature=0.0,
|
| 681 |
top_k=1,
|
| 682 |
+
max_tokens=4000,
|
| 683 |
repetition_penalty=1.2
|
| 684 |
)
|
| 685 |
)
|
|
|
|
| 696 |
with torch.no_grad():
|
| 697 |
outputs = state.llm_model.generate(
|
| 698 |
input_ids,
|
| 699 |
+
max_new_tokens=4000,
|
| 700 |
temperature=0.00,
|
| 701 |
do_sample=True,
|
| 702 |
repetition_penalty=1.2
|
|
|
|
| 774 |
SamplingParams(
|
| 775 |
temperature=0.0,
|
| 776 |
top_k=1,
|
| 777 |
+
max_tokens=4000,
|
| 778 |
repetition_penalty=1.3
|
| 779 |
)
|
| 780 |
)
|