Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -77,7 +77,7 @@ def generate(
|
|
| 77 |
input_ids = input_ids[:, -MAX_INPUT_TOKEN_LENGTH:]
|
| 78 |
gr.Warning(f"Trimmed input from conversation as it was longer than {MAX_INPUT_TOKEN_LENGTH} tokens.")
|
| 79 |
|
| 80 |
-
streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=
|
| 81 |
generate_kwargs = dict(
|
| 82 |
{"input_ids": input_ids},
|
| 83 |
streamer=streamer,
|
|
@@ -91,6 +91,7 @@ def generate(
|
|
| 91 |
)
|
| 92 |
t = Thread(target=model.generate, kwargs=generate_kwargs)
|
| 93 |
t.start()
|
|
|
|
| 94 |
|
| 95 |
outputs = []
|
| 96 |
for text in streamer:
|
|
|
|
| 77 |
input_ids = input_ids[:, -MAX_INPUT_TOKEN_LENGTH:]
|
| 78 |
gr.Warning(f"Trimmed input from conversation as it was longer than {MAX_INPUT_TOKEN_LENGTH} tokens.")
|
| 79 |
|
| 80 |
+
streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=False, skip_special_tokens=True)
|
| 81 |
generate_kwargs = dict(
|
| 82 |
{"input_ids": input_ids},
|
| 83 |
streamer=streamer,
|
|
|
|
| 91 |
)
|
| 92 |
t = Thread(target=model.generate, kwargs=generate_kwargs)
|
| 93 |
t.start()
|
| 94 |
+
t.join()
|
| 95 |
|
| 96 |
outputs = []
|
| 97 |
for text in streamer:
|