am commited on
Commit ·
c39b808
1
Parent(s): aa123d5
app.py
CHANGED
|
@@ -117,21 +117,22 @@ def model_inference(
|
|
| 117 |
# Generate
|
| 118 |
streamer = TextIteratorStreamer(processor, skip_prompt=True, skip_special_tokens=True)
|
| 119 |
generation_args = dict(inputs, streamer=streamer, max_new_tokens=2048)
|
| 120 |
-
generated_text = ""
|
| 121 |
-
|
| 122 |
-
|
| 123 |
-
|
| 124 |
-
|
| 125 |
-
|
| 126 |
-
|
| 127 |
-
|
| 128 |
-
|
| 129 |
-
|
| 130 |
-
|
| 131 |
-
|
| 132 |
-
|
| 133 |
-
|
| 134 |
-
|
|
|
|
| 135 |
|
| 136 |
|
| 137 |
examples=[
|
|
|
|
| 117 |
# Generate
|
| 118 |
streamer = TextIteratorStreamer(processor, skip_prompt=True, skip_special_tokens=True)
|
| 119 |
generation_args = dict(inputs, streamer=streamer, max_new_tokens=2048)
|
| 120 |
+
# generated_text = ""
|
| 121 |
+
|
| 122 |
+
with torch.inference_mode():
|
| 123 |
+
thread = Thread(target=model.generate, kwargs=generation_args)
|
| 124 |
+
thread.start()
|
| 125 |
+
|
| 126 |
+
yield "..."
|
| 127 |
+
buffer = ""
|
| 128 |
+
|
| 129 |
+
|
| 130 |
+
for new_text in streamer:
|
| 131 |
+
buffer += new_text
|
| 132 |
+
# generated_text_without_prompt = buffer#[len(ext_buffer):]
|
| 133 |
+
# time.sleep(0.01)
|
| 134 |
+
# print(f"buffer: {buffer}")
|
| 135 |
+
yield buffer
|
| 136 |
|
| 137 |
|
| 138 |
examples=[
|