am commited on
Commit
c39b808
·
1 Parent(s): aa123d5
Files changed (1) hide show
  1. app.py +16 -15
app.py CHANGED
@@ -117,21 +117,22 @@ def model_inference(
117
  # Generate
118
  streamer = TextIteratorStreamer(processor, skip_prompt=True, skip_special_tokens=True)
119
  generation_args = dict(inputs, streamer=streamer, max_new_tokens=2048)
120
- generated_text = ""
121
-
122
- thread = Thread(target=model.generate, kwargs=generation_args)
123
- thread.start()
124
-
125
- yield "..."
126
- buffer = ""
127
-
128
-
129
- for new_text in streamer:
130
- buffer += new_text
131
- # generated_text_without_prompt = buffer#[len(ext_buffer):]
132
- # time.sleep(0.01)
133
- # print(f"buffer: {buffer}")
134
- yield buffer
 
135
 
136
 
137
  examples=[
 
117
  # Generate
118
  streamer = TextIteratorStreamer(processor, skip_prompt=True, skip_special_tokens=True)
119
  generation_args = dict(inputs, streamer=streamer, max_new_tokens=2048)
120
+ # generated_text = ""
121
+
122
+ with torch.inference_mode():
123
+ thread = Thread(target=model.generate, kwargs=generation_args)
124
+ thread.start()
125
+
126
+ yield "..."
127
+ buffer = ""
128
+
129
+
130
+ for new_text in streamer:
131
+ buffer += new_text
132
+ # generated_text_without_prompt = buffer#[len(ext_buffer):]
133
+ # time.sleep(0.01)
134
+ # print(f"buffer: {buffer}")
135
+ yield buffer
136
 
137
 
138
  examples=[