Spaces:
Build error
Build error
additional chatbot fixes
Browse files
app.py
CHANGED
|
@@ -111,7 +111,6 @@ def chat(
|
|
| 111 |
max_new_tokens=10,
|
| 112 |
temperature=0.1,
|
| 113 |
top_p=1.0,
|
| 114 |
-
stop="<|endoftext|>",
|
| 115 |
seed=42,
|
| 116 |
):
|
| 117 |
global model
|
|
@@ -148,7 +147,7 @@ FRITZ: I am an RNN based Large Language Model (LLM) that use no attention layers
|
|
| 148 |
max_new_tokens = int(max_new_tokens)
|
| 149 |
temperature = float(temperature)
|
| 150 |
top_p = float(top_p)
|
| 151 |
-
stop = [x.strip(' ') for x in stop.split(',')]
|
| 152 |
seed = seed
|
| 153 |
|
| 154 |
assert 1 <= max_new_tokens <= 384
|
|
@@ -165,16 +164,17 @@ FRITZ: I am an RNN based Large Language Model (LLM) that use no attention layers
|
|
| 165 |
|
| 166 |
model.loadContext(newctx=intro+prompt)
|
| 167 |
|
| 168 |
-
generated_text = model.forward(number=max_new_tokens, stopStrings=
|
| 169 |
|
| 170 |
generated_text = generated_text.lstrip("\n ")
|
|
|
|
| 171 |
print(f"{generated_text}")
|
| 172 |
-
|
| 173 |
for stop_word in stop:
|
| 174 |
stop_word = codecs.getdecoder("unicode_escape")(stop_word)[0]
|
| 175 |
if stop_word != '' and stop_word in generated_text:
|
| 176 |
generated_text = generated_text[:generated_text.find(stop_word)]
|
| 177 |
-
|
| 178 |
gc.collect()
|
| 179 |
history.append((prompt, generated_text))
|
| 180 |
return history,history
|
|
@@ -230,7 +230,7 @@ iface = gr.Interface(
|
|
| 230 |
gr.Slider(0.0, 1.0, value=0.85), # top_p
|
| 231 |
gr.Textbox(lines=1, value="<|endoftext|>") # stop
|
| 232 |
],
|
| 233 |
-
outputs=gr.Textbox(lines=25),
|
| 234 |
examples=examples,
|
| 235 |
cache_examples=False,
|
| 236 |
).queue()
|
|
@@ -246,10 +246,9 @@ chatiface = gr.Interface(
|
|
| 246 |
"state",
|
| 247 |
gr.Slider(1, 256, value=60), # max_tokens
|
| 248 |
gr.Slider(0.0, 1.0, value=0.8), # temperature
|
| 249 |
-
gr.Slider(0.0, 1.0, value=0.85)
|
| 250 |
-
gr.Textbox(lines=1, value="USER:,<|endoftext|>") # stop
|
| 251 |
],
|
| 252 |
-
outputs=[gr.Chatbot(color_map=("green", "pink")),"state"],
|
| 253 |
).queue()
|
| 254 |
|
| 255 |
demo = gr.TabbedInterface(
|
|
|
|
| 111 |
max_new_tokens=10,
|
| 112 |
temperature=0.1,
|
| 113 |
top_p=1.0,
|
|
|
|
| 114 |
seed=42,
|
| 115 |
):
|
| 116 |
global model
|
|
|
|
| 147 |
max_new_tokens = int(max_new_tokens)
|
| 148 |
temperature = float(temperature)
|
| 149 |
top_p = float(top_p)
|
| 150 |
+
#stop = [x.strip(' ') for x in stop.split(',')]
|
| 151 |
seed = seed
|
| 152 |
|
| 153 |
assert 1 <= max_new_tokens <= 384
|
|
|
|
| 164 |
|
| 165 |
model.loadContext(newctx=intro+prompt)
|
| 166 |
|
| 167 |
+
generated_text = model.forward(number=max_new_tokens, stopStrings=["<|endoftext|>","USER:"],temp=temperature,top_p_usual=top_p)["output"]
|
| 168 |
|
| 169 |
generated_text = generated_text.lstrip("\n ")
|
| 170 |
+
generated_text = generated_text.rstrip("USER:")
|
| 171 |
print(f"{generated_text}")
|
| 172 |
+
'''
|
| 173 |
for stop_word in stop:
|
| 174 |
stop_word = codecs.getdecoder("unicode_escape")(stop_word)[0]
|
| 175 |
if stop_word != '' and stop_word in generated_text:
|
| 176 |
generated_text = generated_text[:generated_text.find(stop_word)]
|
| 177 |
+
'''
|
| 178 |
gc.collect()
|
| 179 |
history.append((prompt, generated_text))
|
| 180 |
return history,history
|
|
|
|
| 230 |
gr.Slider(0.0, 1.0, value=0.85), # top_p
|
| 231 |
gr.Textbox(lines=1, value="<|endoftext|>") # stop
|
| 232 |
],
|
| 233 |
+
outputs=gr.Textbox(label="Generated Output", lines=25),
|
| 234 |
examples=examples,
|
| 235 |
cache_examples=False,
|
| 236 |
).queue()
|
|
|
|
| 246 |
"state",
|
| 247 |
gr.Slider(1, 256, value=60), # max_tokens
|
| 248 |
gr.Slider(0.0, 1.0, value=0.8), # temperature
|
| 249 |
+
gr.Slider(0.0, 1.0, value=0.85) # top_p
|
|
|
|
| 250 |
],
|
| 251 |
+
outputs=[gr.Chatbot(label="Chat Log", color_map=("green", "pink")),"state"],
|
| 252 |
).queue()
|
| 253 |
|
| 254 |
demo = gr.TabbedInterface(
|