Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -291,7 +291,7 @@ def gen_json(input_text, max_new_tokens):
|
|
| 291 |
@GPU
|
| 292 |
def generate_text(input_text, messages=None):
|
| 293 |
if input_text == "":
|
| 294 |
-
yield "Please enter some text first."
|
| 295 |
return
|
| 296 |
|
| 297 |
token_limit=250
|
|
@@ -301,10 +301,6 @@ def generate_text(input_text, messages=None):
|
|
| 301 |
|
| 302 |
new_tags = label_map
|
| 303 |
|
| 304 |
-
if messages is None:
|
| 305 |
-
messages = []
|
| 306 |
-
else:
|
| 307 |
-
messages = []
|
| 308 |
system="""You are a knowledge assistant. Keep your responses very short."""
|
| 309 |
messages = [{"role": "system", "content": system}]+ [{"role": x["role"], "content": clean_html(x["content"])} for x in messages] +[{"role": "user", "content": input_text}]
|
| 310 |
input_text = tok.apply_chat_template(messages, add_generation_prompt=True, tokenize=False)
|
|
|
|
| 291 |
@GPU
|
| 292 |
def generate_text(input_text, messages=None):
|
| 293 |
if input_text == "":
|
| 294 |
+
yield [{"role": "assistant", "content": "Please enter some text first."}]
|
| 295 |
return
|
| 296 |
|
| 297 |
token_limit=250
|
|
|
|
| 301 |
|
| 302 |
new_tags = label_map
|
| 303 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 304 |
system="""You are a knowledge assistant. Keep your responses very short."""
|
| 305 |
messages = [{"role": "system", "content": system}]+ [{"role": x["role"], "content": clean_html(x["content"])} for x in messages] +[{"role": "user", "content": input_text}]
|
| 306 |
input_text = tok.apply_chat_template(messages, add_generation_prompt=True, tokenize=False)
|