IST199655
commited on
Commit
·
f63e352
1
Parent(s):
e7c3048
app.py
CHANGED
|
@@ -93,14 +93,17 @@ def respond(
|
|
| 93 |
messages.append({"role": "assistant", "content": val[1]})
|
| 94 |
messages.append({"role": "user", "content": message})
|
| 95 |
|
| 96 |
-
# Tokenize the
|
| 97 |
-
inputs = tokenizer(
|
| 98 |
-
|
| 99 |
-
|
|
|
|
|
|
|
|
|
|
| 100 |
# Generate tokens incrementally
|
| 101 |
streamer = TextStreamer(tokenizer, skip_prompt=True)
|
| 102 |
generation_kwargs = {
|
| 103 |
-
"input_ids":
|
| 104 |
"max_new_tokens": max_tokens,
|
| 105 |
"temperature": temperature,
|
| 106 |
"top_p": top_p,
|
|
|
|
| 93 |
messages.append({"role": "assistant", "content": val[1]})
|
| 94 |
messages.append({"role": "user", "content": message})
|
| 95 |
|
| 96 |
+
# Tokenize the messages
|
| 97 |
+
inputs = tokenizer.apply_chat_template(
|
| 98 |
+
messages,
|
| 99 |
+
tokenize = True,
|
| 100 |
+
add_generation_prompt = True, # Must add for generation
|
| 101 |
+
return_tensors = "pt",
|
| 102 |
+
)
|
| 103 |
# Generate tokens incrementally
|
| 104 |
streamer = TextStreamer(tokenizer, skip_prompt=True)
|
| 105 |
generation_kwargs = {
|
| 106 |
+
"input_ids": inputs,
|
| 107 |
"max_new_tokens": max_tokens,
|
| 108 |
"temperature": temperature,
|
| 109 |
"top_p": top_p,
|