added the output [] into llm for llama condition
Browse files
app.py
CHANGED
|
@@ -155,8 +155,10 @@ def bot_comms(input_text: str,
|
|
| 155 |
|
| 156 |
if llm_mode == "switch to llama":
|
| 157 |
streamer = llama_generation(input_text=input_text, history=history, temperature=temperature, max_new_tokens=max_new_tokens)
|
|
|
|
| 158 |
for text in streamer:
|
| 159 |
-
|
|
|
|
| 160 |
|
| 161 |
if llm_mode == "switch to gpt-4o":
|
| 162 |
stream = gpt_generation(input=input_text, llama_output="", mode="gpt-4o")
|
|
|
|
| 155 |
|
| 156 |
if llm_mode == "switch to llama":
|
| 157 |
streamer = llama_generation(input_text=input_text, history=history, temperature=temperature, max_new_tokens=max_new_tokens)
|
| 158 |
+
outputs = []
|
| 159 |
for text in streamer:
|
| 160 |
+
outputs.append(text)
|
| 161 |
+
yield "".join(outputs)
|
| 162 |
|
| 163 |
if llm_mode == "switch to gpt-4o":
|
| 164 |
stream = gpt_generation(input=input_text, llama_output="", mode="gpt-4o")
|