Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -5,7 +5,7 @@ import torch
|
|
| 5 |
|
| 6 |
import gradio as gr
|
| 7 |
import random
|
| 8 |
-
|
| 9 |
client=InferenceClient("mistralai/Mistral-7B-Instruct-v0.2")
|
| 10 |
#STEP 2 from semantic search (read file)
|
| 11 |
# Open the water_cycle.txt file in read mode with UTF-8 encoding
|
|
@@ -111,9 +111,7 @@ def respond(message, history):
|
|
| 111 |
{
|
| 112 |
"role": "system",
|
| 113 |
"content": (
|
| 114 |
-
"You are a very smart, arrogant professor who knows a lot about physics. "
|
| 115 |
-
"You answer the questions from the user directly and concisely as if they were your physics student. Base your response on the provided context."
|
| 116 |
-
"Keep your answers below 100 words!"
|
| 117 |
)
|
| 118 |
},
|
| 119 |
{
|
|
|
|
| 5 |
|
| 6 |
import gradio as gr
|
| 7 |
import random
|
| 8 |
+
#change of LLM below
|
| 9 |
client=InferenceClient("mistralai/Mistral-7B-Instruct-v0.2")
|
| 10 |
#STEP 2 from semantic search (read file)
|
| 11 |
# Open the water_cycle.txt file in read mode with UTF-8 encoding
|
|
|
|
| 111 |
{
|
| 112 |
"role": "system",
|
| 113 |
"content": (
|
| 114 |
+
"You are a very smart, arrogant professor who knows a lot about physics. You answer the questions from the user directly and concisely as if they were your physics student. Base your response on the provided context."
|
|
|
|
|
|
|
| 115 |
)
|
| 116 |
},
|
| 117 |
{
|