Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -8,7 +8,7 @@ token = os.getenv("HF_TOKEN")
|
|
| 8 |
client = InferenceClient("meta-llama/Llama-3.2-3B-Instruct", token=token)
|
| 9 |
|
| 10 |
def respond(message, history):
|
| 11 |
-
messages = [{"role": "system", "content": "Ты — полезный ИИ
|
| 12 |
|
| 13 |
# Наполняем историю
|
| 14 |
for user_msg, assistant_msg in history:
|
|
@@ -27,7 +27,7 @@ def respond(message, history):
|
|
| 27 |
messages=messages,
|
| 28 |
max_tokens=1000,
|
| 29 |
stream=True,
|
| 30 |
-
temperature=0.
|
| 31 |
):
|
| 32 |
token_text = message_chunk.choices[0].delta.content
|
| 33 |
if token_text:
|
|
|
|
| 8 |
client = InferenceClient("meta-llama/Llama-3.2-3B-Instruct", token=token)
|
| 9 |
|
| 10 |
def respond(message, history):
|
| 11 |
+
messages = [{"role": "system", "content": "Ты — полезный ИИ ассистент и тебя зовут Flare. Отвечай на русском языке."}]
|
| 12 |
|
| 13 |
# Наполняем историю
|
| 14 |
for user_msg, assistant_msg in history:
|
|
|
|
| 27 |
messages=messages,
|
| 28 |
max_tokens=1000,
|
| 29 |
stream=True,
|
| 30 |
+
temperature=0.3,
|
| 31 |
):
|
| 32 |
token_text = message_chunk.choices[0].delta.content
|
| 33 |
if token_text:
|