Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -3,7 +3,7 @@ from huggingface_hub import InferenceClient
|
|
| 3 |
import gradio as gr
|
| 4 |
import random
|
| 5 |
|
| 6 |
-
client=InferenceClient("Qwen/Qwen2.5-72B-Instruct")
|
| 7 |
#deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
|
| 8 |
|
| 9 |
def respond(message, history):
|
|
@@ -22,10 +22,10 @@ def respond(message, history):
|
|
| 22 |
|
| 23 |
messages.append(
|
| 24 |
{"role":"user",
|
| 25 |
-
"content":
|
| 26 |
)
|
| 27 |
|
| 28 |
-
response = client.chat_completion(messages, max_tokens=100, temperature=1.3, top_p
|
| 29 |
#temperature and top_p control randomness
|
| 30 |
|
| 31 |
|
|
|
|
| 3 |
import gradio as gr
|
| 4 |
import random
|
| 5 |
|
| 6 |
+
client = InferenceClient("Qwen/Qwen2.5-72B-Instruct")
|
| 7 |
#deepseek-ai/DeepSeek-R1-Distill-Qwen-32B
|
| 8 |
|
| 9 |
def respond(message, history):
|
|
|
|
| 22 |
|
| 23 |
messages.append(
|
| 24 |
{"role":"user",
|
| 25 |
+
"content": message}
|
| 26 |
)
|
| 27 |
|
| 28 |
+
response = client.chat_completion(messages, max_tokens = 100, temperature = 1.3, top_p = .3)
|
| 29 |
#temperature and top_p control randomness
|
| 30 |
|
| 31 |
|