Spaces:
Sleeping
Sleeping
Trying the Qwen LLM
Browse files
app.py
CHANGED
|
@@ -2,7 +2,7 @@ import gradio as gr
|
|
| 2 |
import random
|
| 3 |
from huggingface_hub import InferenceClient
|
| 4 |
|
| 5 |
-
client = InferenceClient("
|
| 6 |
#Changes the LLM being used
|
| 7 |
|
| 8 |
def respond(message, history):
|
|
@@ -47,4 +47,4 @@ def random_message(message, history):
|
|
| 47 |
|
| 48 |
chatbot = gr.ChatInterface(respond, type = "messages")
|
| 49 |
|
| 50 |
-
chatbot.launch()
|
|
|
|
| 2 |
import random
|
| 3 |
from huggingface_hub import InferenceClient
|
| 4 |
|
| 5 |
+
client = InferenceClient("Qwen/Qwen2.5-72B-Instruct")
|
| 6 |
#Changes the LLM being used
|
| 7 |
|
| 8 |
def respond(message, history):
|
|
|
|
| 47 |
|
| 48 |
chatbot = gr.ChatInterface(respond, type = "messages")
|
| 49 |
|
| 50 |
+
chatbot.launch(debug = True)
|