File size: 1,793 Bytes
e238a65
 
66d76cd
152ab18
e238a65
5367ce5
e238a65
 
1baa139
5367ce5
1baa139
e238a65
 
 
 
 
 
 
 
 
2861dbc
e238a65
ecfc0b7
 
e238a65
0216182
 
 
d2582d3
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
import gradio as gr
from huggingface_hub import InferenceClient #InferenceClient class

client = InferenceClient("deepseek-ai/DeepSeek-R1-Distill-Qwen-32B") #Create an instance of InferenceClient connected to the Qwen/Qwen2.5-7B-Instruct text-generation model
#this client will handle making requests to the model to generate responses


def respond(message, history): #function for Gradio to call
#Gradio passes arguments as parameters: the user's most recent input which is a string ("message"), and "history" which is the list of past messages
#I have to put it in this order because Gradio will always past the current user input first and then the convo history  
# however for now, this chatbot won't use the history parameter anyway
    messages = [{"role": "system", "content": "You are a friendly chatbot."}] #dict in list to store messages

    #Add convo history to the messages if there's convo history
    if history: 
        messages.extend(history)

    messages.append({"role": "user", "content": message}) #add the current user’s message to the messages list

    # chat completion API call forwarding the messages & other params to model
    response = client.chat_completion(messages, max_tokens=100, temperature = 2, top_p=0.95) #deepseek R1 recomended temp range: 0.5-0.7
    return response.choices[0].message.content.strip()

# defining chatbot
chatbot = gr.ChatInterface(respond, title = "", description = "") #using gradio to quickly build a chatbot UI (w/ convo history & user input)
# passing fxn into a fxn, passing echo for gradio to call each time the user sends a message
# Adding parentheses would call the function and pass its return value instead, I didn't include () because I want Gradio to call it later, not right now

chatbot.launch() #launch chatbot