Elieon commited on
Commit
ef332e0
·
verified ·
1 Parent(s): e8f2f5b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -8
app.py CHANGED
@@ -11,13 +11,14 @@ client = InferenceClient(token=HF_TOKEN)
11
 
12
  # Chat function
13
  def respond(message, history, max_tokens, temperature, top_p):
14
- # Build prompt in OpenAI-style
15
  prompt = [{"role": "system", "content": system_message}]
16
- for msg in history:
17
- if msg["role"] == "user":
18
- prompt.append({"role": "user", "content": msg["content"]})
19
- elif msg["role"] == "assistant":
20
- prompt.append({"role": "assistant", "content": msg["content"]})
 
21
  prompt.append({"role": "user", "content": message})
22
 
23
  response = []
@@ -32,7 +33,7 @@ def respond(message, history, max_tokens, temperature, top_p):
32
 
33
  for chunk in stream:
34
  if not chunk.choices:
35
- continue # skip empty chunks safely
36
  delta = chunk.choices[0].delta
37
  token = getattr(delta, "content", None)
38
  if token:
@@ -42,7 +43,6 @@ def respond(message, history, max_tokens, temperature, top_p):
42
  # UI
43
  app = gr.ChatInterface(
44
  fn=respond,
45
- type="messages", # ✅ fixes Gradio warning
46
  additional_inputs=[
47
  gr.Slider(16, 2048, value=512, step=1, label="Max Tokens"),
48
  gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature"),
 
11
 
12
  # Chat function
13
  def respond(message, history, max_tokens, temperature, top_p):
14
+ # Build prompt
15
  prompt = [{"role": "system", "content": system_message}]
16
+
17
+ # history is a list of (user, assistant) tuples
18
+ for user_msg, assistant_msg in history:
19
+ prompt.append({"role": "user", "content": user_msg})
20
+ prompt.append({"role": "assistant", "content": assistant_msg})
21
+
22
  prompt.append({"role": "user", "content": message})
23
 
24
  response = []
 
33
 
34
  for chunk in stream:
35
  if not chunk.choices:
36
+ continue
37
  delta = chunk.choices[0].delta
38
  token = getattr(delta, "content", None)
39
  if token:
 
43
  # UI
44
  app = gr.ChatInterface(
45
  fn=respond,
 
46
  additional_inputs=[
47
  gr.Slider(16, 2048, value=512, step=1, label="Max Tokens"),
48
  gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature"),