Abeersherif commited on
Commit
9253184
·
verified ·
1 Parent(s): 3b12881

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -8
app.py CHANGED
@@ -1,10 +1,10 @@
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
3
 
 
 
4
 
5
- MODEL_NAME = "smol-medical-meadow-FT"
6
-
7
- # Load model & tokenizer
8
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
9
  model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
10
 
@@ -15,19 +15,17 @@ pipe = pipeline(
15
  )
16
 
17
  def respond(message, history, system_message, max_tokens, temperature, top_p):
18
- # Build chat-style messages
19
  messages = [{"role": "system", "content": system_message}]
20
  messages.extend(history)
21
  messages.append({"role": "user", "content": message})
22
 
23
- # Convert to model-specific chat template
24
  prompt = tokenizer.apply_chat_template(
25
  messages,
26
  tokenize=False,
27
  add_generation_prompt=True,
28
  )
29
 
30
- # Generate
31
  out = pipe(
32
  prompt,
33
  max_new_tokens=max_tokens,
@@ -36,18 +34,20 @@ def respond(message, history, system_message, max_tokens, temperature, top_p):
36
  do_sample=True,
37
  )[0]["generated_text"]
38
 
39
- # Extract assistant-only text
40
  if "<|im_start|>assistant" in out:
41
  out = out.split("<|im_start|>assistant", 1)[-1]
42
  out = out.replace("<|im_end|>", "").strip()
43
 
44
  return out
45
 
 
46
  chatbot = gr.ChatInterface(
47
  fn=respond,
48
  type="messages",
49
  additional_inputs=[
50
- gr.Textbox("Give short answers with brief logical reasoning.", label="System message"),
 
51
  gr.Slider(1, 512, value=256, step=1, label="Max new tokens"),
52
  gr.Slider(0.1, 4.0, value=0.7, step=0.1, label="Temperature"),
53
  gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-p"),
 
1
  import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
3
 
4
+ # ✅ Your fine-tuned model on Hugging Face
5
+ MODEL_NAME = "Abeersherif/Medical_Homework2"
6
 
7
+ # Load model & tokenizer from HF
 
 
8
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
9
  model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
10
 
 
15
  )
16
 
17
  def respond(message, history, system_message, max_tokens, temperature, top_p):
18
+ # Build chat-style messages for the model’s chat template
19
  messages = [{"role": "system", "content": system_message}]
20
  messages.extend(history)
21
  messages.append({"role": "user", "content": message})
22
 
 
23
  prompt = tokenizer.apply_chat_template(
24
  messages,
25
  tokenize=False,
26
  add_generation_prompt=True,
27
  )
28
 
 
29
  out = pipe(
30
  prompt,
31
  max_new_tokens=max_tokens,
 
34
  do_sample=True,
35
  )[0]["generated_text"]
36
 
37
+ # Extract only the assistant part (for Smol-style chat templates)
38
  if "<|im_start|>assistant" in out:
39
  out = out.split("<|im_start|>assistant", 1)[-1]
40
  out = out.replace("<|im_end|>", "").strip()
41
 
42
  return out
43
 
44
+
45
  chatbot = gr.ChatInterface(
46
  fn=respond,
47
  type="messages",
48
  additional_inputs=[
49
+ gr.Textbox("You are a helpful medical assistant. Answer concisely with brief reasoning.",
50
+ label="System message"),
51
  gr.Slider(1, 512, value=256, step=1, label="Max new tokens"),
52
  gr.Slider(0.1, 4.0, value=0.7, step=0.1, label="Temperature"),
53
  gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-p"),