Castial commited on
Commit
5f82807
·
verified ·
1 Parent(s): 5ab387b

Upload 2 files

Browse files
Files changed (2) hide show
  1. app.py +9 -7
  2. requirements.txt +1 -1
app.py CHANGED
@@ -1,18 +1,20 @@
1
  import gradio as gr
2
  import torch
3
  from transformers import Qwen2ForCausalLM, AutoTokenizer
 
4
 
 
 
5
 
6
- checkpoint = "Castial/FinQwen-2.5b-v2"
7
 
8
- # Load model + tokenizer
9
- model = Qwen2ForCausalLM.from_pretrained(
10
- checkpoint,
11
  device_map="auto",
12
- dtype="auto",
13
  )
14
- tokenizer = AutoTokenizer.from_pretrained(checkpoint)
15
 
 
16
 
17
  def respond(
18
  message,
@@ -80,5 +82,5 @@ chatbot = gr.ChatInterface(
80
  with gr.Blocks() as demo:
81
  chatbot.render()
82
 
83
- if name == "main":
84
  demo.launch()
 
1
  import gradio as gr
2
  import torch
3
  from transformers import Qwen2ForCausalLM, AutoTokenizer
4
+ from peft import PeftModel
5
 
6
+ base_model_id = "Qwen/Qwen2.5-3B-Instruct" # base
7
+ adapter_path = "Castial/FinQwen-2.5b-v2" # your fine-tuned adapter
8
 
 
9
 
10
+ tokenizer = AutoTokenizer.from_pretrained(base_model_id)
11
+ base_model = AutoModelForCausalLM.from_pretrained(
12
+ base_model_id,
13
  device_map="auto",
14
+ torch_dtype="auto",
15
  )
 
16
 
17
+ model = PeftModel.from_pretrained(base_model, adapter_path)
18
 
19
  def respond(
20
  message,
 
82
  with gr.Blocks() as demo:
83
  chatbot.render()
84
 
85
+ if __name__ == "__main__":
86
  demo.launch()
requirements.txt CHANGED
@@ -2,5 +2,5 @@ torch
2
  transformers
3
  gradio
4
  accelerate
5
-
6
 
 
2
  transformers
3
  gradio
4
  accelerate
5
+ peft
6