tuandunghcmut commited on
Commit
7ccc6c1
·
1 Parent(s): 9aaf868
Files changed (3) hide show
  1. app.py +14 -9
  2. requirements.txt +1 -1
  3. temp.py +5 -0
app.py CHANGED
@@ -72,15 +72,20 @@ class QwenModelHandler:
72
  )
73
 
74
  # Load with standard precision on CPU
75
- self.model = AutoModelForCausalLM.from_pretrained(
76
- model_path,
77
- torch_dtype=torch.float32,
78
- device_map="cpu",
79
- trust_remote_code=True,
80
- # Explicitly disable quantization
81
- load_in_8bit=False,
82
- load_in_4bit=False
83
- )
 
 
 
 
 
84
 
85
  if self.tokenizer.pad_token is None and self.tokenizer.eos_token is not None:
86
  self.tokenizer.pad_token = self.tokenizer.eos_token
 
72
  )
73
 
74
  # Load with standard precision on CPU
75
+ from peft import PeftModel
76
+ from transformers import AutoModelForCausalLM
77
+
78
+ base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-Coder-1.5B-Instruct")
79
+ self.model = PeftModel.from_pretrained(base_model, "tuandunghcmut/Qwen25_Coder_MultipleChoice_v4")
80
+ # self.model = AutoModelForCausalLM.from_pretrained(
81
+ # model_path,
82
+ # torch_dtype=torch.float32,
83
+ # device_map="cpu",
84
+ # trust_remote_code=True,
85
+ # # Explicitly disable quantization
86
+ # load_in_8bit=False,
87
+ # load_in_4bit=False
88
+ # )
89
 
90
  if self.tokenizer.pad_token is None and self.tokenizer.eos_token is not None:
91
  self.tokenizer.pad_token = self.tokenizer.eos_token
requirements.txt CHANGED
@@ -4,4 +4,4 @@ transformers
4
  pyyaml
5
  accelerate
6
  peft
7
- bitsandbytes
 
4
  pyyaml
5
  accelerate
6
  peft
7
+ # bitsandbytes
temp.py ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ from peft import PeftModel
2
+ from transformers import AutoModelForCausalLM
3
+
4
+ base_model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen2.5-Coder-1.5B-Instruct")
5
+ model = PeftModel.from_pretrained(base_model, "tuandunghcmut/Qwen25_Coder_MultipleChoice_v4")