Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -122,15 +122,15 @@ model = Phi3WithImage(phi3_model_name, image_encoder, image_embed_dim, bnb_conf
|
|
| 122 |
# Load LoRA model
|
| 123 |
model.phi3 = PeftModel.from_pretrained(model.phi3, lora_model_path, device_map="auto", offload_dir='./offload')
|
| 124 |
model.phi3 = model.phi3.merge_and_unload()
|
| 125 |
-
model.phi3.save_pretrained("merged_model_fp16")
|
| 126 |
|
| 127 |
#model.phi3 = PeftModel.from_pretrained(model.phi3, lora_model_path, device_map="auto")
|
| 128 |
-
model.phi3 = AutoModelForCausalLM.from_pretrained(
|
| 129 |
-
"merged_model_fp16",
|
| 130 |
-
quantization_config=bnb_config,
|
| 131 |
-
torch_dtype=torch.bfloat16,
|
| 132 |
-
device_map="auto"
|
| 133 |
-
)
|
| 134 |
model.eval() # Set to evaluation mode
|
| 135 |
|
| 136 |
# 3. Inference Function
|
|
|
|
| 122 |
# Load LoRA model
|
| 123 |
model.phi3 = PeftModel.from_pretrained(model.phi3, lora_model_path, device_map="auto", offload_dir='./offload')
|
| 124 |
model.phi3 = model.phi3.merge_and_unload()
|
| 125 |
+
#model.phi3.save_pretrained("merged_model_fp16")
|
| 126 |
|
| 127 |
#model.phi3 = PeftModel.from_pretrained(model.phi3, lora_model_path, device_map="auto")
|
| 128 |
+
#model.phi3 = AutoModelForCausalLM.from_pretrained(
|
| 129 |
+
# "merged_model_fp16",
|
| 130 |
+
# quantization_config=bnb_config,
|
| 131 |
+
# torch_dtype=torch.bfloat16,
|
| 132 |
+
# device_map="auto"
|
| 133 |
+
#)
|
| 134 |
model.eval() # Set to evaluation mode
|
| 135 |
|
| 136 |
# 3. Inference Function
|