Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -13,11 +13,9 @@ tokenizer = AutoTokenizer.from_pretrained(base_model_name)
|
|
| 13 |
|
| 14 |
# 加载基础模型到 CPU
|
| 15 |
base_model = AutoModelForCausalLM.from_pretrained(base_model_name, torch_dtype=torch.float32, device_map=None)
|
| 16 |
-
base_model.to("cpu") # 确保在 CPU 上运行
|
| 17 |
|
| 18 |
# 加载 LoRA 适配器
|
| 19 |
model = PeftModel.from_pretrained(base_model, adapter_model_name)
|
| 20 |
-
model.to("cpu") # 确保适配器也在 CPU 上运行
|
| 21 |
|
| 22 |
# 运行测试
|
| 23 |
def chat_with_model(prompt):
|
|
|
|
| 13 |
|
| 14 |
# 加载基础模型到 CPU
|
| 15 |
base_model = AutoModelForCausalLM.from_pretrained(base_model_name, torch_dtype=torch.float32, device_map=None)
|
|
|
|
| 16 |
|
| 17 |
# 加载 LoRA 适配器
|
| 18 |
model = PeftModel.from_pretrained(base_model, adapter_model_name)
|
|
|
|
| 19 |
|
| 20 |
# 运行测试
|
| 21 |
def chat_with_model(prompt):
|