Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
|
@@ -10,11 +10,13 @@ adapter_model_name = "tiankuo1111/LLAMA3-TEST"
|
|
| 10 |
# 加载 tokenizer
|
| 11 |
tokenizer = AutoTokenizer.from_pretrained(base_model_name)
|
| 12 |
|
| 13 |
-
#
|
| 14 |
-
base_model = AutoModelForCausalLM.from_pretrained(base_model_name, torch_dtype=torch.
|
|
|
|
| 15 |
|
| 16 |
# 加载 LoRA 适配器
|
| 17 |
model = PeftModel.from_pretrained(base_model, adapter_model_name)
|
|
|
|
| 18 |
|
| 19 |
# 运行测试
|
| 20 |
def chat_with_model(prompt):
|
|
|
|
| 10 |
# 加载 tokenizer
|
| 11 |
tokenizer = AutoTokenizer.from_pretrained(base_model_name)
|
| 12 |
|
| 13 |
+
# 加载基础模型到 CPU
|
| 14 |
+
base_model = AutoModelForCausalLM.from_pretrained(base_model_name, torch_dtype=torch.float32, device_map=None)
|
| 15 |
+
base_model.to("cpu") # 确保在 CPU 上运行
|
| 16 |
|
| 17 |
# 加载 LoRA 适配器
|
| 18 |
model = PeftModel.from_pretrained(base_model, adapter_model_name)
|
| 19 |
+
model.to("cpu") # 确保适配器也在 CPU 上运行
|
| 20 |
|
| 21 |
# 运行测试
|
| 22 |
def chat_with_model(prompt):
|