Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -11,17 +11,14 @@ tokenizer = AutoTokenizer.from_pretrained(BASE_MODEL)
|
|
| 11 |
tokenizer.pad_token = tokenizer.eos_token
|
| 12 |
|
| 13 |
# Base model (CPU)
|
| 14 |
-
# 1. Base model (disk offload aktif)
|
| 15 |
model = AutoModelForCausalLM.from_pretrained(
|
| 16 |
BASE_MODEL,
|
| 17 |
torch_dtype=torch.float32,
|
| 18 |
-
device_map="
|
| 19 |
-
offload_dir="offload",
|
| 20 |
-
offload_state_dict=True,
|
| 21 |
low_cpu_mem_usage=True
|
| 22 |
)
|
| 23 |
|
| 24 |
-
#
|
| 25 |
model = PeftModel.from_pretrained(model, LORA_PATH)
|
| 26 |
model.eval()
|
| 27 |
|
|
|
|
| 11 |
tokenizer.pad_token = tokenizer.eos_token
|
| 12 |
|
| 13 |
# Base model (CPU)
|
|
|
|
| 14 |
model = AutoModelForCausalLM.from_pretrained(
|
| 15 |
BASE_MODEL,
|
| 16 |
torch_dtype=torch.float32,
|
| 17 |
+
device_map={"": "cpu"},
|
|
|
|
|
|
|
| 18 |
low_cpu_mem_usage=True
|
| 19 |
)
|
| 20 |
|
| 21 |
+
# Load LoRA
|
| 22 |
model = PeftModel.from_pretrained(model, LORA_PATH)
|
| 23 |
model.eval()
|
| 24 |
|