Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
|
@@ -20,7 +20,9 @@ model = LlamaForCausalLM.from_pretrained(
|
|
| 20 |
"decapoda-research/llama-7b-hf",
|
| 21 |
load_in_8bit=True,
|
| 22 |
torch_dtype=torch.float16,
|
| 23 |
-
device_map="auto",
|
|
|
|
|
|
|
| 24 |
)
|
| 25 |
model = PeftModel.from_pretrained(
|
| 26 |
model, "tloen/alpaca-lora-7b",
|
|
@@ -28,7 +30,8 @@ model = PeftModel.from_pretrained(
|
|
| 28 |
)
|
| 29 |
|
| 30 |
device = "cpu"
|
| 31 |
-
model.
|
|
|
|
| 32 |
model.eval()
|
| 33 |
|
| 34 |
def evaluate_raw_prompt(
|
|
|
|
| 20 |
"decapoda-research/llama-7b-hf",
|
| 21 |
load_in_8bit=True,
|
| 22 |
torch_dtype=torch.float16,
|
| 23 |
+
# device_map="auto",
|
| 24 |
+
device_map={"":"cpu"},
|
| 25 |
+
max_memory={"cpu":"12GiB"}
|
| 26 |
)
|
| 27 |
model = PeftModel.from_pretrained(
|
| 28 |
model, "tloen/alpaca-lora-7b",
|
|
|
|
| 30 |
)
|
| 31 |
|
| 32 |
device = "cpu"
|
| 33 |
+
print("model device :", model.device)
|
| 34 |
+
# model.to(device)
|
| 35 |
model.eval()
|
| 36 |
|
| 37 |
def evaluate_raw_prompt(
|