akcoderspark commited on
Commit
89fb040
·
verified ·
1 Parent(s): 219db16

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -11,8 +11,8 @@ model_name = "cognitivecomputations/dolphin-2_6-phi-2"
11
  tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
12
  model = AutoModelForCausalLM.from_pretrained(
13
  model_name,
14
- torch_dtype=torch.float16,
15
- device_map="auto",
16
  low_cpu_mem_usage=True,
17
  trust_remote_code=True
18
  )
@@ -38,7 +38,7 @@ def generate():
38
 
39
  prompt = f"<|im_start|>system\n{system_msg}<|im_end|>\n<|im_start|>user\n{user_msg}<|im_end|>\n<|im_start|>assistant\n"
40
 
41
- inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
42
 
43
  with torch.no_grad():
44
  outputs = model.generate(
 
11
  tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
12
  model = AutoModelForCausalLM.from_pretrained(
13
  model_name,
14
+ torch_dtype=torch.float32, # ✅ Changed to float32 for CPU
15
+ device_map="cpu", # ✅ Explicitly use CPU
16
  low_cpu_mem_usage=True,
17
  trust_remote_code=True
18
  )
 
38
 
39
  prompt = f"<|im_start|>system\n{system_msg}<|im_end|>\n<|im_start|>user\n{user_msg}<|im_end|>\n<|im_start|>assistant\n"
40
 
41
+ inputs = tokenizer(prompt, return_tensors="pt")
42
 
43
  with torch.no_grad():
44
  outputs = model.generate(