Sanjay002 commited on
Commit
e6a86d3
Β·
verified Β·
1 Parent(s): 734cabe

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -4
app.py CHANGED
@@ -6,21 +6,24 @@ from peft import PeftModel
6
  base_model = "ybelkada/falcon-7b-sharded-bf16"
7
  adapter_model = "Sanjay002/falcon-7b-mental-health-finetuned"
8
 
 
 
9
  tokenizer = AutoTokenizer.from_pretrained(base_model, trust_remote_code=True)
10
  model = AutoModelForCausalLM.from_pretrained(
11
  base_model,
12
- device_map="auto",
13
- torch_dtype=torch.bfloat16,
14
  trust_remote_code=True
15
  )
16
 
17
  model = PeftModel.from_pretrained(model, adapter_model)
 
18
  model.eval()
19
 
20
  def chat(message):
21
- inputs = tokenizer(message, return_tensors="pt").to("cuda")
22
  outputs = model.generate(**inputs, max_new_tokens=150, do_sample=True, temperature=0.7)
23
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)
24
  return response
25
 
26
- gr.Interface(fn=chat, inputs="text", outputs="text", title="🧠 Mental Health Chatbot").queue().launch(share=False)
 
6
  base_model = "ybelkada/falcon-7b-sharded-bf16"
7
  adapter_model = "Sanjay002/falcon-7b-mental-health-finetuned"
8
 
9
+ device = "cuda" if torch.cuda.is_available() else "cpu"
10
+
11
  tokenizer = AutoTokenizer.from_pretrained(base_model, trust_remote_code=True)
12
  model = AutoModelForCausalLM.from_pretrained(
13
  base_model,
14
+ device_map=None, # Don't map to GPU
15
+ torch_dtype=torch.float32 if device == "cpu" else torch.bfloat16,
16
  trust_remote_code=True
17
  )
18
 
19
  model = PeftModel.from_pretrained(model, adapter_model)
20
+ model.to(device)
21
  model.eval()
22
 
23
  def chat(message):
24
+ inputs = tokenizer(message, return_tensors="pt").to(device)
25
  outputs = model.generate(**inputs, max_new_tokens=150, do_sample=True, temperature=0.7)
26
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)
27
  return response
28
 
29
+ gr.Interface(fn=chat, inputs="text", outputs="text", title="🧠 Mental Health Chatbot").queue().launch()