Adityabhatia0204 commited on
Commit
ba337c7
·
verified ·
1 Parent(s): 06964e3

Update chat.py

Browse files
Files changed (1) hide show
  1. chat.py +9 -17
chat.py CHANGED
@@ -1,23 +1,13 @@
1
  from peft import AutoPeftModelForCausalLM
2
- from transformers import GenerationConfig
3
- from transformers import AutoTokenizer
4
  import torch
5
  import os
6
- import os
7
 
 
8
  os.environ["HF_HOME"] = "./hf_home"
9
  os.environ["TRANSFORMERS_CACHE"] = "./hf_home/transformers"
10
- os.makedirs("./hf_home/transformers", exist_ok=True)
11
-
12
-
13
- # Set Hugging Face cache directory to a folder you have access to
14
- os.environ["HF_HOME"] = "/data"
15
- os.environ["TRANSFORMERS_CACHE"] = "/data/transformers"
16
-
17
- # Create cache directory if doesn't exist
18
- os.makedirs("/data/transformers", exist_ok=True)
19
-
20
 
 
21
  tokenizer = AutoTokenizer.from_pretrained("Vasanth/mistral-finetuned-alpaca")
22
 
23
  model = AutoPeftModelForCausalLM.from_pretrained(
@@ -25,7 +15,8 @@ model = AutoPeftModelForCausalLM.from_pretrained(
25
  low_cpu_mem_usage=True,
26
  return_dict=True,
27
  torch_dtype=torch.float16,
28
- device_map="cuda")
 
29
 
30
  generation_config = GenerationConfig(
31
  do_sample=True,
@@ -36,7 +27,8 @@ generation_config = GenerationConfig(
36
  )
37
 
38
  def chatbot(message):
39
- input_str = "###Human: " + message + " ###Assistant: "
40
- inputs = tokenizer(input_str, return_tensors="pt").to("cuda")
41
  outputs = model.generate(**inputs, generation_config=generation_config)
42
- return tokenizer.decode(outputs[0], skip_special_tokens=True).replace(input_str, '')
 
 
1
  from peft import AutoPeftModelForCausalLM
2
+ from transformers import GenerationConfig, AutoTokenizer
 
3
  import torch
4
  import os
 
5
 
6
+ # Set huggingface cache directory
7
  os.environ["HF_HOME"] = "./hf_home"
8
  os.environ["TRANSFORMERS_CACHE"] = "./hf_home/transformers"
 
 
 
 
 
 
 
 
 
 
9
 
10
+ # Load tokenizer and model
11
  tokenizer = AutoTokenizer.from_pretrained("Vasanth/mistral-finetuned-alpaca")
12
 
13
  model = AutoPeftModelForCausalLM.from_pretrained(
 
15
  low_cpu_mem_usage=True,
16
  return_dict=True,
17
  torch_dtype=torch.float16,
18
+ device_map="auto"
19
+ )
20
 
21
  generation_config = GenerationConfig(
22
  do_sample=True,
 
27
  )
28
 
29
  def chatbot(message):
30
+ input_str = f"###Human: {message} ###Assistant: "
31
+ inputs = tokenizer(input_str, return_tensors="pt").to(model.device)
32
  outputs = model.generate(**inputs, generation_config=generation_config)
33
+ response = tokenizer.decode(outputs[0], skip_special_tokens=True).replace(input_str, "")
34
+ return response.strip()