emya commited on
Commit
2f9c1ea
·
1 Parent(s): 8ad2f92

use base and new models

Browse files
Files changed (1) hide show
  1. app.py +14 -2
app.py CHANGED
@@ -1,8 +1,20 @@
1
  import gradio as gr
2
 
3
- from transformers import pipeline
4
 
5
- pipe = pipeline("translation", model="emya/vicuna-7b-v1.5-steve-jobs-8bit-v1")
 
 
 
 
 
 
 
 
 
 
 
 
6
 
7
  def predict(text):
8
  prompt = f"{text} (Answer in a few sentences)"
 
1
  import gradio as gr
2
 
3
+ from transformers import AutoModelForCausalLM, pipeline
4
 
5
+ model_name = "lmsys/vicuna-7b-v1.5"
6
+ base_model = AutoModelForCausalLM.from_pretrained(
7
+ model_name,
8
+ low_cpu_mem_usage=True,
9
+ return_dict=True,
10
+ #torch_dtype=torch.float16,
11
+ device_map="auto",
12
+ load_in_8bit=True,
13
+ )
14
+ new_model = "emya/vicuna-7b-v1.5-steve-jobs-8bit-v1"
15
+ model = PeftModel.from_pretrained(base_model, new_model, load_in_8bit=True)
16
+
17
+ pipe = pipeline("translation", model=model)
18
 
19
  def predict(text):
20
  prompt = f"{text} (Answer in a few sentences)"