Spaces:
Sleeping
Sleeping
Update chat interface logic
Browse files
app.py
CHANGED
|
@@ -137,21 +137,19 @@ MODEL_ID = "JLee0/staria-pdf-chatbot-lora"
|
|
| 137 |
BASE_ID = "MLP-KTLim/llama-3-Korean-Bllossom-8B"
|
| 138 |
|
| 139 |
from peft import PeftModel
|
| 140 |
-
|
| 141 |
-
tokenizer = AutoTokenizer.from_pretrained(BASE_ID)
|
| 142 |
-
tokenizer.pad_token = tokenizer.eos_token
|
| 143 |
|
| 144 |
base_model = AutoModelForCausalLM.from_pretrained(
|
| 145 |
-
|
| 146 |
-
|
| 147 |
-
|
| 148 |
-
|
| 149 |
-
|
| 150 |
-
|
| 151 |
-
|
| 152 |
-
|
| 153 |
-
|
| 154 |
-
|
| 155 |
|
| 156 |
rag_embedder = SentenceTransformer("JLee0/rag-embedder-staria-10epochs")
|
| 157 |
SYSTEM_PROMPT = PROMPTS["staria_after"]
|
|
|
|
| 137 |
BASE_ID = "MLP-KTLim/llama-3-Korean-Bllossom-8B"
|
| 138 |
|
| 139 |
from peft import PeftModel
|
| 140 |
+
import torch
|
|
|
|
|
|
|
| 141 |
|
| 142 |
base_model = AutoModelForCausalLM.from_pretrained(
|
| 143 |
+
BASE_ID,
|
| 144 |
+
torch_dtype=torch.float16,
|
| 145 |
+
load_in_4bit=True,
|
| 146 |
+
device_map="auto"
|
| 147 |
+
)
|
| 148 |
+
|
| 149 |
+
model = PeftModel.from_pretrained(base_model, MODEL_ID)
|
| 150 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
| 151 |
+
tokenizer.pad_token = tokenizer.eos_token
|
| 152 |
+
tokenizer.padding_side = "right"
|
| 153 |
|
| 154 |
rag_embedder = SentenceTransformer("JLee0/rag-embedder-staria-10epochs")
|
| 155 |
SYSTEM_PROMPT = PROMPTS["staria_after"]
|