Spaces:
Sleeping
Sleeping
Update chat interface logic
Browse files
app.py
CHANGED
|
@@ -134,16 +134,24 @@ def answer_query(
|
|
| 134 |
EMB_KEY_PATH = "staria_keys_embed.json"
|
| 135 |
VALUE_TEXT_PATH = "staria_values.json"
|
| 136 |
MODEL_ID = "JLee0/staria-pdf-chatbot-lora"
|
|
|
|
| 137 |
|
| 138 |
-
|
|
|
|
|
|
|
| 139 |
tokenizer.pad_token = tokenizer.eos_token
|
| 140 |
|
| 141 |
-
|
| 142 |
-
|
| 143 |
-
MODEL_ID,
|
| 144 |
load_in_8bit=True,
|
| 145 |
device_map="auto"
|
| 146 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 147 |
|
| 148 |
rag_embedder = SentenceTransformer("JLee0/rag-embedder-staria-10epochs")
|
| 149 |
SYSTEM_PROMPT = PROMPTS["staria_after"]
|
|
|
|
| 134 |
EMB_KEY_PATH = "staria_keys_embed.json"
|
| 135 |
VALUE_TEXT_PATH = "staria_values.json"
|
| 136 |
MODEL_ID = "JLee0/staria-pdf-chatbot-lora"
|
| 137 |
+
BASE_ID = "MLP-KTLim/llama-3-Korean-Bllossom-8B"
|
| 138 |
|
| 139 |
+
from peft import PeftModel
|
| 140 |
+
|
| 141 |
+
tokenizer = AutoTokenizer.from_pretrained(BASE_ID)
|
| 142 |
tokenizer.pad_token = tokenizer.eos_token
|
| 143 |
|
| 144 |
+
base_model = AutoModelForCausalLM.from_pretrained(
|
| 145 |
+
BASE_ID,
|
|
|
|
| 146 |
load_in_8bit=True,
|
| 147 |
device_map="auto"
|
| 148 |
)
|
| 149 |
+
model = PeftModel.from_pretrained(
|
| 150 |
+
base_model,
|
| 151 |
+
MODEL_ID,
|
| 152 |
+
device_map="auto"
|
| 153 |
+
)
|
| 154 |
+
|
| 155 |
|
| 156 |
rag_embedder = SentenceTransformer("JLee0/rag-embedder-staria-10epochs")
|
| 157 |
SYSTEM_PROMPT = PROMPTS["staria_after"]
|