Spaces:
Sleeping
Sleeping
new model
Browse files
app.py
CHANGED
|
@@ -179,7 +179,7 @@ qa_prompt = ChatPromptTemplate.from_template(
|
|
| 179 |
# =============================
|
| 180 |
# 4) LLM — local, token-free
|
| 181 |
# =============================
|
| 182 |
-
LOCAL_MODEL_ID = os.environ.get("LOCAL_LLM", "
|
| 183 |
tok = AutoTokenizer.from_pretrained(LOCAL_MODEL_ID)
|
| 184 |
mdl = AutoModelForSeq2SeqLM.from_pretrained(LOCAL_MODEL_ID)
|
| 185 |
|
|
@@ -187,7 +187,6 @@ gen = pipeline(
|
|
| 187 |
task="text2text-generation",
|
| 188 |
model=mdl,
|
| 189 |
tokenizer=tok,
|
| 190 |
-
max_new_tokens=512,
|
| 191 |
do_sample=False, # deterministic; helps JSON adherence
|
| 192 |
)
|
| 193 |
llm = HuggingFacePipeline(pipeline=gen)
|
|
|
|
| 179 |
# =============================
|
| 180 |
# 4) LLM — local, token-free
|
| 181 |
# =============================
|
| 182 |
+
LOCAL_MODEL_ID = os.environ.get("LOCAL_LLM", "mistralai/Mixtral-8x7B-v0.1")
|
| 183 |
tok = AutoTokenizer.from_pretrained(LOCAL_MODEL_ID)
|
| 184 |
mdl = AutoModelForSeq2SeqLM.from_pretrained(LOCAL_MODEL_ID)
|
| 185 |
|
|
|
|
| 187 |
task="text2text-generation",
|
| 188 |
model=mdl,
|
| 189 |
tokenizer=tok,
|
|
|
|
| 190 |
do_sample=False, # deterministic; helps JSON adherence
|
| 191 |
)
|
| 192 |
llm = HuggingFacePipeline(pipeline=gen)
|