try mistral model
Browse files- app/chains.py +4 -2
app/chains.py
CHANGED
|
@@ -14,10 +14,12 @@ load_dotenv()
|
|
| 14 |
|
| 15 |
# data_indexer = DataIndexer()
|
| 16 |
|
| 17 |
-
|
|
|
|
|
|
|
| 18 |
|
| 19 |
llm = HuggingFaceEndpoint(
|
| 20 |
-
model=
|
| 21 |
huggingfacehub_api_token=os.environ['HF_TOKEN'],
|
| 22 |
max_new_tokens=512,
|
| 23 |
stop_sequences=[tokenizer.eos_token],
|
|
|
|
| 14 |
|
| 15 |
# data_indexer = DataIndexer()
|
| 16 |
|
| 17 |
+
MODEL_ID = "mistralai/Mistral-7B-Instruct-v0.3"
|
| 18 |
+
|
| 19 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
| 20 |
|
| 21 |
llm = HuggingFaceEndpoint(
|
| 22 |
+
model=MODEL_ID,
|
| 23 |
huggingfacehub_api_token=os.environ['HF_TOKEN'],
|
| 24 |
max_new_tokens=512,
|
| 25 |
stop_sequences=[tokenizer.eos_token],
|