Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -1,5 +1,4 @@
|
|
| 1 |
-
|
| 2 |
-
!huggingface-cli login
|
| 3 |
import streamlit as st
|
| 4 |
from llama_index.core import VectorStoreIndex, SimpleDirectoryReader
|
| 5 |
from llama_index.core import ServiceContext
|
|
@@ -40,8 +39,8 @@ llm = HuggingFaceLLM(
|
|
| 40 |
generate_kwargs={"temperature": 0.7, "do_sample": False},
|
| 41 |
system_prompt=system_prompt,
|
| 42 |
query_wrapper_prompt=query_wrapper_prompt,
|
| 43 |
-
tokenizer_name="mistralai/Mistral-7B-Instruct-v0.1",
|
| 44 |
-
model_name="mistralai/Mistral-7B-Instruct-v0.1",
|
| 45 |
device_map="auto",
|
| 46 |
stopping_ids=[50278, 50279, 50277, 1, 0],
|
| 47 |
tokenizer_kwargs={"max_length": 4096},
|
|
|
|
| 1 |
+
import gradio as gr
|
|
|
|
| 2 |
import streamlit as st
|
| 3 |
from llama_index.core import VectorStoreIndex, SimpleDirectoryReader
|
| 4 |
from llama_index.core import ServiceContext
|
|
|
|
| 39 |
generate_kwargs={"temperature": 0.7, "do_sample": False},
|
| 40 |
system_prompt=system_prompt,
|
| 41 |
query_wrapper_prompt=query_wrapper_prompt,
|
| 42 |
+
tokenizer_name=gr.Interface.load("mistralai/Mistral-7B-Instruct-v0.1"),
|
| 43 |
+
model_name=gr.Interface.load("mistralai/Mistral-7B-Instruct-v0.1"),
|
| 44 |
device_map="auto",
|
| 45 |
stopping_ids=[50278, 50279, 50277, 1, 0],
|
| 46 |
tokenizer_kwargs={"max_length": 4096},
|