moved imports
Browse files
app.py
CHANGED
|
@@ -3,25 +3,13 @@ import os
|
|
| 3 |
import tiktoken
|
| 4 |
import streamlit as st
|
| 5 |
|
| 6 |
-
from llama_index.llms.gemini import Gemini
|
| 7 |
-
from llama_index.llms.huggingface import HuggingFaceLLM
|
| 8 |
-
from llama_index.llms.huggingface_api import HuggingFaceInferenceAPI
|
| 9 |
-
from llama_index.llms.mistralai import MistralAI
|
| 10 |
-
from llama_index.llms.openai import OpenAI
|
| 11 |
-
|
| 12 |
-
from llama_index.embeddings.openai import OpenAIEmbedding
|
| 13 |
-
from llama_index.embeddings.huggingface import HuggingFaceEmbedding, HuggingFaceInferenceAPIEmbedding
|
| 14 |
-
from llama_index.embeddings.mistralai import MistralAIEmbedding
|
| 15 |
-
|
| 16 |
from llama_index.core import (
|
| 17 |
VectorStoreIndex,
|
| 18 |
Settings,
|
| 19 |
)
|
| 20 |
|
| 21 |
from llama_parse import LlamaParse
|
| 22 |
-
|
| 23 |
-
from transformers import AutoTokenizer
|
| 24 |
-
|
| 25 |
from streamlit_pdf_viewer import pdf_viewer
|
| 26 |
|
| 27 |
MAX_OUTPUT_TOKENS = 2048
|
|
@@ -97,6 +85,10 @@ def main():
|
|
| 97 |
raise NotImplementedError(f"{provider} is not supported yet")
|
| 98 |
elif provider == 'huggingface':
|
| 99 |
if llm_name is not None and embed_name is not None:
|
|
|
|
|
|
|
|
|
|
|
|
|
| 100 |
os.environ['HFTOKEN'] = str(llm_key)
|
| 101 |
Settings.llm = HuggingFaceInferenceAPI(
|
| 102 |
model_name=llm_name,
|
|
@@ -114,6 +106,9 @@ def main():
|
|
| 114 |
)
|
| 115 |
Settings.context_window = 4096
|
| 116 |
elif provider == 'mistralai':
|
|
|
|
|
|
|
|
|
|
| 117 |
os.environ['MISTRAL_API_KEY'] = str(llm_key)
|
| 118 |
Settings.llm = MistralAI(
|
| 119 |
model=llm_name,
|
|
@@ -130,6 +125,9 @@ def main():
|
|
| 130 |
)
|
| 131 |
Settings.context_window = 4096 # max possible
|
| 132 |
elif provider == 'openai':
|
|
|
|
|
|
|
|
|
|
| 133 |
os.environ["OPENAI_API_KEY"] = str(llm_key)
|
| 134 |
Settings.llm = OpenAI(
|
| 135 |
model=llm_name,
|
|
|
|
| 3 |
import tiktoken
|
| 4 |
import streamlit as st
|
| 5 |
|
| 6 |
+
# from llama_index.llms.gemini import Gemini
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 7 |
from llama_index.core import (
|
| 8 |
VectorStoreIndex,
|
| 9 |
Settings,
|
| 10 |
)
|
| 11 |
|
| 12 |
from llama_parse import LlamaParse
|
|
|
|
|
|
|
|
|
|
| 13 |
from streamlit_pdf_viewer import pdf_viewer
|
| 14 |
|
| 15 |
MAX_OUTPUT_TOKENS = 2048
|
|
|
|
| 85 |
raise NotImplementedError(f"{provider} is not supported yet")
|
| 86 |
elif provider == 'huggingface':
|
| 87 |
if llm_name is not None and embed_name is not None:
|
| 88 |
+
from llama_index.llms.huggingface_api import HuggingFaceInferenceAPI
|
| 89 |
+
from llama_index.embeddings.huggingface import HuggingFaceInferenceAPIEmbedding
|
| 90 |
+
from transformers import AutoTokenizer
|
| 91 |
+
|
| 92 |
os.environ['HFTOKEN'] = str(llm_key)
|
| 93 |
Settings.llm = HuggingFaceInferenceAPI(
|
| 94 |
model_name=llm_name,
|
|
|
|
| 106 |
)
|
| 107 |
Settings.context_window = 4096
|
| 108 |
elif provider == 'mistralai':
|
| 109 |
+
from llama_index.llms.mistralai import MistralAI
|
| 110 |
+
from llama_index.embeddings.mistralai import MistralAIEmbedding
|
| 111 |
+
|
| 112 |
os.environ['MISTRAL_API_KEY'] = str(llm_key)
|
| 113 |
Settings.llm = MistralAI(
|
| 114 |
model=llm_name,
|
|
|
|
| 125 |
)
|
| 126 |
Settings.context_window = 4096 # max possible
|
| 127 |
elif provider == 'openai':
|
| 128 |
+
from llama_index.llms.openai import OpenAI
|
| 129 |
+
from llama_index.embeddings.openai import OpenAIEmbedding
|
| 130 |
+
|
| 131 |
os.environ["OPENAI_API_KEY"] = str(llm_key)
|
| 132 |
Settings.llm = OpenAI(
|
| 133 |
model=llm_name,
|