kxrrrr commited on
Commit
aabb83e
Β·
verified Β·
1 Parent(s): 5a29131

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +22 -10
app.py CHANGED
@@ -1,7 +1,24 @@
1
  import streamlit as st
2
  from llama_index.core import VectorStoreIndex, SimpleDirectoryReader, Settings
3
  from llama_index.embeddings.huggingface import HuggingFaceEmbedding
4
- from llama_index.llms.huggingface import HuggingFaceLLM
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
 
6
  st.set_page_config(page_title="llama_index_demo", page_icon="πŸ¦œπŸ”—")
7
  st.title("llama_index_demo")
@@ -10,19 +27,14 @@ st.title("llama_index_demo")
10
  @st.cache_resource
11
  def init_models():
12
  embed_model = HuggingFaceEmbedding(
13
- model_name="/root/model/sentence-transformer"
14
  )
15
  Settings.embed_model = embed_model
16
 
17
- llm = HuggingFaceLLM(
18
- model_name="/root/model/internlm2-chat-1_8b",
19
- tokenizer_name="/root/model/internlm2-chat-1_8b",
20
- model_kwargs={"trust_remote_code": True},
21
- tokenizer_kwargs={"trust_remote_code": True}
22
- )
23
  Settings.llm = llm
24
 
25
- documents = SimpleDirectoryReader("/root/llamaindex_demo/data").load_data()
26
  index = VectorStoreIndex.from_documents(documents)
27
  query_engine = index.as_query_engine()
28
 
@@ -69,4 +81,4 @@ if st.session_state.messages[-1]["role"] != "assistant":
69
  placeholder = st.empty()
70
  placeholder.markdown(response)
71
  message = {"role": "assistant", "content": response}
72
- st.session_state.messages.append(message)
 
1
  import streamlit as st
2
  from llama_index.core import VectorStoreIndex, SimpleDirectoryReader, Settings
3
  from llama_index.embeddings.huggingface import HuggingFaceEmbedding
4
+ from llama_index.legacy.callbacks import CallbackManager
5
+ from llama_index.llms.openai_like import OpenAILike
6
+ import os
7
+
8
+ # Create an instance of CallbackManager
9
+ callback_manager = CallbackManager()
10
+
11
+ api_base_url = "https://internlm-chat.intern-ai.org.cn/puyu/api/v1/"
12
+ model = "internlm2.5-latest"
13
+ api_key = os.getenv("API_KEY")
14
+
15
+ # api_base_url = "https://api.siliconflow.cn/v1"
16
+ # model = "internlm/internlm2_5-7b-chat"
17
+ # api_key = "请呫写 API Key"
18
+
19
+ llm =OpenAILike(model=model, api_base=api_base_url, api_key=api_key, is_chat_model=True,callback_manager=callback_manager)
20
+
21
+
22
 
23
  st.set_page_config(page_title="llama_index_demo", page_icon="πŸ¦œπŸ”—")
24
  st.title("llama_index_demo")
 
27
  @st.cache_resource
28
  def init_models():
29
  embed_model = HuggingFaceEmbedding(
30
+ model_name="sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2"
31
  )
32
  Settings.embed_model = embed_model
33
 
34
+ #η”¨εˆε§‹εŒ–llm
 
 
 
 
 
35
  Settings.llm = llm
36
 
37
+ documents = SimpleDirectoryReader("./data").load_data()
38
  index = VectorStoreIndex.from_documents(documents)
39
  query_engine = index.as_query_engine()
40
 
 
81
  placeholder = st.empty()
82
  placeholder.markdown(response)
83
  message = {"role": "assistant", "content": response}
84
+ st.session_state.messages.append(message)