Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -28,8 +28,24 @@ with st.sidebar:
|
|
| 28 |
|
| 29 |
def main() -> None:
|
| 30 |
st.header("X-Pipe Wiki 机器人 💬")
|
| 31 |
-
robot_manager = XPipeRobotManagerFactory.get_or_create(XPipeRobotRevision.SIMPLE_OPENAI_VERSION_0)
|
| 32 |
-
robot = robot_manager.get_robot()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
query = st.text_input("X-Pipe Wiki 问题:")
|
| 34 |
if query:
|
| 35 |
response = robot.ask(question=query)
|
|
|
|
| 28 |
|
| 29 |
def main() -> None:
|
| 30 |
st.header("X-Pipe Wiki 机器人 💬")
|
| 31 |
+
# robot_manager = XPipeRobotManagerFactory.get_or_create(XPipeRobotRevision.SIMPLE_OPENAI_VERSION_0)
|
| 32 |
+
# robot = robot_manager.get_robot()
|
| 33 |
+
from llama_index import StorageContext
|
| 34 |
+
store = StorageContext.from_defaults(persist_dir="./dataset")
|
| 35 |
+
from langchain_manager.manager import LangChainAzureManager
|
| 36 |
+
lc_manager = LangChainAzureManager()
|
| 37 |
+
from llama_index import ServiceContext
|
| 38 |
+
from llama_index import LangchainEmbedding
|
| 39 |
+
service = ServiceContext.from_defaults(llm=lc_manager.get_llm(), embed_model=LangchainEmbedding(lc_manager.get_embedding()))
|
| 40 |
+
|
| 41 |
+
from llama_index import load_index_from_storage
|
| 42 |
+
index = load_index_from_storage(store)
|
| 43 |
+
query_engine = index.as_query_engine(service_context=service)
|
| 44 |
+
robot = AzureOpenAIXPipeWikiRobot(query_engine)
|
| 45 |
+
query = st.text_input("X-Pipe Wiki 问题:")
|
| 46 |
+
if query:
|
| 47 |
+
response = robot.ask(question=query)
|
| 48 |
+
st.write(response)
|
| 49 |
query = st.text_input("X-Pipe Wiki 问题:")
|
| 50 |
if query:
|
| 51 |
response = robot.ask(question=query)
|