HenryY2023 commited on
Commit
faab237
·
verified ·
1 Parent(s): f2d9373

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +41 -6
app.py CHANGED
@@ -8,10 +8,14 @@ from langchain.text_splitter import RecursiveCharacterTextSplitter, SemanticChun
8
  from langchain.embeddings import HuggingFaceEmbeddings
9
  from langchain.vectorstores import FAISS
10
  from langchain.chains import ConversationalRetrievalChain
11
- from langchain.llms import ZhipuAI
12
- from langchain.prompts import PromptTemplate
13
- from langchain.memory import ConversationBufferMemory
14
  from langchain.schema import HumanMessage, AIMessage
 
 
 
 
 
 
 
15
 
16
  # Configuration from environment variables
17
  ZHIPU_MODEL = os.environ.get("ZHIPU_MODEL", "chatglm3-6b")
@@ -21,6 +25,37 @@ CHUNK_OVERLAP = int(os.environ.get("CHUNK_OVERLAP", "200"))
21
  USE_SEMANTIC_CHUNKING = os.environ.get("USE_SEMANTIC_CHUNKING", "true").lower() == "true"
22
  MEMORY_K = int(os.environ.get("MEMORY_K", "5")) # Number of messages to remember
23
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  # Database setup
25
  DB_PATH = "/data/chat_history.db"
26
 
@@ -118,10 +153,10 @@ def initialize_system(pdf_path):
118
  vector_store = FAISS.from_texts(texts, embeddings)
119
  retriever = vector_store.as_retriever()
120
 
121
- # Initialize Zhipu LLM
122
- llm = ZhipuAI(
123
- model=ZHIPU_MODEL,
124
  api_key=os.environ["ZHIPU_API_KEY"],
 
125
  temperature=ZHIPU_TEMPERATURE
126
  )
127
 
 
8
  from langchain.embeddings import HuggingFaceEmbeddings
9
  from langchain.vectorstores import FAISS
10
  from langchain.chains import ConversationalRetrievalChain
 
 
 
11
  from langchain.schema import HumanMessage, AIMessage
12
+ from langchain.memory import ConversationBufferMemory
13
+ from langchain.prompts import PromptTemplate
14
+ from langchain.llms.base import LLM
15
+ from typing import Optional, List, Dict, Any
16
+
17
+ # Import the new Zhipu AI SDK
18
+ from zhipuai import ZhipuAI
19
 
20
  # Configuration from environment variables
21
  ZHIPU_MODEL = os.environ.get("ZHIPU_MODEL", "chatglm3-6b")
 
25
  USE_SEMANTIC_CHUNKING = os.environ.get("USE_SEMANTIC_CHUNKING", "true").lower() == "true"
26
  MEMORY_K = int(os.environ.get("MEMORY_K", "5")) # Number of messages to remember
27
 
28
+ # Custom LLM wrapper for the new Zhipu AI SDK
29
+ class ZhipuAILLM(LLM):
30
+ client: ZhipuAI
31
+ model: str = "chatglm3-6b"
32
+ temperature: float = 0.1
33
+
34
+ def __init__(self, api_key: str, **kwargs):
35
+ super().__init__(**kwargs)
36
+ self.client = ZhipuAI(api_key=api_key)
37
+ self.model = kwargs.get("model", self.model)
38
+ self.temperature = kwargs.get("temperature", self.temperature)
39
+
40
+ @property
41
+ def _llm_type(self) -> str:
42
+ return "zhipuai"
43
+
44
+ def _call(
45
+ self,
46
+ prompt: str,
47
+ stop: Optional[List[str]] = None,
48
+ run_manager: Optional[Any] = None,
49
+ **kwargs: Any,
50
+ ) -> str:
51
+ response = self.client.chat.completions.create(
52
+ model=self.model,
53
+ messages=[{"role": "user", "content": prompt}],
54
+ temperature=self.temperature,
55
+ **kwargs
56
+ )
57
+ return response.choices[0].message.content
58
+
59
  # Database setup
60
  DB_PATH = "/data/chat_history.db"
61
 
 
153
  vector_store = FAISS.from_texts(texts, embeddings)
154
  retriever = vector_store.as_retriever()
155
 
156
+ # Initialize Zhipu LLM using our custom wrapper
157
+ llm = ZhipuAILLM(
 
158
  api_key=os.environ["ZHIPU_API_KEY"],
159
+ model=ZHIPU_MODEL,
160
  temperature=ZHIPU_TEMPERATURE
161
  )
162