peichao.dong commited on
Commit ·
3772cf0
1
Parent(s): 5c65101
update embedding config
Browse files- .gitignore +1 -0
- documents/abstract.faiss/index.faiss +0 -0
- documents/abstract.faiss/index.pkl +2 -2
- documents/business_context.faiss/index.faiss +0 -0
- documents/business_context.faiss/index.pkl +0 -0
- documents/bussiness_context/business_context.md +3 -0
- embedding.py +1 -1
- promopts.py +1 -0
- requirements.txt +3 -1
.gitignore
CHANGED
|
@@ -1,3 +1,4 @@
|
|
| 1 |
__pycache__/*
|
| 2 |
**/__pycache__/*
|
| 3 |
.DS_Store
|
|
|
|
|
|
| 1 |
__pycache__/*
|
| 2 |
**/__pycache__/*
|
| 3 |
.DS_Store
|
| 4 |
+
.idea/*
|
documents/abstract.faiss/index.faiss
CHANGED
|
Binary files a/documents/abstract.faiss/index.faiss and b/documents/abstract.faiss/index.faiss differ
|
|
|
documents/abstract.faiss/index.pkl
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2b5614785dff26ba4c3cb191b8aef6c88a7e90fc50f0414df5285863e58dd833
|
| 3 |
+
size 58026
|
documents/business_context.faiss/index.faiss
CHANGED
|
Binary files a/documents/business_context.faiss/index.faiss and b/documents/business_context.faiss/index.faiss differ
|
|
|
documents/business_context.faiss/index.pkl
CHANGED
|
Binary files a/documents/business_context.faiss/index.pkl and b/documents/business_context.faiss/index.pkl differ
|
|
|
documents/bussiness_context/business_context.md
CHANGED
|
@@ -14,3 +14,6 @@ FeatureConfig 用于配置某个 Feature 中控制前端展示效果的配置项
|
|
| 14 |
|
| 15 |
用户白名单圈人条件需要上传用户id的白名单,仅在白名单里的用户可以获取到相关feature
|
| 16 |
地理位置配置端需要设置圈定地区的地理位置编号列表,客户端请求接口是传递地理位置编号参数,位置编号匹配的数据用户可见
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
|
| 15 |
用户白名单圈人条件需要上传用户id的白名单,仅在白名单里的用户可以获取到相关feature
|
| 16 |
地理位置配置端需要设置圈定地区的地理位置编号列表,客户端请求接口是传递地理位置编号参数,位置编号匹配的数据用户可见
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
新增实验需要提供实验名称、目标、分组信息(包括分组标识、描述、比例)
|
embedding.py
CHANGED
|
@@ -48,7 +48,7 @@ class CustomEmbedding:
|
|
| 48 |
CONDENSE_QUESTION_PROMPT = PromptTemplate.from_template(_template)
|
| 49 |
question_generator = LLMChain(llm=llm, prompt=CONDENSE_QUESTION_PROMPT)
|
| 50 |
|
| 51 |
-
doc_chain = load_qa_chain(llm, chain_type="
|
| 52 |
qa = ConversationalRetrievalChain( retriever= docsearch.as_retriever(search_kwargs={"k": 1}),
|
| 53 |
question_generator=question_generator,
|
| 54 |
combine_docs_chain=doc_chain,
|
|
|
|
| 48 |
CONDENSE_QUESTION_PROMPT = PromptTemplate.from_template(_template)
|
| 49 |
question_generator = LLMChain(llm=llm, prompt=CONDENSE_QUESTION_PROMPT)
|
| 50 |
|
| 51 |
+
doc_chain = load_qa_chain(llm, chain_type="stuff")
|
| 52 |
qa = ConversationalRetrievalChain( retriever= docsearch.as_retriever(search_kwargs={"k": 1}),
|
| 53 |
question_generator=question_generator,
|
| 54 |
combine_docs_chain=doc_chain,
|
promopts.py
CHANGED
|
@@ -19,6 +19,7 @@ Answer: the answer I responded to the question
|
|
| 19 |
Thought: I know enough to explain the user story
|
| 20 |
Scenarios: List all possible scenarios with concrete example in Given/When/Then style
|
| 21 |
|
|
|
|
| 22 |
|
| 23 |
{history}
|
| 24 |
Answer:{input}"""
|
|
|
|
| 19 |
Thought: I know enough to explain the user story
|
| 20 |
Scenarios: List all possible scenarios with concrete example in Given/When/Then style
|
| 21 |
|
| 22 |
+
Please use Chinese! Begin!
|
| 23 |
|
| 24 |
{history}
|
| 25 |
Answer:{input}"""
|
requirements.txt
CHANGED
|
@@ -6,4 +6,6 @@ sentence_transformers
|
|
| 6 |
gradio_tools
|
| 7 |
faiss-cpu
|
| 8 |
tiktoken
|
| 9 |
-
pygpt4all
|
|
|
|
|
|
|
|
|
| 6 |
gradio_tools
|
| 7 |
faiss-cpu
|
| 8 |
tiktoken
|
| 9 |
+
pygpt4all
|
| 10 |
+
chromadb
|
| 11 |
+
unstructured
|