Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -69,7 +69,8 @@ def get_chain(temperature,selected_model):
|
|
| 69 |
child_splitter = RecursiveCharacterTextSplitter(chunk_size=300,
|
| 70 |
chunk_overlap=50)
|
| 71 |
retriever = load_retriever(docstore_path,chroma_path,embeddings,child_splitter,parent_splitter)
|
| 72 |
-
llm_api = 'glpat-AMzMevbqaVjp4HbLcVum'
|
|
|
|
| 73 |
llm = ChatOpenAI(model_name=selected_model,
|
| 74 |
temperature=temperature,
|
| 75 |
openai_api_key=llm_api,
|
|
@@ -88,7 +89,7 @@ def get_chain(temperature,selected_model):
|
|
| 88 |
|
| 89 |
# Original Question: {question}
|
| 90 |
|
| 91 |
-
# Answer
|
| 92 |
|
| 93 |
"""
|
| 94 |
response_prompt = ChatPromptTemplate.from_template(response_prompt_template)
|
|
|
|
| 69 |
child_splitter = RecursiveCharacterTextSplitter(chunk_size=300,
|
| 70 |
chunk_overlap=50)
|
| 71 |
retriever = load_retriever(docstore_path,chroma_path,embeddings,child_splitter,parent_splitter)
|
| 72 |
+
# llm_api = 'glpat-AMzMevbqaVjp4HbLcVum'
|
| 73 |
+
llm_api = os.getenv("blablador_api")
|
| 74 |
llm = ChatOpenAI(model_name=selected_model,
|
| 75 |
temperature=temperature,
|
| 76 |
openai_api_key=llm_api,
|
|
|
|
| 89 |
|
| 90 |
# Original Question: {question}
|
| 91 |
|
| 92 |
+
# Answer:
|
| 93 |
|
| 94 |
"""
|
| 95 |
response_prompt = ChatPromptTemplate.from_template(response_prompt_template)
|