Spaces:
Sleeping
Sleeping
Update learning_platform.py
Browse files- learning_platform.py +7 -7
learning_platform.py
CHANGED
|
@@ -5,11 +5,11 @@ from datetime import datetime
|
|
| 5 |
import json
|
| 6 |
import logging
|
| 7 |
from langchain.chat_models import ChatOpenAI
|
| 8 |
-
from
|
| 9 |
-
from
|
| 10 |
from langchain.memory import ConversationBufferMemory
|
| 11 |
from langchain.prompts import ChatPromptTemplate, MessagesPlaceholder
|
| 12 |
-
from langchain.chains import
|
| 13 |
from prompts import CoursePromptTemplates
|
| 14 |
from models import *
|
| 15 |
from sqlalchemy.orm import Session
|
|
@@ -63,7 +63,7 @@ class EnhancedCourseBuilder:
|
|
| 63 |
MessagesPlaceholder("chat_history"),
|
| 64 |
["human", "{input}"]
|
| 65 |
])
|
| 66 |
-
self.history_aware_retriever =
|
| 67 |
llm=self.llm,
|
| 68 |
retriever=self.vector_store.as_retriever(),
|
| 69 |
rephrase_prompt=contextualize_q_prompt
|
|
@@ -82,15 +82,15 @@ class EnhancedCourseBuilder:
|
|
| 82 |
])
|
| 83 |
|
| 84 |
# Create the question-answer chain
|
| 85 |
-
self.question_answer_chain =
|
| 86 |
llm=self.llm,
|
| 87 |
prompt=qa_system_prompt
|
| 88 |
)
|
| 89 |
|
| 90 |
# Create the retrieval chain
|
| 91 |
-
self.rag_chain =
|
| 92 |
retriever=self.history_aware_retriever,
|
| 93 |
-
|
| 94 |
)
|
| 95 |
|
| 96 |
async def create_course(self, topic: str, difficulty: str, user_id: int) -> Course:
|
|
|
|
| 5 |
import json
|
| 6 |
import logging
|
| 7 |
from langchain.chat_models import ChatOpenAI
|
| 8 |
+
from langchain_community.embeddings import OpenAIEmbeddings
|
| 9 |
+
from langchain_community.vectorstores import FAISS
|
| 10 |
from langchain.memory import ConversationBufferMemory
|
| 11 |
from langchain.prompts import ChatPromptTemplate, MessagesPlaceholder
|
| 12 |
+
from langchain.chains import create_stuff_documents_chain, create_history_aware_retriever, create_retrieval_chain
|
| 13 |
from prompts import CoursePromptTemplates
|
| 14 |
from models import *
|
| 15 |
from sqlalchemy.orm import Session
|
|
|
|
| 63 |
MessagesPlaceholder("chat_history"),
|
| 64 |
["human", "{input}"]
|
| 65 |
])
|
| 66 |
+
self.history_aware_retriever = create_history_aware_retriever(
|
| 67 |
llm=self.llm,
|
| 68 |
retriever=self.vector_store.as_retriever(),
|
| 69 |
rephrase_prompt=contextualize_q_prompt
|
|
|
|
| 82 |
])
|
| 83 |
|
| 84 |
# Create the question-answer chain
|
| 85 |
+
self.question_answer_chain = create_stuff_documents_chain(
|
| 86 |
llm=self.llm,
|
| 87 |
prompt=qa_system_prompt
|
| 88 |
)
|
| 89 |
|
| 90 |
# Create the retrieval chain
|
| 91 |
+
self.rag_chain = create_retrieval_chain(
|
| 92 |
retriever=self.history_aware_retriever,
|
| 93 |
+
combine_docs_chain=self.question_answer_chain
|
| 94 |
)
|
| 95 |
|
| 96 |
async def create_course(self, topic: str, difficulty: str, user_id: int) -> Course:
|