Spaces:
Sleeping
Sleeping
| import os | |
| import logging | |
| from typing import Optional, List | |
| from langchain_community.document_loaders import DirectoryLoader, TextLoader | |
| from langchain_community.vectorstores import FAISS | |
| from langchain_text_splitters import RecursiveCharacterTextSplitter | |
| logger = logging.getLogger(__name__) | |
| _embeddings_model = None | |
| _rag_handler_instance = None | |
| VECTOR_STORE_PATH = "/tmp/vector_store" | |
| def get_embeddings_model(): | |
| global _embeddings_model | |
| if _embeddings_model is None: | |
| from langchain_huggingface import HuggingFaceEmbeddings | |
| logger.info("Initialisation du modèle d'embeddings...") | |
| _embeddings_model = HuggingFaceEmbeddings( | |
| model_name='sentence-transformers/all-MiniLM-L6-v2', | |
| model_kwargs={'device': 'cpu'}, | |
| encode_kwargs={'normalize_embeddings': True} | |
| ) | |
| logger.info("✅ Modèle d'embeddings initialisé avec succès") | |
| return _embeddings_model | |
| class RAGHandler: | |
| def __init__(self, knowledge_base_path: str = "/app/knowledge_base", lazy_init: bool = True): | |
| self.knowledge_base_path = knowledge_base_path | |
| self.embeddings = None | |
| self.vector_store = None | |
| self._initialized = False | |
| os.makedirs(VECTOR_STORE_PATH, exist_ok=True) | |
| if not lazy_init: | |
| self._initialize() | |
| def _initialize(self): | |
| if self._initialized: | |
| return | |
| logger.info("Initialisation du RAG Handler...") | |
| self.embeddings = get_embeddings_model() | |
| if self.embeddings is None: | |
| logger.error("Impossible d'initialiser les embeddings") | |
| return | |
| self.vector_store = self._load_or_create_vector_store(self.knowledge_base_path) | |
| self._initialized = True | |
| logger.info("✅ RAG Handler initialisé avec succès") | |
| def _load_documents(self, path: str) -> List: | |
| if not os.path.exists(path): | |
| logger.warning(f"Répertoire {path} non trouvé") | |
| return [] | |
| loader = DirectoryLoader( | |
| path, | |
| glob="**/*.md", | |
| loader_cls=TextLoader, | |
| loader_kwargs={"encoding": "utf-8"} | |
| ) | |
| logger.info(f"Chargement des documents depuis : {path}") | |
| documents = loader.load() | |
| logger.info(f"✅ {len(documents)} documents chargés") | |
| return documents | |
| def _create_vector_store(self, knowledge_base_path: str) -> Optional[FAISS]: | |
| documents = self._load_documents(knowledge_base_path) | |
| if not documents: | |
| logger.warning("Aucun document trouvé - création d'un vector store vide") | |
| from langchain.schema import Document | |
| dummy_doc = Document( | |
| page_content="Document de test pour initialiser le vector store", | |
| metadata={"source": "dummy"} | |
| ) | |
| documents = [dummy_doc] | |
| logger.info(f"{len(documents)} documents chargés. Création des vecteurs...") | |
| text_splitter = RecursiveCharacterTextSplitter(chunk_size=1000, chunk_overlap=100) | |
| texts = text_splitter.split_documents(documents) | |
| vector_store = FAISS.from_documents(texts, self.embeddings) | |
| vector_store.save_local(VECTOR_STORE_PATH) | |
| logger.info(f"✅ Vector store créé et sauvegardé dans : {VECTOR_STORE_PATH}") | |
| return vector_store | |
| def _load_or_create_vector_store(self, knowledge_base_path: str) -> Optional[FAISS]: | |
| index_path = os.path.join(VECTOR_STORE_PATH, "index.faiss") | |
| if os.path.exists(index_path): | |
| logger.info(f"Chargement du vector store existant depuis : {VECTOR_STORE_PATH}") | |
| return FAISS.load_local( | |
| VECTOR_STORE_PATH, | |
| embeddings=self.embeddings, | |
| allow_dangerous_deserialization=True | |
| ) | |
| else: | |
| logger.info("Aucun vector store trouvé. Création d'un nouveau...") | |
| return self._create_vector_store(knowledge_base_path) | |
| def get_relevant_feedback(self, query: str, k: int = 1) -> List[str]: | |
| if not self._initialized: | |
| self._initialize() | |
| if not self.vector_store: | |
| logger.warning("Vector store non disponible - retour de conseils génériques") | |
| return [ | |
| "Préparez vos réponses aux questions comportementales", | |
| "Montrez votre motivation pour le poste", | |
| "Donnez des exemples concrets de vos réalisations" | |
| ] | |
| results = self.vector_store.similarity_search(query, k=k) | |
| feedback = [doc.page_content for doc in results if doc.page_content.strip()] | |
| if not feedback: | |
| return ["Conseil général: Préparez-vous bien pour les entretiens futurs."] | |
| return feedback | |
| def get_rag_handler() -> Optional[RAGHandler]: | |
| global _rag_handler_instance | |
| if _rag_handler_instance is None: | |
| _rag_handler_instance = RAGHandler(lazy_init=True) | |
| return _rag_handler_instance |