File size: 4,670 Bytes
f17525a 65562e0 9f3cdd5 f17525a afd7f8c f17525a 9f3cdd5 e73a9c6 9f3cdd5 65562e0 f17525a 65562e0 f17525a 65562e0 f17525a 9f3cdd5 65562e0 f17525a 65562e0 f17525a 65562e0 9f3cdd5 65562e0 afd7f8c f17525a 65562e0 f17525a 65562e0 f17525a 65562e0 f17525a 65562e0 f17525a 65562e0 afd7f8c f17525a afd7f8c f17525a 65562e0 afd7f8c 65562e0 afd7f8c 65562e0 afd7f8c 65562e0 afd7f8c 65562e0 afd7f8c f17525a afd7f8c f17525a 65562e0 afd7f8c 65562e0 afd7f8c 65562e0 afd7f8c f17525a 65562e0 f17525a afd7f8c 65562e0 afd7f8c 65562e0 afd7f8c 65562e0 afd7f8c 65562e0 afd7f8c 65562e0 afd7f8c 65562e0 f17525a afd7f8c 65562e0 f17525a 65562e0 afd7f8c 65562e0 f17525a 65562e0 f17525a 65562e0 afd7f8c f17525a 65562e0 f17525a afd7f8c | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 | from __future__ import annotations
import os
import re
import textwrap
import subprocess
import sys
from pathlib import Path
import faiss
import numpy as np
import requests
import spacy
from bs4 import BeautifulSoup
from huggingface_hub import InferenceClient
from pypdf import PdfReader
from sentence_transformers import SentenceTransformer
# ββ Config βββββββββββββββββββββββββββββββββββββββββββββ
EMBED_MODEL = "sentence-transformers/all-MiniLM-L6-v2"
LLM_MODEL = "HuggingFaceH4/zephyr-7b-beta"
CHUNK_SIZE = 400
CHUNK_OVERLAP = 80
TOP_K = 4
# ββ Engine βββββββββββββββββββββββββββββββββββββββββββββ
class RAGEngine:
def __init__(self):
print("Loading embedding model...")
self.embedder = SentenceTransformer(EMBED_MODEL)
self.hf_client = InferenceClient(token=os.getenv("HF_TOKEN"))
self._load_spacy()
self.reset()
def _load_spacy(self):
try:
self.nlp = spacy.load("en_core_web_sm")
except:
subprocess.run(
[sys.executable, "-m", "spacy", "download", "en_core_web_sm"],
check=True,
)
self.nlp = spacy.load("en_core_web_sm")
def reset(self):
self.chunks = []
self.index = None
@property
def ready(self):
return self.index is not None and len(self.chunks) > 0
# ββ Loaders βββββββββββββββββββββββββββββββββββββ
def load_pdf(self, path):
reader = PdfReader(path)
text = " ".join(p.extract_text() or "" for p in reader.pages)
if not text.strip():
raise ValueError("No text found in PDF")
self._build_index(text)
return f"β
PDF loaded ({len(self.chunks)} chunks)"
def load_url(self, url):
r = requests.get(url, timeout=15, headers={"User-Agent": "Mozilla/5.0"})
r.raise_for_status()
soup = BeautifulSoup(r.text, "html.parser")
for tag in soup(["script", "style"]):
tag.decompose()
text = soup.get_text(" ", strip=True)
if not text.strip():
raise ValueError("No text found in URL")
self._build_index(text)
return f"β
URL loaded ({len(self.chunks)} chunks)"
def load_text(self, text):
if not text.strip():
raise ValueError("Empty text")
self._build_index(text)
return f"β
Text loaded ({len(self.chunks)} chunks)"
# ββ Chunking βββββββββββββββββββββββββββββββββββββ
def _chunk(self, text):
text = re.sub(r"\s+", " ", text)
chunks, i = [], 0
while i < len(text):
chunks.append(text[i:i + CHUNK_SIZE])
i += CHUNK_SIZE - CHUNK_OVERLAP
return [c for c in chunks if len(c.strip()) > 30]
# ββ Indexing βββββββββββββββββββββββββββββββββββββ
def _build_index(self, text):
self.chunks = self._chunk(text)
emb = self.embedder.encode(self.chunks, show_progress_bar=False)
emb = np.array(emb).astype("float32")
faiss.normalize_L2(emb)
self.index = faiss.IndexFlatIP(emb.shape[1])
self.index.add(emb)
# ββ Retrieval βββββββββββββββββββββββββββββββββββββ
def _retrieve(self, query):
emb = self.embedder.encode([query], show_progress_bar=False)
emb = np.array(emb).astype("float32")
faiss.normalize_L2(emb)
_, idx = self.index.search(emb, TOP_K)
return [self.chunks[i] for i in idx[0] if i < len(self.chunks)]
# ββ Answer βββββββββββββββββββββββββββββββββββββββ
def answer(self, query):
if not self.ready:
return "β οΈ Please load data first."
chunks = self._retrieve(query)
prompt = f"""
Use ONLY this context to answer:
{chunks}
Question: {query}
"""
try:
res = self.hf_client.text_generation(
prompt,
model=LLM_MODEL,
max_new_tokens=300,
temperature=0.3,
)
return res.strip()
except Exception as e:
return f"β οΈ API Error: {e}\n\n{chunks[0]}" |