| |
| """ |
| embedding_backend.py -- Pluggable embedding backends for the intake gate. |
| |
| Two backends ship in Phase 2: |
| |
| - SentenceTransformerEmbedder default, zero external daemon. |
| Uses sentence-transformers/all-MiniLM-L6-v2. |
| - OllamaEmbedder opt-in, requires a local ollama daemon with |
| `nomic-embed-text` pulled. |
| |
| The ``Embedder`` Protocol lets the rest of the intake gate stay |
| backend-agnostic. Both implementations return L2-normalised float32 vectors |
| so downstream cosine similarity is a single dot product. |
| |
| Backend selection is centralised in ``get_embedder(name)``; callers pass the |
| string from ``ctx_config.intake.embedding.backend``. Heavy imports |
| (``sentence_transformers``, ``requests``) happen lazily inside the concrete |
| class to keep the module cheap to import. |
| |
| Security: |
| |
| ``OllamaEmbedder`` only talks to a locally-bound host (localhost / |
| 127.0.0.1 / ::1) to prevent SSRF via a poisoned config pointing the |
| embedder at internal metadata endpoints. Non-http(s) schemes are |
| rejected. To reach a non-local ollama deployment, callers must opt in |
| explicitly by passing ``allow_remote=True`` — this is a deliberate |
| friction surface. |
| """ |
|
|
| from __future__ import annotations |
|
|
| import os |
| from dataclasses import dataclass, field |
| from typing import Any, Protocol, Sequence, runtime_checkable |
| from urllib.parse import urlparse |
|
|
| import numpy as np |
|
|
|
|
| DEFAULT_ST_MODEL = "sentence-transformers/all-MiniLM-L6-v2" |
| DEFAULT_OLLAMA_MODEL = "nomic-embed-text" |
| DEFAULT_OLLAMA_URL = "http://localhost:11434" |
| _ST_ENCODE_BATCH_SIZE = 512 |
|
|
| |
| |
| _NOMIC_EMBED_TEXT_DIM = 768 |
| _LOCAL_HOSTS = frozenset({"localhost", "127.0.0.1", "::1"}) |
|
|
|
|
| @runtime_checkable |
| class Embedder(Protocol): |
| """One method: embed a batch of texts to a 2-D float32 matrix. |
| |
| ``dim`` and ``name`` are declared as ``@property`` so implementations are |
| free to compute them lazily — important because models can be expensive |
| to load and callers shouldn't pay that cost just to introspect metadata. |
| """ |
|
|
| @property |
| def dim(self) -> int: ... |
|
|
| @property |
| def name(self) -> str: ... |
|
|
| def embed(self, texts: Sequence[str]) -> np.ndarray: ... |
|
|
|
|
| def _l2_normalize(mat: np.ndarray) -> np.ndarray: |
| """Row-wise L2 normalisation with zero-vector safety.""" |
| norms = np.linalg.norm(mat, axis=1, keepdims=True) |
| norms = np.where(norms == 0.0, 1.0, norms) |
| return (mat / norms).astype(np.float32, copy=False) |
|
|
|
|
| @dataclass |
| class SentenceTransformerEmbedder: |
| """Local-only embedder; loads the model on first ``embed`` call.""" |
|
|
| model_name: str = DEFAULT_ST_MODEL |
| |
| |
| _model: Any = field(init=False, default=None, repr=False, compare=False) |
|
|
| @property |
| def name(self) -> str: |
| return f"sentence-transformers:{self.model_name}" |
|
|
| @property |
| def dim(self) -> int: |
| |
| |
| |
| |
| if self._model is None: |
| return -1 |
| return int(self._model.get_sentence_embedding_dimension()) |
|
|
| def _ensure_loaded(self) -> None: |
| if self._model is not None: |
| return |
| try: |
| from sentence_transformers import SentenceTransformer |
| except ImportError as exc: |
| raise RuntimeError( |
| "sentence-transformers is not installed; " |
| "pip install sentence-transformers or switch intake.embedding.backend " |
| "to 'ollama'" |
| ) from exc |
| self._model = SentenceTransformer(self.model_name) |
|
|
| def embed(self, texts: Sequence[str]) -> np.ndarray: |
| if not texts: |
| return np.zeros((0, 0), dtype=np.float32) |
| self._ensure_loaded() |
| vecs = self._model.encode( |
| list(texts), |
| batch_size=min(len(texts), _ST_ENCODE_BATCH_SIZE), |
| convert_to_numpy=True, |
| normalize_embeddings=False, |
| show_progress_bar=False, |
| ) |
| return _l2_normalize(np.asarray(vecs, dtype=np.float32)) |
|
|
|
|
| class OllamaEmbedderError(RuntimeError): |
| """Raised when an Ollama request fails. Carries the failing text index.""" |
|
|
| def __init__(self, index: int, message: str) -> None: |
| super().__init__(f"[text #{index}] {message}") |
| self.index = index |
|
|
|
|
| @dataclass |
| class OllamaEmbedder: |
| """HTTP-backed embedder for a local ollama daemon. Opt-in. |
| |
| Partial failures fail fast: if text *k* of *N* errors, an |
| ``OllamaEmbedderError`` is raised with ``.index == k`` and no |
| partial batch is returned. Callers who want best-effort behaviour |
| must wrap single-text calls themselves. |
| """ |
|
|
| model_name: str = DEFAULT_OLLAMA_MODEL |
| base_url: str = DEFAULT_OLLAMA_URL |
| timeout: float = 30.0 |
| allow_remote: bool = False |
|
|
| def __post_init__(self) -> None: |
| parsed = urlparse(self.base_url) |
| if parsed.scheme not in ("http", "https"): |
| raise ValueError( |
| f"base_url scheme must be http or https: {self.base_url!r}" |
| ) |
| host = (parsed.hostname or "").lower() |
| if not host: |
| raise ValueError(f"base_url has no host: {self.base_url!r}") |
| if not self.allow_remote and host not in _LOCAL_HOSTS: |
| raise ValueError( |
| f"base_url host {host!r} is not local; pass allow_remote=True " |
| f"to explicitly allow a remote ollama instance" |
| ) |
|
|
| @property |
| def name(self) -> str: |
| return f"ollama:{self.model_name}" |
|
|
| @property |
| def dim(self) -> int: |
| |
| |
| |
| return _NOMIC_EMBED_TEXT_DIM if self.model_name == DEFAULT_OLLAMA_MODEL else -1 |
|
|
| def embed(self, texts: Sequence[str]) -> np.ndarray: |
| if not texts: |
| return np.zeros((0, 0), dtype=np.float32) |
| try: |
| import requests |
| except ImportError as exc: |
| raise RuntimeError( |
| "requests is required for the ollama backend; pip install requests" |
| ) from exc |
|
|
| url = f"{self.base_url.rstrip('/')}/api/embeddings" |
| rows: list[list[float]] = [] |
| for idx, text in enumerate(texts): |
| try: |
| resp = requests.post( |
| url, |
| json={"model": self.model_name, "prompt": text}, |
| timeout=self.timeout, |
| ) |
| resp.raise_for_status() |
| payload = resp.json() |
| except Exception as exc: |
| raise OllamaEmbedderError(idx, str(exc)) from exc |
| if "embedding" not in payload: |
| raise OllamaEmbedderError( |
| idx, f"response missing 'embedding' key: {payload!r}" |
| ) |
| rows.append(payload["embedding"]) |
| return _l2_normalize(np.asarray(rows, dtype=np.float32)) |
|
|
|
|
| def get_embedder( |
| backend: str = "sentence-transformers", |
| *, |
| model: str | None = None, |
| base_url: str | None = None, |
| allow_remote: bool = False, |
| ) -> Embedder: |
| """Factory: map a backend name to a concrete ``Embedder``. |
| |
| The default is ``sentence-transformers`` (no external daemon required). |
| ``ollama`` is opt-in and requires the user to have ollama running with the |
| chosen model pulled. ``allow_remote`` must be set explicitly to reach a |
| non-local ollama host. |
| """ |
| key = (backend or "").strip().lower() |
| if key in ("", "sentence-transformers", "st", "sbert"): |
| return SentenceTransformerEmbedder(model_name=model or DEFAULT_ST_MODEL) |
| if key in ("ollama", "ol"): |
| return OllamaEmbedder( |
| model_name=model or DEFAULT_OLLAMA_MODEL, |
| base_url=base_url or os.environ.get("OLLAMA_URL", DEFAULT_OLLAMA_URL), |
| allow_remote=allow_remote, |
| ) |
| raise ValueError( |
| f"unknown embedding backend {backend!r}; expected " |
| f"'sentence-transformers' or 'ollama'" |
| ) |
|
|