|
|
""" |
|
|
Genomic Semantic Search API with FastAPI |
|
|
========================================= |
|
|
Search genomic sequences using your pre-trained transformer embeddings. |
|
|
""" |
|
|
|
|
|
import pickle |
|
|
from pathlib import Path |
|
|
from typing import Optional |
|
|
import numpy as np |
|
|
import pandas as pd |
|
|
from fastapi import FastAPI, HTTPException |
|
|
from fastapi.staticfiles import StaticFiles |
|
|
from fastapi.responses import HTMLResponse, FileResponse |
|
|
from fastapi.middleware.cors import CORSMiddleware |
|
|
from pydantic import BaseModel |
|
|
import faiss |
|
|
import torch |
|
|
import torch.nn as nn |
|
|
from x_transformers import TransformerWrapper, Encoder |
|
|
import tiktoken |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
DATA_DIR = Path("data") |
|
|
INDEX_PATH = DATA_DIR / "data/faiss.index" |
|
|
METADATA_PATH = DATA_DIR / "data/metadata.pkl" |
|
|
EMBEDDINGS_PATH = DATA_DIR / "data/embeddings.npy" |
|
|
|
|
|
|
|
|
MODEL_WEIGHTS_PATH = DATA_DIR / "data/bpe_plus_special_tokens_model.pt" |
|
|
TOKENIZER_PATH = DATA_DIR / "data/bpe_plus_special_tokens_tokenizer.json" |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class GenomicTransformer(nn.Module): |
|
|
def __init__(self, vocab_size=40000, hidden_dim=512, layers=12, heads=8, max_length=6000): |
|
|
super().__init__() |
|
|
self.model = TransformerWrapper( |
|
|
num_tokens=vocab_size, |
|
|
max_seq_len=max_length, |
|
|
attn_layers=Encoder( |
|
|
dim=hidden_dim, |
|
|
depth=layers, |
|
|
heads=heads, |
|
|
rotary_pos_emb=True, |
|
|
attn_orthog_projected_values=True, |
|
|
attn_orthog_projected_values_per_head=True, |
|
|
attn_flash=True |
|
|
) |
|
|
) |
|
|
|
|
|
def forward(self, input_ids, return_embeddings=False): |
|
|
return self.model(input_ids, return_embeddings=return_embeddings) |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
app = FastAPI( |
|
|
title="Genomic Semantic Search", |
|
|
description="Search genomic sequences using transformer embeddings", |
|
|
version="1.0.0" |
|
|
) |
|
|
|
|
|
app.add_middleware( |
|
|
CORSMiddleware, |
|
|
allow_origins=["*"], |
|
|
allow_credentials=True, |
|
|
allow_methods=["*"], |
|
|
allow_headers=["*"], |
|
|
) |
|
|
|
|
|
|
|
|
device: torch.device = None |
|
|
model: Optional[GenomicTransformer] = None |
|
|
encoder: Optional[tiktoken.Encoding] = None |
|
|
index: Optional[faiss.IndexFlatIP] = None |
|
|
metadata: Optional[pd.DataFrame] = None |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
class SearchRequest(BaseModel): |
|
|
query: str |
|
|
top_k: int = 10 |
|
|
|
|
|
class SearchResult(BaseModel): |
|
|
rank: int |
|
|
score: float |
|
|
sequence: str |
|
|
metadata: dict |
|
|
|
|
|
class SearchResponse(BaseModel): |
|
|
query: str |
|
|
results: list[SearchResult] |
|
|
total_indexed: int |
|
|
|
|
|
class IndexStats(BaseModel): |
|
|
total_documents: int |
|
|
embedding_dimension: int |
|
|
model_name: str |
|
|
device: str |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@app.on_event("startup") |
|
|
async def startup(): |
|
|
"""Load the model, tokenizer, and FAISS index on startup.""" |
|
|
global device, model, encoder, index, metadata |
|
|
|
|
|
|
|
|
device = torch.device("cuda" if torch.cuda.is_available() else "cpu") |
|
|
print(f"Using device: {device}") |
|
|
|
|
|
|
|
|
print("Loading tokenizer...") |
|
|
if TOKENIZER_PATH.exists(): |
|
|
with open(TOKENIZER_PATH, "rb") as f: |
|
|
tokenizer_data = pickle.load(f) |
|
|
encoder = tiktoken.Encoding( |
|
|
name="genomic_bpe", |
|
|
pat_str=tokenizer_data['pattern'], |
|
|
mergeable_ranks=tokenizer_data['mergable_ranks'], |
|
|
special_tokens={} |
|
|
) |
|
|
print("Tokenizer loaded successfully") |
|
|
else: |
|
|
print(f"WARNING: Tokenizer not found at {TOKENIZER_PATH}") |
|
|
|
|
|
|
|
|
print("Loading model...") |
|
|
if MODEL_WEIGHTS_PATH.exists(): |
|
|
model = GenomicTransformer( |
|
|
vocab_size=40_000, hidden_dim=512, layers=12, heads=8 |
|
|
) |
|
|
weights = torch.load(MODEL_WEIGHTS_PATH, map_location=device) |
|
|
model.load_state_dict(weights) |
|
|
model = model.to(device) |
|
|
model.eval() |
|
|
print("Model loaded successfully") |
|
|
else: |
|
|
print(f"WARNING: Model weights not found at {MODEL_WEIGHTS_PATH}") |
|
|
|
|
|
|
|
|
if INDEX_PATH.exists() and METADATA_PATH.exists(): |
|
|
print("Loading FAISS index...") |
|
|
index = faiss.read_index(str(INDEX_PATH)) |
|
|
with open(METADATA_PATH, "rb") as f: |
|
|
metadata = pickle.load(f) |
|
|
print(f"Index loaded with {index.ntotal} documents") |
|
|
else: |
|
|
print(f"WARNING: Index not found at {INDEX_PATH}") |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
@app.get("/", response_class=HTMLResponse) |
|
|
async def root(): |
|
|
"""Serve the search frontend.""" |
|
|
return FileResponse("index.html") |
|
|
|
|
|
@app.get("/api/health") |
|
|
async def health(): |
|
|
"""Health check endpoint.""" |
|
|
return { |
|
|
"status": "healthy", |
|
|
"model_loaded": model is not None, |
|
|
"index_loaded": index is not None, |
|
|
"tokenizer_loaded": encoder is not None, |
|
|
"device": str(device) |
|
|
} |
|
|
|
|
|
@app.get("/api/stats", response_model=IndexStats) |
|
|
async def get_stats(): |
|
|
"""Get statistics about the current index.""" |
|
|
if index is None: |
|
|
raise HTTPException(status_code=404, detail="No index loaded") |
|
|
|
|
|
return IndexStats( |
|
|
total_documents=index.ntotal, |
|
|
embedding_dimension=index.d, |
|
|
model_name="GenomicTransformer (512d, 12 layers)", |
|
|
device=str(device) |
|
|
) |
|
|
|
|
|
@app.post("/api/search", response_model=SearchResponse) |
|
|
async def search(request: SearchRequest): |
|
|
""" |
|
|
Perform semantic search over genomic sequences. |
|
|
|
|
|
- **query**: The genomic sequence to search for (e.g., "ATCGATCG...") |
|
|
- **top_k**: Number of results to return (default: 10) |
|
|
""" |
|
|
if index is None or metadata is None: |
|
|
raise HTTPException(status_code=404, detail="No index loaded") |
|
|
if model is None or encoder is None: |
|
|
raise HTTPException(status_code=503, detail="Model or tokenizer not loaded") |
|
|
if index.ntotal == 0: |
|
|
raise HTTPException(status_code=404, detail="Index is empty") |
|
|
|
|
|
|
|
|
try: |
|
|
encodings = encoder.encode_ordinary(request.query) |
|
|
query_tensor = torch.tensor([encodings]).long().to(device) |
|
|
|
|
|
with torch.no_grad(): |
|
|
query_embedding = model(query_tensor, return_embeddings=True) |
|
|
query_embedding = query_embedding.mean(dim=1).cpu().numpy() |
|
|
|
|
|
query_embedding = query_embedding.astype(np.float32) |
|
|
except Exception as e: |
|
|
raise HTTPException(status_code=400, detail=f"Failed to encode query: {str(e)}") |
|
|
|
|
|
|
|
|
k = min(request.top_k, index.ntotal) |
|
|
scores, indices = index.search(query_embedding, k) |
|
|
|
|
|
|
|
|
results = [] |
|
|
for rank, (score, idx) in enumerate(zip(scores[0], indices[0]), 1): |
|
|
if idx == -1: |
|
|
continue |
|
|
|
|
|
row = metadata.iloc[idx] |
|
|
meta_dict = row.to_dict() |
|
|
sequence = meta_dict.pop("__sequence__", "") |
|
|
|
|
|
results.append(SearchResult( |
|
|
rank=rank, |
|
|
score=float(score), |
|
|
sequence=sequence, |
|
|
metadata=meta_dict |
|
|
)) |
|
|
|
|
|
return SearchResponse( |
|
|
query=request.query[:100] + "..." if len(request.query) > 100 else request.query, |
|
|
results=results, |
|
|
total_indexed=index.ntotal |
|
|
) |
|
|
|
|
|
@app.get("/api/sample") |
|
|
async def get_sample(n: int = 5): |
|
|
"""Get a sample of indexed documents.""" |
|
|
if metadata is None: |
|
|
raise HTTPException(status_code=404, detail="No index loaded") |
|
|
|
|
|
sample = metadata.head(n) |
|
|
return { |
|
|
"total": len(metadata), |
|
|
"sample": sample.to_dict(orient="records") |
|
|
} |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
if __name__ == "__main__": |
|
|
import uvicorn |
|
|
uvicorn.run(app, host="0.0.0.0", port=8080) |