virus / README.md
Hack90's picture
Upload folder using huggingface_hub
aee08b4 verified
# Semantic Search API
A production-ready semantic search service built with FastAPI. Upload your data (sequences + metadata), create embeddings automatically, and search using natural language queries.
## Features
- **Semantic/Latent Search**: Find similar sequences based on meaning, not just keywords
- **FastAPI Backend**: Modern, fast, async Python web framework
- **FAISS Index**: Efficient similarity search at scale
- **Sentence Transformers**: State-of-the-art embedding models
- **Beautiful UI**: Dark-themed, responsive search interface
- **CSV Upload**: Easy data import via web interface or API
- **Persistent Storage**: Index persists across restarts
## Quick Start
### 1. Install Dependencies
```bash
pip install -r requirements.txt
```
### 2. Run the Server
```bash
python app.py
# or
uvicorn app:app --reload --host 0.0.0.0 --port 8000
```
### 3. Open the UI
Navigate to `http://localhost:8000` in your browser.
### 4. Upload Your Data
- Drag & drop a CSV file or click to browse
- Select the column containing your sequences
- Click "Create Index"
- Start searching!
## Data Format
Your CSV should have at least one column containing the text sequences you want to search. All other columns become searchable metadata.
Example:
```csv
sequence,category,source,date
"Machine learning is transforming industries",tech,blog,2024-01-15
"The quick brown fox jumps over the lazy dog",example,pangram,2024-01-10
"Embeddings capture semantic meaning",ml,paper,2024-01-20
```
## API Endpoints
### Search
```bash
POST /api/search
Content-Type: application/json
{
"query": "artificial intelligence",
"top_k": 10
}
```
### Upload CSV
```bash
POST /api/upload-csv?sequence_column=text
Content-Type: multipart/form-data
file: your_data.csv
```
### Create Index (JSON)
```bash
POST /api/index
Content-Type: application/json
{
"sequence_column": "text",
"data": [
{"text": "Hello world", "category": "greeting"},
{"text": "Machine learning", "category": "tech"}
]
}
```
### Get Stats
```bash
GET /api/stats
```
### Get Sample
```bash
GET /api/sample?n=5
```
### Delete Index
```bash
DELETE /api/index
```
## Programmatic Usage
You can also create indexes directly from Python:
```python
from create_index import create_index_from_dataframe, search_index
import pandas as pd
# Create your dataframe
df = pd.DataFrame({
'sequence': [
'The mitochondria is the powerhouse of the cell',
'DNA stores genetic information',
'Proteins are made of amino acids'
],
'category': ['biology', 'genetics', 'biochemistry'],
'difficulty': ['easy', 'medium', 'medium']
})
# Create the index
create_index_from_dataframe(df, sequence_column='sequence')
# Search
results = search_index("cellular energy production", top_k=3)
for r in results:
print(f"Score: {r['score']:.3f} | {r['sequence'][:50]}...")
```
## Configuration
Edit these values in `app.py` to customize:
```python
# Embedding model (from sentence-transformers)
EMBEDDING_MODEL = "all-MiniLM-L6-v2" # Fast, 384 dimensions
# Alternatives:
# "all-mpnet-base-v2" # Higher quality, 768 dimensions
# "paraphrase-multilingual-MiniLM-L12-v2" # Multilingual support
# "all-MiniLM-L12-v2" # Balanced quality/speed
```
## Project Structure
```
semantic_search/
├── app.py # FastAPI application
├── create_index.py # Programmatic index creation
├── requirements.txt # Python dependencies
├── static/
│ └── index.html # Search UI
├── data/ # Created at runtime
│ ├── faiss.index # FAISS index file
│ ├── metadata.pkl # DataFrame with metadata
│ └── embeddings.npy # Raw embeddings (optional)
└── README.md
```
## How It Works
1. **Embedding Creation**: When you upload data, each sequence is converted to a dense vector (embedding) using a sentence transformer model
2. **FAISS Indexing**: Embeddings are stored in a FAISS index optimized for similarity search
3. **Search**: Your query is embedded using the same model, then FAISS finds the most similar vectors using cosine similarity
4. **Results**: The original sequences and metadata are returned, ranked by similarity
## Performance Tips
- **Model Choice**: `all-MiniLM-L6-v2` is fast and good for most use cases. Use `all-mpnet-base-v2` for higher quality at the cost of speed.
- **Batch Size**: For large datasets, the model processes in batches automatically
- **GPU**: If you have a CUDA-capable GPU, install `faiss-gpu` instead of `faiss-cpu` for faster indexing
## License
MIT