File size: 2,139 Bytes
5e33af5
bb60941
 
5e33af5
 
 
e03ab78
5e33af5
 
 
bb60941
5e33af5
 
bb60941
2ea10f9
bb60941
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
---
title: NetraEmbed
emoji: πŸ‘οΈ
colorFrom: yellow
colorTo: purple
sdk: gradio
sdk_version: 6.0.2
app_file: app.py
pinned: false
license: mit
short_description: Universal Multilingual Multimodal Document Retrieval
---

# NetraEmbed - Universal Multilingual Multimodal Document Retrieval

This Space demonstrates **NetraEmbed** and **ColNetraEmbed**, state-of-the-art multilingual multimodal document retrieval models based on the BiGemma3 and ColGemma3 architectures.

## Features

- **NetraEmbed (BiGemma3)**: Single-vector embedding with Matryoshka representation for fast retrieval
- **ColNetraEmbed (ColGemma3)**: Multi-vector embedding with late interaction for high-quality retrieval with attention heatmaps
- **ZeroGPU Integration**: Efficient dynamic GPU allocation for on-demand model loading
- **PDF Document Support**: Upload PDFs and perform semantic search across pages
- **Side-by-side Comparison**: Compare both models simultaneously

## Citation

If you use NetraEmbed or ColNetraEmbed in your research, please cite:

```bibtex
@misc{kolavi2025m3druniversalmultilingualmultimodal,
  title={M3DR: Towards Universal Multilingual Multimodal Document Retrieval},
  author={Adithya S Kolavi and Vyoman Jain},
  year={2025},
  eprint={2512.03514},
  archivePrefix={arXiv},
  primaryClass={cs.IR},
  url={https://arxiv.org/abs/2512.03514}
}
```

## Links

- πŸ“„ [Paper](https://arxiv.org/abs/2512.03514)
- πŸ’» [GitHub](https://github.com/adithya-s-k/colpali)
- πŸ€— [Models on Hugging Face](https://huggingface.co/Cognitive-Lab)
- 🌐 [CognitiveLab Website](https://www.cognitivelab.in)

## Usage

1. **Load Model**: Select your preferred model (NetraEmbed, ColNetraEmbed, or Both) and click "Load Model"
2. **Upload PDF**: Upload a PDF document to index
3. **Index Document**: Click "Index Document" to process and embed the pages
4. **Query**: Enter your search query and click "Search" to retrieve relevant pages

This Space uses ZeroGPU for dynamic GPU allocation. Models are loaded on-demand when functions are called.

Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference