Necent's picture
Upload README.md with huggingface_hub
1fdf3e0 verified
---
language:
- en
- ru
tags:
- efficientrag
- multi-hop-qa
- token-classification
- sequence-classification
- deberta-v3
license: mit
base_model: microsoft/mdeberta-v3-base
---
# EfficientRAG Labeler (mdeberta-v3-base)
**Labeler** component of [EfficientRAG](https://arxiv.org/abs/2408.04259) — dual-headed DeBERTa model for multi-hop retrieval.
## What it does
Given a query and a retrieved chunk, the Labeler:
1. **Sequence classification**: Is this chunk relevant (`CONTINUE`) or irrelevant (`TERMINATE`)?
2. **Token classification**: Which tokens in the chunk are useful for answering?
## Architecture
- Base: `microsoft/mdeberta-v3-base` (86M params, multilingual)
- Custom dual head: `DebertaForSequenceTokenClassification`
- Token head: binary per-token (useful/useless)
- Sequence head: binary per-chunk (CONTINUE/TERMINATE)
## Training
| | |
|--|--|
| Data | 30,818 samples (HotpotQA EN + Dragon-derec RU) |
| Epochs | 2 |
| Batch size | 4 |
| LR | 5e-6 |
| Max length | 384 |
| Hardware | Apple M3 Pro, ~3.4 hours |
## Usage
## Results on DRAGON benchmark
| Metric | Baseline | EfficientRAG | Delta |
|--------|----------|-------------|-------|
| MRR (multi-hop) | 0.736 | 0.798 | **+0.062** |
| MRR (overall) | 0.783 | 0.822 | **+0.040** |
| Precision | 0.187 | 0.582 | **+0.395** |
## Related
- Training data: [Necent/efficientrag-labeler-training-data](https://huggingface.co/datasets/Necent/efficientrag-labeler-training-data)
- Filter model: [Necent/efficientrag-filter-mdeberta-v3-base](https://huggingface.co/Necent/efficientrag-filter-mdeberta-v3-base)
- Paper: [EfficientRAG (arXiv:2408.04259)](https://arxiv.org/abs/2408.04259)