Necent's picture
Upload README.md with huggingface_hub
4cf68c6 verified
---
language:
- en
- ru
tags:
- efficientrag
- multi-hop-qa
- token-classification
- deberta-v3
license: mit
base_model: microsoft/mdeberta-v3-base
---
# EfficientRAG Filter (mdeberta-v3-base)
**Filter** component of [EfficientRAG](https://arxiv.org/abs/2408.04259) — constructs next-hop queries via token selection.
## What it does
Given the original question + extracted useful tokens, the Filter selects which tokens to keep in the next retrieval query. This is extractive (no generation) — it picks words from the input.
## Architecture
- Base: `microsoft/mdeberta-v3-base` (86M params, multilingual)
- Standard `DebertaV2ForTokenClassification` with 2 labels (keep/discard)
## Training
| | |
|--|--|
| Data | 5,691 samples (HotpotQA EN + Dragon-derec RU) |
| Epochs | 2 |
| Batch size | 4 |
| LR | 1e-5 |
| Max length | 128 |
| Hardware | Apple M3 Pro, ~17 minutes |
## Usage
## Related
- Training data: [Necent/efficientrag-filter-training-data](https://huggingface.co/datasets/Necent/efficientrag-filter-training-data)
- Labeler model: [Necent/efficientrag-labeler-mdeberta-v3-base](https://huggingface.co/Necent/efficientrag-labeler-mdeberta-v3-base)
- Paper: [EfficientRAG (arXiv:2408.04259)](https://arxiv.org/abs/2408.04259)