Update README.md
Browse files
README.md
CHANGED
|
@@ -14,4 +14,87 @@ configs:
|
|
| 14 |
- split: queries
|
| 15 |
path: queries.jsonl
|
| 16 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 17 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 14 |
- split: queries
|
| 15 |
path: queries.jsonl
|
| 16 |
---
|
| 17 |
+
# 📚 Translated LONG2RAG (MTEB-Style Retrieval Dataset)
|
| 18 |
+
|
| 19 |
+
## Dataset Summary
|
| 20 |
+
|
| 21 |
+
This dataset is a **translated version** of the [LONG2RAG benchmark](https://github.com/QZH-777/longrag) (Qi et al., EMNLP Findings 2024), adapted into **MTEB-style retrieval format** for evaluating multilingual **retrieval-augmented generation (RAG)** and **long-context retrieval** systems.
|
| 22 |
+
|
| 23 |
+
LONG2RAG was originally designed to evaluate how well large language models (LLMs) incorporate key points from retrieved long documents into long-form answers. It includes **280 complex, practical questions** across **10 domains** and **8 question categories**, each paired with **5 retrieved documents** (avg. length ~2,444 words).
|
| 24 |
+
|
| 25 |
+
This translated version preserves the structure but reformats it into **query–document relevance pairs** suitable for **retrieval evaluation** under the [Massive Text Embedding Benchmark (MTEB)](https://huggingface.co/collections/mteb/mteb-benchmark-63f5f98f79c33120b8f94d1d).
|
| 26 |
+
|
| 27 |
---
|
| 28 |
+
|
| 29 |
+
## Supported Tasks and Leaderboards
|
| 30 |
+
|
| 31 |
+
* **Task Category:** Retrieval
|
| 32 |
+
* **Task:** Given a natural language query, rank candidate documents by relevance.
|
| 33 |
+
* **MTEB Integration:** Compatible with `mteb` evaluation framework.
|
| 34 |
+
|
| 35 |
+
---
|
| 36 |
+
|
| 37 |
+
## Languages
|
| 38 |
+
|
| 39 |
+
* **Original:** English
|
| 40 |
+
* **This release:** Translated into Persian
|
| 41 |
+
|
| 42 |
+
---
|
| 43 |
+
|
| 44 |
+
|
| 45 |
+
## Dataset Details
|
| 46 |
+
|
| 47 |
+
### Queries
|
| 48 |
+
- **280** complex, uncontaminated, long-form questions.
|
| 49 |
+
|
| 50 |
+
### Corpus
|
| 51 |
+
- Retrieved real-world documents (**5 per query**).
|
| 52 |
+
|
| 53 |
+
### Relevance Labels
|
| 54 |
+
- Binary (**relevant / not relevant**).
|
| 55 |
+
|
| 56 |
+
---
|
| 57 |
+
|
| 58 |
+
## Domains and Question Categories
|
| 59 |
+
|
| 60 |
+
### Domains (10)
|
| 61 |
+
- AI
|
| 62 |
+
- Biology
|
| 63 |
+
- Economics
|
| 64 |
+
- Film
|
| 65 |
+
- History
|
| 66 |
+
- Music
|
| 67 |
+
- Religion
|
| 68 |
+
- Sports
|
| 69 |
+
- Technology
|
| 70 |
+
- Others
|
| 71 |
+
|
| 72 |
+
### Question Categories (8)
|
| 73 |
+
- Factual
|
| 74 |
+
- Explanatory
|
| 75 |
+
- Comparative
|
| 76 |
+
- Subjective
|
| 77 |
+
- Methodological
|
| 78 |
+
- Causal
|
| 79 |
+
- Hypothetical
|
| 80 |
+
- Predictive
|
| 81 |
+
|
| 82 |
+
---
|
| 83 |
+
|
| 84 |
+
## Data Splits
|
| 85 |
+
|
| 86 |
+
- **test**: 280 queries
|
| 87 |
+
|
| 88 |
+
Each query has **5 candidate documents**, aligned with **MTEB retrieval style**.
|
| 89 |
+
|
| 90 |
+
---
|
| 91 |
+
|
| 92 |
+
## Citation
|
| 93 |
+
|
| 94 |
+
```bibtex
|
| 95 |
+
@inproceedings{qi2024long2rag,
|
| 96 |
+
title = {LONG2RAG: Evaluating Long-Context \& Long-Form Retrieval-Augmented Generation with Key Point Recall},
|
| 97 |
+
author = {Qi, Zehan and Xu, Rongwu and Guo, Zhijiang and Wang, Cunxiang and Zhang, Hao and Xu, Wei},
|
| 98 |
+
booktitle = {Findings of the Association for Computational Linguistics: EMNLP 2024},
|
| 99 |
+
year = {2024}
|
| 100 |
+
}
|