Datasets:
Modalities:
Text
Formats:
parquet
Languages:
English
Size:
1K - 10K
Tags:
rag
retrieval-augmented-generation
multi-hop-reasoning
hotpotqa
information-retrieval
question-answering
License:
File size: 7,806 Bytes
01b271c 2e52e7f 01b271c 2e52e7f 01b271c 2e52e7f 01b271c 2e52e7f 01b271c 2e52e7f 01b271c 2e52e7f 01b271c 2e52e7f 01b271c 2e52e7f 01b271c 2e52e7f 01b271c 2e52e7f 01b271c 2e52e7f 01b271c 53c6b1b 01b271c 2e52e7f | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 | ---
license: apache-2.0
task_categories:
- question-answering
- text-retrieval
language:
- en
tags:
- rag
- retrieval-augmented-generation
- multi-hop-reasoning
- hotpotqa
- information-retrieval
- question-answering
- evaluation
size_categories:
- 1K<n<10K
---
# StratRAG
**StratRAG** is a retrieval evaluation dataset for benchmarking Retrieval-Augmented Generation (RAG) systems on multi-hop reasoning tasks. It is derived from [HotpotQA](https://hotpotqa.github.io/) (distractor setting) and structured specifically for evaluating retrieval strategies — including sparse (BM25), dense, and hybrid approaches — in realistic, noisy document pool conditions.
---
## Why StratRAG?
Most RAG benchmarks either evaluate end-to-end generation quality or assume clean, small document sets. StratRAG addresses a gap: **retrieval evaluation under multi-hop, distractor-heavy conditions**, where:
- Each question requires reasoning across **2 gold documents**
- The retriever must find those 2 docs inside a pool of **15 candidates** (13 distractors)
- Questions span 3 types: **bridge**, **comparison**, and **yes-no**
This makes it suitable for measuring Recall@k, MRR, NDCG, and faithfulness of retrieved context before any generation step.
---
## Dataset Structure
### Splits
| Split | Rows |
|------------|------|
| train | 2000 |
| validation | 200 |
### Schema
```python
{
"id": str, # e.g. "train_000042"
"query": str, # the multi-hop question
"reference_answer": str, # ground-truth answer string
"doc_pool": [ # always exactly 15 documents
{
"doc_id": str, # globally unique doc identifier
"text": str, # title + paragraph body
"source": str, # paragraph title (from HotpotQA)
}
],
"gold_doc_indices": [int], # indices into doc_pool (always [0, 1])
# gold docs are always placed first
"metadata": {
"split": str, # "train" or "val"
"question_type": str, # "bridge" | "comparison" | "yes-no"
},
"created_at": str, # ISO-8601 UTC timestamp
"provenance": {
"base": str, # "hotpot_qa(distractor)"
"seed": int, # 42
}
}
```
### Key design decisions
- **Gold docs are always at indices 0 and 1** in `doc_pool`. This makes it trivial to compute oracle retrieval metrics and verify your retriever's upper bound.
- **13 distractor documents** per row are drawn from HotpotQA's built-in distractor paragraphs — these are topically related and intentionally difficult to distinguish from gold docs.
- **Empty-text paragraphs are filtered out** — HotpotQA contains some paragraphs with no sentence content; these are excluded from distractor slots to ensure all 15 docs have real text.
---
## Usage
```python
from datasets import load_dataset
ds = load_dataset("Aryanp088/StratRAG")
# Inspect a single training example
row = ds["train"][0]
print("Query:", row["query"])
print("Answer:", row["reference_answer"])
print("Question type:", row["metadata"]["question_type"])
print("Gold doc indices:", row["gold_doc_indices"])
print("Number of docs in pool:", len(row["doc_pool"]))
# Access gold documents directly
for idx in row["gold_doc_indices"]:
print(f"\nGold doc [{idx}]:", row["doc_pool"][idx]["text"][:200])
```
---
## Evaluation Example
```python
from datasets import load_dataset
ds = load_dataset("Aryanp088/StratRAG", split="validation")
def recall_at_k(row, k=2):
"""
Simulate a retriever that returns the first k docs (random baseline).
Replace retrieved_indices with your retriever's output.
"""
retrieved_indices = list(range(k)) # replace with your retriever
gold = set(row["gold_doc_indices"])
hits = len(gold & set(retrieved_indices))
return hits / len(gold)
scores = [recall_at_k(row, k=2) for row in ds]
print(f"Random Recall@2: {sum(scores)/len(scores):.3f}") # ~0.133 (2/15)
```
To benchmark a real retriever (e.g. BM25):
```python
from rank_bm25 import BM25Okapi
def bm25_recall_at_k(row, k=2):
docs = [doc["text"].split() for doc in row["doc_pool"]]
query = row["query"].split()
bm25 = BM25Okapi(docs)
scores = bm25.get_scores(query)
top_k = sorted(range(len(scores)), key=lambda i: scores[i], reverse=True)[:k]
gold = set(row["gold_doc_indices"])
return len(gold & set(top_k)) / len(gold)
scores = [bm25_recall_at_k(row, k=2) for row in ds]
print(f"BM25 Recall@2: {sum(scores)/len(scores):.3f}")
```
---
## Question Type Distribution
| Type | Train | Validation |
|------------|-------|------------|
| bridge | 1775 | 171 |
| yes-no | 113 | 12 |
| comparison | 112 | 17 |
---
## Provenance & Reproducibility
All rows are derived from HotpotQA (distractor configuration) with `seed=42`.
```python
from datasets import load_dataset
hotpot = load_dataset("hotpot_qa", "distractor")
```
---
---
## Benchmark Results
Evaluated on the validation split (n=200) using three retrieval strategies.
### Overall Results
| Retriever | Recall@1 | Recall@2 | Recall@5 | MRR | NDCG@5 |
|-----------|----------|----------|----------|--------|--------|
| Random | 0.0525 | 0.1425 | 0.3300 | 0.3190 | 0.2336 |
| BM25 | 0.3950 | 0.6000 | 0.8150 | 0.8732 | 0.7624 |
| Dense (MiniLM-L6-v2) | 0.4175 | 0.6500 | 0.8600 | 0.9035 | 0.8087 |
| **Hybrid (BM25 + Dense)** | **0.4400** | **0.6975** | **0.9050** | **0.9310** | **0.8543** |
> Hybrid retriever uses equal-weight (α=0.5) min-max normalized score fusion.
### Hybrid Retriever — By Question Type
| Question Type | n | Recall@2 | MRR | NDCG@5 |
|---------------|-----|----------|--------|--------|
| bridge | 171 | 0.6696 | 0.9281 | 0.8418 |
| comparison | 17 | 0.8824 | 0.9706 | 0.9473 |
| yes-no | 12 | 0.8333 | 0.9167 | 0.9007 |
**Key findings:**
- Hybrid retrieval consistently outperforms sparse and dense individually across all metrics
- Dense retrieval outperforms BM25 on all metrics, highlighting the importance of semantic matching for multi-hop questions
- Bridge questions are the hardest retrieval type (Recall@2 = 0.67), as they require cross-document reasoning without strong lexical overlap
- Comparison and yes-no questions benefit more from BM25's keyword matching (higher Recall@2)
- Recall@5 of 0.905 for Hybrid shows that 90% of the time, both gold documents appear in the top 5 — a strong upper bound for downstream generation
---
## Limitations
- **English only** — inherited from HotpotQA
- **Wikipedia-domain** — all documents are Wikipedia paragraphs; may not generalize to other domains without adaptation
- **2,200 total rows** — suitable for retriever evaluation and fine-tuning signal, not large-scale pretraining
- **Gold position is fixed** — gold docs are always at indices 0 and 1. Shuffle `doc_pool` before training retrievers to avoid position bias
```python
import random
random.shuffle(row["doc_pool"]) # shuffle before use in training
```
---
## Citation
If you use StratRAG in your work, please cite:
```bibtex
@dataset{patodiya2026stratrag,
author = {Patodiya, Aryan},
title = {StratRAG: A Multi-Hop Retrieval Evaluation Dataset for RAG Systems},
year = {2026},
publisher = {Hugging Face},
url = {https://huggingface.co/datasets/Aryanp088/StratRAG}
}
```
---
## Author
**Aryan Patodiya** — ML Systems Engineer
MS Computer Science @ California State University, Fresno
[Portfolio](https://aryanp-portfolio.netlify.app) · [GitHub](https://github.com/aryanpatodiya08) · aryanpatodiya018@gmail.com |