File size: 7,806 Bytes
01b271c
2e52e7f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
01b271c
 
2e52e7f
01b271c
2e52e7f
01b271c
2e52e7f
 
 
 
 
01b271c
2e52e7f
 
 
 
 
 
 
01b271c
2e52e7f
 
 
 
 
 
 
 
 
 
 
01b271c
2e52e7f
 
 
 
 
 
 
 
 
01b271c
2e52e7f
 
 
 
 
 
 
 
 
 
 
01b271c
2e52e7f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
01b271c
2e52e7f
 
 
01b271c
2e52e7f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
01b271c
53c6b1b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
01b271c
2e52e7f
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
---
license: apache-2.0
task_categories:
  - question-answering
  - text-retrieval
language:
  - en
tags:
  - rag
  - retrieval-augmented-generation
  - multi-hop-reasoning
  - hotpotqa
  - information-retrieval
  - question-answering
  - evaluation
size_categories:
  - 1K<n<10K
---

# StratRAG

**StratRAG** is a retrieval evaluation dataset for benchmarking Retrieval-Augmented Generation (RAG) systems on multi-hop reasoning tasks. It is derived from [HotpotQA](https://hotpotqa.github.io/) (distractor setting) and structured specifically for evaluating retrieval strategies — including sparse (BM25), dense, and hybrid approaches — in realistic, noisy document pool conditions.

---

## Why StratRAG?

Most RAG benchmarks either evaluate end-to-end generation quality or assume clean, small document sets. StratRAG addresses a gap: **retrieval evaluation under multi-hop, distractor-heavy conditions**, where:

- Each question requires reasoning across **2 gold documents**
- The retriever must find those 2 docs inside a pool of **15 candidates** (13 distractors)
- Questions span 3 types: **bridge**, **comparison**, and **yes-no**

This makes it suitable for measuring Recall@k, MRR, NDCG, and faithfulness of retrieved context before any generation step.

---

## Dataset Structure

### Splits

| Split      | Rows |
|------------|------|
| train      | 2000 |
| validation | 200  |

### Schema
```python
{
  "id":               str,          # e.g. "train_000042"
  "query":            str,          # the multi-hop question
  "reference_answer": str,          # ground-truth answer string
  "doc_pool": [                     # always exactly 15 documents
    {
      "doc_id": str,                # globally unique doc identifier
      "text":   str,                # title + paragraph body
      "source": str,                # paragraph title (from HotpotQA)
    }
  ],
  "gold_doc_indices": [int],        # indices into doc_pool (always [0, 1])
                                    # gold docs are always placed first
  "metadata": {
    "split":         str,           # "train" or "val"
    "question_type": str,           # "bridge" | "comparison" | "yes-no"
  },
  "created_at": str,                # ISO-8601 UTC timestamp
  "provenance": {
    "base": str,                    # "hotpot_qa(distractor)"
    "seed": int,                    # 42
  }
}
```

### Key design decisions

- **Gold docs are always at indices 0 and 1** in `doc_pool`. This makes it trivial to compute oracle retrieval metrics and verify your retriever's upper bound.
- **13 distractor documents** per row are drawn from HotpotQA's built-in distractor paragraphs — these are topically related and intentionally difficult to distinguish from gold docs.
- **Empty-text paragraphs are filtered out** — HotpotQA contains some paragraphs with no sentence content; these are excluded from distractor slots to ensure all 15 docs have real text.

---

## Usage
```python
from datasets import load_dataset

ds = load_dataset("Aryanp088/StratRAG")

# Inspect a single training example
row = ds["train"][0]
print("Query:", row["query"])
print("Answer:", row["reference_answer"])
print("Question type:", row["metadata"]["question_type"])
print("Gold doc indices:", row["gold_doc_indices"])
print("Number of docs in pool:", len(row["doc_pool"]))

# Access gold documents directly
for idx in row["gold_doc_indices"]:
    print(f"\nGold doc [{idx}]:", row["doc_pool"][idx]["text"][:200])
```

---

## Evaluation Example
```python
from datasets import load_dataset

ds = load_dataset("Aryanp088/StratRAG", split="validation")

def recall_at_k(row, k=2):
    """
    Simulate a retriever that returns the first k docs (random baseline).
    Replace retrieved_indices with your retriever's output.
    """
    retrieved_indices = list(range(k))   # replace with your retriever
    gold = set(row["gold_doc_indices"])
    hits = len(gold & set(retrieved_indices))
    return hits / len(gold)

scores = [recall_at_k(row, k=2) for row in ds]
print(f"Random Recall@2: {sum(scores)/len(scores):.3f}")  # ~0.133 (2/15)
```

To benchmark a real retriever (e.g. BM25):
```python
from rank_bm25 import BM25Okapi

def bm25_recall_at_k(row, k=2):
    docs    = [doc["text"].split() for doc in row["doc_pool"]]
    query   = row["query"].split()
    bm25    = BM25Okapi(docs)
    scores  = bm25.get_scores(query)
    top_k   = sorted(range(len(scores)), key=lambda i: scores[i], reverse=True)[:k]
    gold    = set(row["gold_doc_indices"])
    return len(gold & set(top_k)) / len(gold)

scores = [bm25_recall_at_k(row, k=2) for row in ds]
print(f"BM25 Recall@2: {sum(scores)/len(scores):.3f}")
```

---

## Question Type Distribution

| Type       | Train | Validation |
|------------|-------|------------|
| bridge     | 1775  | 171        |
| yes-no     | 113   | 12         |
| comparison | 112   | 17         |

---

## Provenance & Reproducibility

All rows are derived from HotpotQA (distractor configuration) with `seed=42`.
```python
from datasets import load_dataset
hotpot = load_dataset("hotpot_qa", "distractor")
```

---


---

## Benchmark Results

Evaluated on the validation split (n=200) using three retrieval strategies.

### Overall Results

| Retriever | Recall@1 | Recall@2 | Recall@5 | MRR    | NDCG@5 |
|-----------|----------|----------|----------|--------|--------|
| Random    | 0.0525   | 0.1425   | 0.3300   | 0.3190 | 0.2336 |
| BM25      | 0.3950   | 0.6000   | 0.8150   | 0.8732 | 0.7624 |
| Dense (MiniLM-L6-v2) | 0.4175 | 0.6500 | 0.8600 | 0.9035 | 0.8087 |
| **Hybrid (BM25 + Dense)** | **0.4400** | **0.6975** | **0.9050** | **0.9310** | **0.8543** |

> Hybrid retriever uses equal-weight (α=0.5) min-max normalized score fusion.

### Hybrid Retriever — By Question Type

| Question Type | n   | Recall@2 | MRR    | NDCG@5 |
|---------------|-----|----------|--------|--------|
| bridge        | 171 | 0.6696   | 0.9281 | 0.8418 |
| comparison    | 17  | 0.8824   | 0.9706 | 0.9473 |
| yes-no        | 12  | 0.8333   | 0.9167 | 0.9007 |

**Key findings:**
- Hybrid retrieval consistently outperforms sparse and dense individually across all metrics
- Dense retrieval outperforms BM25 on all metrics, highlighting the importance of semantic matching for multi-hop questions
- Bridge questions are the hardest retrieval type (Recall@2 = 0.67), as they require cross-document reasoning without strong lexical overlap
- Comparison and yes-no questions benefit more from BM25's keyword matching (higher Recall@2)
- Recall@5 of 0.905 for Hybrid shows that 90% of the time, both gold documents appear in the top 5 — a strong upper bound for downstream generation

---

## Limitations

- **English only** — inherited from HotpotQA
- **Wikipedia-domain** — all documents are Wikipedia paragraphs; may not generalize to other domains without adaptation
- **2,200 total rows** — suitable for retriever evaluation and fine-tuning signal, not large-scale pretraining
- **Gold position is fixed** — gold docs are always at indices 0 and 1. Shuffle `doc_pool` before training retrievers to avoid position bias
```python
import random
random.shuffle(row["doc_pool"])   # shuffle before use in training
```

---

## Citation

If you use StratRAG in your work, please cite:
```bibtex
@dataset{patodiya2026stratrag,
  author    = {Patodiya, Aryan},
  title     = {StratRAG: A Multi-Hop Retrieval Evaluation Dataset for RAG Systems},
  year      = {2026},
  publisher = {Hugging Face},
  url       = {https://huggingface.co/datasets/Aryanp088/StratRAG}
}
```

---

## Author

**Aryan Patodiya** — ML Systems Engineer  
MS Computer Science @ California State University, Fresno  
[Portfolio](https://aryanp-portfolio.netlify.app) · [GitHub](https://github.com/aryanpatodiya08) · aryanpatodiya018@gmail.com