NanoCrumb / README.md
quicktensor's picture
Upload folder using huggingface_hub
2ac7793 verified
---
license: apache-2.0
task_categories:
- text-retrieval
language:
- en
tags:
- information-retrieval
- benchmark
- clinical-trials
- code-search
- legal-qa
size_categories:
- 10K<n<100K
configs:
- config_name: clinical_trial
data_files:
- split: queries
path: "clinical_trial/queries.jsonl"
- split: documents
path: "clinical_trial/documents.jsonl"
- split: qrels
path: "clinical_trial/qrels.jsonl"
- config_name: code_retrieval
data_files:
- split: queries
path: "code_retrieval/queries.jsonl"
- split: documents
path: "code_retrieval/documents.jsonl"
- split: qrels
path: "code_retrieval/qrels.jsonl"
- config_name: legal_qa
data_files:
- split: queries
path: "legal_qa/queries.jsonl"
- split: documents
path: "legal_qa/documents.jsonl"
- split: qrels
path: "legal_qa/qrels.jsonl"
- config_name: paper_retrieval
data_files:
- split: queries
path: "paper_retrieval/queries.jsonl"
- split: documents
path: "paper_retrieval/documents.jsonl"
- split: qrels
path: "paper_retrieval/qrels.jsonl"
- config_name: set_operation_entity_retrieval
data_files:
- split: queries
path: "set_operation_entity_retrieval/queries.jsonl"
- split: documents
path: "set_operation_entity_retrieval/documents.jsonl"
- split: qrels
path: "set_operation_entity_retrieval/qrels.jsonl"
- config_name: stack_exchange
data_files:
- split: queries
path: "stack_exchange/queries.jsonl"
- split: documents
path: "stack_exchange/documents.jsonl"
- split: qrels
path: "stack_exchange/qrels.jsonl"
- config_name: theorem_retrieval
data_files:
- split: queries
path: "theorem_retrieval/queries.jsonl"
- split: documents
path: "theorem_retrieval/documents.jsonl"
- split: qrels
path: "theorem_retrieval/qrels.jsonl"
- config_name: tip_of_the_tongue
data_files:
- split: queries
path: "tip_of_the_tongue/queries.jsonl"
- split: documents
path: "tip_of_the_tongue/documents.jsonl"
- split: qrels
path: "tip_of_the_tongue/qrels.jsonl"
---
# NanoCrumb Dataset
A curated subset of the [Crumb](https://huggingface.co/datasets/jfkback/crumb) retrieval dataset, designed for rapid experimentation and evaluation of information retrieval systems.
## Dataset Summary
**NanoCrumb** distills the large Crumb dataset (10.5 GB, 6.36M rows) into a manageable benchmark while maintaining task diversity across 8 different retrieval domains.
- **Total Size**: ~125 MB (JSONL format)
- **Queries**: 400 (50 per task split)
- **Documents**: 30,040 unique passages
- **Query-Document Pairs**: 31,754
- **Configs**: 8 task-specific configs
## Configs (Task Splits)
Each config represents a different retrieval domain:
| Config Name | Queries | Documents | Docs/Query (avg) | Description |
|------------|---------|-----------|------------------|-------------|
| `clinical_trial` | 50 | 22,251 | 464 | Match patients to clinical trials |
| `paper_retrieval` | 50 | 4,402 | 102 | Find relevant academic papers |
| `set_operation_entity_retrieval` | 50 | 1,533 | 31 | Entity-based retrieval |
| `code_retrieval` | 50 | 1,206 | 24 | Find relevant code snippets |
| `tip_of_the_tongue` | 50 | 363 | 7 | Recall items from vague descriptions |
| `stack_exchange` | 50 | 125 | 3 | Find relevant Q&A posts |
| `legal_qa` | 50 | 86 | 2 | Legal question answering |
| `theorem_retrieval` | 50 | 74 | 2 | Find mathematical theorems |
## Dataset Structure
Each config contains three splits:
### `queries`
- `query_id`: Unique query identifier (string)
- `query_content`: The query text (string)
- `instruction`: Task-specific instructions (string)
- `passage_qrels`: List of relevant passages with graded relevance scores (list)
- `task_split`: Task domain name (string)
- `metadata`: Additional task-specific information (string)
- `use_max_p`: Boolean flag for MaxP aggregation (bool)
### `documents`
- `document_id`: Unique document identifier (string)
- `document_content`: The passage text (string)
- `parent_id`: Links passages to source documents (string)
- `task_split`: Task domain name (string)
- `metadata`: Document metadata (string)
### `qrels`
- `query_id`: Query identifier (string)
- `document_id`: Document identifier (string)
- `relevance_score`: Graded relevance 0.0-2.0 (float)
- `binary_relevance`: Binary relevance 0 or 1 (int)
- `task_split`: Task domain name (string)
## Usage
```python
from datasets import load_dataset
# Load a specific config (task split)
clinical_data = load_dataset("YOUR_USERNAME/nanocrumb", "clinical_trial")
# Access the splits
queries = clinical_data['queries']
documents = clinical_data['documents']
qrels = clinical_data['qrels']
# Load all configs
all_configs = [
"clinical_trial", "code_retrieval", "legal_qa", "paper_retrieval",
"set_operation_entity_retrieval", "stack_exchange",
"theorem_retrieval", "tip_of_the_tongue"
]
for config_name in all_configs:
data = load_dataset("YOUR_USERNAME/nanocrumb", config_name)
print(f"{config_name}: {len(data['queries'])} queries")
```
## Sampling Methodology
For each task split:
1. **Query Selection**: Randomly sampled 50 queries from evaluation set (seed=42)
2. **Document Selection**:
- Include ALL positive documents (binary_relevance=1)
- Fill remainder with hard negatives (relevance=0) to reach ~100 docs per query
- Target: ~5,000 documents per task split
3. **Deduplication**: Documents shared across queries are deduplicated within each config
## Use Cases
- 🚀 **Rapid prototyping** of retrieval models
- 🧪 **Quick benchmarking** without downloading large datasets
- 📚 **Educational purposes** for learning IR techniques
- 🔬 **Ablation studies** across diverse domains
## Citation
If you use NanoCrumb, please cite the original Crumb dataset:
```bibtex
@misc{crumb2024,
title={Crumb: A Comprehensive Retrieval Benchmark},
author={[Original Crumb Authors]},
year={2024},
url={https://huggingface.co/datasets/jfkback/crumb}
}
```
## License
This dataset inherits the license from the original [Crumb dataset](https://huggingface.co/datasets/jfkback/crumb).