Accel-IR / README.md
qdai's picture
Update README.md
8fccced verified
---
license: mit
language:
- en
tags:
- science
- accelerator-physics
- particle-accelerator
pretty_name: Accel-IR
size_categories:
- 1K<n<10K
task_categories:
- text-retrieval
- question-answering
configs:
- config_name: expert_core
data_files:
- split: test
path: "Accel_IR_expert_core.csv"
- config_name: augmented
data_files:
- split: test
path: "Accel_IR_augmented.csv"
---
# Accel-IR Benchmark: A Gold Standard for Particle Accelerator Physics
This repository contains the **Accel-IR Benchmark**, a domain-specific Information Retrieval (IR) dataset for particle accelerator physics. It was developed as part of the Master's Thesis *"From Dataset to Optimization: A Benchmarking Framework for Information Retrieval in the Particle Accelerator Domain"* by **Qing Dai** (University of Zurich, 2025), in collaboration with the **Paul Scherrer Institute (PSI)**.
## Dataset Configurations
This benchmark is available in two configurations. You can load specific versions based on your evaluation needs:
| Configuration | Rows | Description | Use Case |
| :--- | :--- | :--- | :--- |
| **`expert_core`** | 390 | Purely **expert-annotated** pairs. Labeled by 7 domain experts (PhDs/Researchers) from PSI. | Precise evaluation against human ground truth. |
| **`augmented`** | 1,357 | The `expert_core` + **curated hard negatives**. The negatives were generated using a expert-validated automatic annotation pipeline. | Realistic IR evaluation with many more negatives than positives. |
## Dataset Structure
### Data Fields
Each row in the dataset represents a query-document pair with the following columns:
- **`Source`**: The referenced paper or an IPAC publication, source of the chunks.
- **`Question`**: The domain-specific scientific question.
- **`Answer`**: Answer to the question.
- **`Question_type`**: The category of the question, simulating diverse information needs:
- `Fact`: Specific details or parameters.
- `Definition`: Explanations of concepts/terms.
- `Reasoning`: Logic behind phenomena or mechanisms.
- `Summary`: Key points or conclusions.
- **`Referenced_file(s)`**: Referenced papers for the questions, provided by experts.
- **`chunk_text`**: The text passage retrieved from domain-expert-referenced papers or IPAC conference papers.
- **`expert_annotation`** *(Core only)*: The raw relevance score given by domain experts on a 5-point Likert scale:
- `1`: Irrelevant
- `2`: Partially Irrelevant
- `3`: Hard to Decide (Excluded from Core)
- `4`: Partially Relevant
- `5`: Relevant
- **`specific to paper`**: Indicates if the question is "Context-Dependent" (answerable *only* by the referenced paper) or "General" (answerable by broader domain knowledge).
- **`Label`**: The binary ground truth used for evaluation metrics (nDCG/MAP).
- `1` (Relevant): Derived from expert scores 4 & 5.
- `0` (Irrelevant): Derived from expert scores 1 & 2, or pipeline hard negatives.
## Creation Methodology
1. **Expert Core**:
- Created by **7 domain experts** from the Electron Beam Instrumentation Group at PSI.
- Experts reviewed query-chunk pairs and annotated them on a 1-5 scale using a custom interface.
- Pairs labeled as "3 - Not Sure" were removed to ensure no ambiguity.
2. **Augmentation (Hard Negatives)**:
- To simulate realistic retrieval scenarios where negatives far outnumber positives, the core dataset was augmented.
- **Hard Negatives** were generated using an expert-validated automatic annotation pipeline.
## Usage
You can load the datasets using the Hugging Face `datasets` library.
### Load the Expert Core (390 pairs)
```python
from datasets import load_dataset
# Load the pure expert-annotated subset
ds_core = load_dataset("qdai/Accel-IR", "expert_core", split="test")
print(ds_core[0])
```
## Citation
If you use this dataset, please cite:
> Qing Dai, "From Dataset to Optimization: A Benchmarking Framework for Information Retrieval in the Particle Accelerator Domain", Master's Thesis, University of Zurich, 2025.