File size: 4,050 Bytes
5ad4bb0
 
 
 
 
e146b12
 
 
5ad4bb0
7764405
 
e146b12
 
 
bb3bd24
8fccced
bb3bd24
 
8fccced
 
 
 
 
e146b12
 
8fccced
 
 
e146b12
 
8fccced
 
 
 
 
 
 
 
 
 
e146b12
8fccced
 
 
 
 
 
 
 
 
 
 
 
 
e146b12
 
8fccced
e146b12
 
8fccced
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e146b12
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
---
license: mit
language:
- en
tags:
- science
- accelerator-physics
- particle-accelerator
pretty_name: Accel-IR
size_categories:
- 1K<n<10K
task_categories:
- text-retrieval
- question-answering
configs:
- config_name: expert_core
  data_files:
  - split: test
    path: "Accel_IR_expert_core.csv"
- config_name: augmented
  data_files:
  - split: test
    path: "Accel_IR_augmented.csv"
---

# Accel-IR Benchmark: A Gold Standard for Particle Accelerator Physics

This repository contains the **Accel-IR Benchmark**, a domain-specific Information Retrieval (IR) dataset for particle accelerator physics. It was developed as part of the Master's Thesis *"From Dataset to Optimization: A Benchmarking Framework for Information Retrieval in the Particle Accelerator Domain"* by **Qing Dai** (University of Zurich, 2025), in collaboration with the **Paul Scherrer Institute (PSI)**.


## Dataset Configurations

This benchmark is available in two configurations. You can load specific versions based on your evaluation needs:

| Configuration | Rows | Description | Use Case |
| :--- | :--- | :--- | :--- |
| **`expert_core`** | 390 | Purely **expert-annotated** pairs. Labeled by 7 domain experts (PhDs/Researchers) from PSI.  | Precise evaluation against human ground truth. |
| **`augmented`** | 1,357 | The `expert_core` + **curated hard negatives**. The negatives were generated using a expert-validated automatic annotation pipeline. | Realistic IR evaluation with many more negatives than positives. |

## Dataset Structure

### Data Fields
Each row in the dataset represents a query-document pair with the following columns:
- **`Source`**:  The referenced paper or an IPAC publication, source of the chunks.
- **`Question`**: The domain-specific scientific question.
- **`Answer`**: Answer to the question.
- **`Question_type`**: The category of the question, simulating diverse information needs:
  - `Fact`: Specific details or parameters.
  - `Definition`: Explanations of concepts/terms.
  - `Reasoning`: Logic behind phenomena or mechanisms.
  - `Summary`: Key points or conclusions.
- **`Referenced_file(s)`**: Referenced papers for the questions, provided by experts.
- **`chunk_text`**: The text passage retrieved from domain-expert-referenced papers or IPAC conference papers.
- **`expert_annotation`** *(Core only)*: The raw relevance score given by domain experts on a 5-point Likert scale:
  - `1`: Irrelevant
  - `2`: Partially Irrelevant
  - `3`: Hard to Decide (Excluded from Core)
  - `4`: Partially Relevant
  - `5`: Relevant
- **`specific to paper`**: Indicates if the question is "Context-Dependent" (answerable *only* by the referenced paper) or "General" (answerable by broader domain knowledge).
- **`Label`**: The binary ground truth used for evaluation metrics (nDCG/MAP).
  - `1` (Relevant): Derived from expert scores 4 & 5.
  - `0` (Irrelevant): Derived from expert scores 1 & 2, or pipeline hard negatives.
  
## Creation Methodology

1.  **Expert Core**:
    -   Created by **7 domain experts** from the Electron Beam Instrumentation Group at PSI.
    -   Experts reviewed query-chunk pairs and annotated them on a 1-5 scale using a custom interface.
    -   Pairs labeled as "3 - Not Sure" were removed to ensure no ambiguity.

2.  **Augmentation (Hard Negatives)**:
    -   To simulate realistic retrieval scenarios where negatives far outnumber positives, the core dataset was augmented.
    -   **Hard Negatives** were generated using an expert-validated automatic annotation pipeline.

## Usage

You can load the datasets using the Hugging Face `datasets` library.

### Load the Expert Core (390 pairs)
```python
from datasets import load_dataset

# Load the pure expert-annotated subset
ds_core = load_dataset("qdai/Accel-IR", "expert_core", split="test")

print(ds_core[0])
```


## Citation
If you use this dataset, please cite:
> Qing Dai, "From Dataset to Optimization: A Benchmarking Framework for Information Retrieval in the Particle Accelerator Domain", Master's Thesis, University of Zurich, 2025.