murnanedaniel commited on
Commit
9bc5ea1
·
verified ·
1 Parent(s): 5f9f180

Update hard_scatter/ggf/v1 dataset card

Browse files
Files changed (1) hide show
  1. README.md +4 -298
README.md CHANGED
@@ -845,7 +845,7 @@ configs:
845
 
846
  ## Dataset Description
847
 
848
- This dataset contains simulated high-energy physics collision events for Higgs boson production from gluon-gluon fusion with no pileup (single interaction per event) generated using the **Open Data Detector (ODD)** geometry within the **ACTS (A Common Tracking Software)** framework, representing a generic collider detector similar to those at the LHC.
849
 
850
  ### Dataset Summary
851
 
@@ -989,7 +989,7 @@ Truth information about generated particles before detector simulation.
989
  | `time` | list<float64> | Production time (ns) |
990
  | `num_tracker_hits` | list<int64> | Number of hits in tracker |
991
  | `num_calo_hits` | list<int64> | Number of hits in calorimeter |
992
- | `vertex_primary` | list<int64> | Primary vertex flag (1=primary, 0=secondary) |
993
  | `parent_id` | list<float64> | ID of parent particle |
994
 
995
  **Typical event**: ~200-500 particles per event
@@ -1077,312 +1077,18 @@ train_val, test = train_test_split(all_events, test_size=0.15, random_state=42)
1077
  train, val = train_test_split(train_val, test_size=0.176, random_state=42) # 0.176 * 0.85 ≈ 0.15
1078
  ```
1079
 
1080
- ## Dataset Creation
1081
-
1082
- ### Curation Rationale
1083
-
1084
- This dataset was created to support machine learning research in high-energy physics, specifically for:
1085
-
1086
- 1. **Benchmarking tracking algorithms**: Compare traditional and ML-based track reconstruction methods
1087
- 2. **Hierarchical representation learning**: Study information flow from detector hits → tracks → particles
1088
- 3. **Physics analysis**: Develop ML models for event classification and particle identification
1089
- 4. **Open science**: Provide publicly accessible, realistic detector simulation data
1090
-
1091
- This dataset contains simulated Higgs boson events with no pileup,
1092
- useful for developing and testing machine learning algorithms
1093
- for particle tracking and physics analysis.
1094
-
1095
-
1096
- ### Source Data
1097
-
1098
- #### Initial Data Collection and Normalization
1099
-
1100
- The data is generated through the following simulation chain:
1101
-
1102
- 1. **Event Generation**: Events generated using a Monte Carlo event generator
1103
- 2. **Detector Simulation**: Particle propagation through the Open Data Detector using ACTS
1104
- 3. **Digitization**: Conversion of energy deposits to realistic detector signals
1105
- 4. **Reconstruction**: Track finding and fitting using ACTS tracking algorithms
1106
- 5. **Format Conversion**: EDM4HEP → Parquet using the ColliderML data pipeline
1107
-
1108
- #### Who are the source data producers?
1109
-
1110
- The data is produced by the **ColliderML collaboration** as part of the **ATLAS ITk ML Reconstruction** project at NERSC (National Energy Research Scientific Computing Center).
1111
-
1112
- ### Annotations
1113
-
1114
- #### Annotation process
1115
-
1116
- The dataset includes truth-level annotations automatically generated during the simulation:
1117
-
1118
- - **Particle-level truth**: Generator-level particle information
1119
- - **Hit-to-particle associations**: Which particle created each detector hit
1120
- - **Track-to-particle matching**: `majority_particle_id` links reconstructed tracks to truth particles
1121
-
1122
- These annotations enable supervised learning for tasks like:
1123
- - Track efficiency (did we reconstruct this particle?)
1124
- - Track purity (how many hits belong to the correct particle?)
1125
- - Fake rate (how many tracks are not matched to real particles?)
1126
-
1127
- #### Who are the annotators?
1128
-
1129
- N/A (Annotations are from simulation ground truth)
1130
-
1131
- ### Personal and Sensitive Information
1132
-
1133
- This dataset contains only simulated physics data. No personal or sensitive information is included.
1134
-
1135
- ## Considerations for Using the Data
1136
-
1137
- ### Social Impact of Dataset
1138
-
1139
- This dataset supports fundamental physics research and ML algorithm development. It has no direct social impact but contributes to:
1140
-
1141
- - Open science and reproducible research
1142
- - Education in HEP and ML
1143
- - Development of algorithms that may have broader applications (e.g., pattern recognition, tracking in medical imaging)
1144
-
1145
- ### Discussion of Biases
1146
-
1147
- As a simulated dataset, biases may arise from:
1148
-
1149
- 1. **Generator-level biases**: The event generator's modeling of the physics process
1150
- 2. **Detector simulation biases**: Approximations in material interactions, detector response
1151
- 3. **Reconstruction biases**: Algorithm choices in track finding and fitting
1152
- 4. **Pileup modeling**: This dataset has no pileup; real LHC data has 20-60 simultaneous collisions
1153
-
1154
- Users should be aware that models trained on this data may not generalize to:
1155
- - Real detector data (requires calibration and alignment)
1156
- - Different detector geometries
1157
- - Different pileup conditions
1158
-
1159
- ### Other Known Limitations
1160
-
1161
- - **Limited statistics**: ~100000 events (consider data augmentation for large models)
1162
- - **Single physics process**: Only Higgs boson production from gluon-gluon fusion; does not include background processes
1163
- - **Idealized detector**: ODD is a generic detector, not an exact replica of ATLAS/CMS
1164
- - **Simplified simulation**: Some detector effects may be simplified
1165
-
1166
- ## Additional Information
1167
-
1168
- ### Dataset Curators
1169
-
1170
- This dataset is maintained by the ColliderML team:
1171
-
1172
- - Primary contact: daniel.thomas.murnane@cern.ch
1173
- - Collaboration: ATLAS ITk ML Reconstruction working group
1174
- - Infrastructure: NERSC (National Energy Research Scientific Computing Center)
1175
-
1176
- ### Licensing Information
1177
-
1178
- This dataset is released under the **Creative Commons Attribution 4.0 International (CC BY 4.0)** license.
1179
-
1180
- You are free to:
1181
- - **Share**: Copy and redistribute the material
1182
- - **Adapt**: Remix, transform, and build upon the material
1183
-
1184
- Under the following terms:
1185
- - **Attribution**: You must give appropriate credit and indicate if changes were made
1186
-
1187
- ### Citation Information
1188
-
1189
- If you use this dataset in your research, please cite:
1190
-
1191
- ```bibtex
1192
- @dataset{colliderml_ggf_v1_2025,
1193
- title={ {ColliderML: ColliderML Higgs Boson Production from Gluon-Gluon Fusion (No Pileup)} },
1194
- author={ {ColliderML Collaboration} },
1195
- year={ 2025 },
1196
- publisher={NERSC},
1197
- howpublished={\url{ https://huggingface.co/datasets/OpenDataDetector/ColliderML_higgs_pu0 }},
1198
- note={Simulation performed using ACTS and the Open Data Detector}
1199
- }
1200
- ```
1201
-
1202
- ### Contributions
1203
-
1204
- This dataset was produced using:
1205
-
1206
- - **ACTS (A Common Tracking Software)**: https://acts.readthedocs.io/
1207
- - **Open Data Detector**: https://acts.readthedocs.io/en/latest/examples/open_data_detector.html
1208
- - **EDM4HEP**: https://edm4hep.web.cern.ch/
1209
- - **ColliderML Pipeline**: https://github.com/ATLAS-ITk-ML/colliderml
1210
-
1211
- ## How to Use This Dataset
1212
-
1213
- ### Loading the Dataset
1214
-
1215
- The dataset is hosted on the NERSC public portal and can be streamed directly without downloading:
1216
-
1217
- ```python
1218
- from datasets import load_dataset
1219
-
1220
-
1221
- # Load particles
1222
- particles_ds = load_dataset(
1223
- "OpenDataDetector/ColliderML_higgs_pu0",
1224
- "particles",
1225
- split="train"
1226
- )
1227
-
1228
- # Load tracker_hits
1229
- tracker_hits_ds = load_dataset(
1230
- "OpenDataDetector/ColliderML_higgs_pu0",
1231
- "tracker_hits",
1232
- split="train"
1233
- )
1234
-
1235
- # Load calo_hits
1236
- calo_hits_ds = load_dataset(
1237
- "OpenDataDetector/ColliderML_higgs_pu0",
1238
- "calo_hits",
1239
- split="train"
1240
- )
1241
-
1242
- # Load tracks
1243
- tracks_ds = load_dataset(
1244
- "OpenDataDetector/ColliderML_higgs_pu0",
1245
- "tracks",
1246
- split="train"
1247
- )
1248
-
1249
- ```
1250
-
1251
- ### Example: Iterating Over Events
1252
-
1253
- ```python
1254
- import numpy as np
1255
-
1256
- # Iterate over first 10 events
1257
- for i, event in enumerate(particles_ds.take(10)):
1258
- event_id = event['event_id']
1259
- n_particles = len(event['particle_id'])
1260
-
1261
- print(f"Event {event_id}: {n_particles} particles")
1262
-
1263
- # Access list columns as numpy arrays
1264
- px = np.array(event['px'])
1265
- py = np.array(event['py'])
1266
- pz = np.array(event['pz'])
1267
-
1268
- # Compute transverse momentum
1269
- pt = np.sqrt(px**2 + py**2)
1270
- print(f" Mean pt: {pt.mean():.2f} GeV")
1271
- ```
1272
-
1273
- ### Example: Computing Track Features
1274
-
1275
- ```python
1276
- import numpy as np
1277
-
1278
- for event in tracks_ds.take(5):
1279
- # Get track parameters
1280
- qop = np.array(event['qop'])
1281
- theta = np.array(event['theta'])
1282
- phi = np.array(event['phi'])
1283
-
1284
- # Compute derived quantities
1285
- pt = np.abs(1.0 / qop) * np.sin(theta)
1286
- eta = -np.log(np.tan(theta / 2.0))
1287
-
1288
- print(f"Event {event['event_id']}: {len(qop)} tracks")
1289
- print(f" pt range: [{pt.min():.2f}, {pt.max():.2f}] GeV")
1290
- print(f" eta range: [{eta.min():.2f}, {eta.max():.2f}]")
1291
- ```
1292
-
1293
- ### Example: Matching Tracks to Particles
1294
-
1295
- ```python
1296
- # Load both datasets
1297
- particles = load_dataset("OpenDataDetector/ColliderML_higgs_pu0", "particles", split="train")
1298
- tracks = load_dataset("OpenDataDetector/ColliderML_higgs_pu0", "tracks", split="train")
1299
-
1300
- # Process event-by-event
1301
- for particle_event, track_event in zip(particles, tracks):
1302
- assert particle_event['event_id'] == track_event['event_id']
1303
-
1304
- # Get particle information
1305
- particle_ids = np.array(particle_event['particle_id'])
1306
- particle_px = np.array(particle_event['px'])
1307
- particle_py = np.array(particle_event['py'])
1308
-
1309
- # Get track information
1310
- track_particle_ids = np.array(track_event['majority_particle_id'])
1311
-
1312
- # Compute truth pt for particles
1313
- particle_pt = np.sqrt(particle_px**2 + particle_py**2)
1314
-
1315
- # Find matched tracks
1316
- for i, pid in enumerate(track_particle_ids):
1317
- if pid in particle_ids:
1318
- idx = np.where(particle_ids == pid)[0][0]
1319
- truth_pt = particle_pt[idx]
1320
- print(f"Track {i}: matched to particle {pid}, pt={truth_pt:.2f} GeV")
1321
- ```
1322
-
1323
- ### Data Location
1324
-
1325
- The Parquet files are hosted at:
1326
-
1327
- ```
1328
- https://portal.nersc.gov/cfs/m4958/ColliderML/hard_scatter/ggf/v1/parquet
1329
-
1330
- ├── truth/
1331
- │ └── particles/
1332
- │ └── *.parquet (100 files)
1333
-
1334
- ├── reco/
1335
- │ └── tracker_hits/
1336
- │ └── *.parquet (100 files)
1337
-
1338
- ├── reco/
1339
- │ └── calo_hits/
1340
- │ └── *.parquet (100 files)
1341
-
1342
- ├── reco/
1343
- │ └── tracks/
1344
- │ └── *.parquet (100 files)
1345
-
1346
- ```
1347
-
1348
- ### File Naming Convention
1349
-
1350
- Files follow the pattern:
1351
- ```
1352
- <campaign>.<dataset>.<version>.<category>.<object>.<event_range>.parquet
1353
- ```
1354
-
1355
- Example: `hard_scatter.ggf.v1.truth.particles.events0-999.parquet`
1356
- - Campaign: `hard_scatter`
1357
- - Dataset: `ggf`
1358
- - Version: `v1`
1359
- - Category: `truth`
1360
- - Object: one of particles, tracker_hits, calo_hits, tracks
1361
- - Event range: `eventsXXXX-YYYY` (inclusive)
1362
-
1363
- ### Performance Tips
1364
-
1365
- 1. **Streaming**: Use the dataset API for efficient memory usage
1366
- 2. **Batch processing**: Process events in chunks for better performance
1367
- 3. **Selective loading**: Only load the data types you need
1368
- 4. **Caching**: Use dataset caching for repeated experiments
1369
-
1370
- ### Related Datasets
1371
-
1372
-
1373
-
1374
  ### Support
1375
 
1376
  For questions, issues, or feature requests:
1377
  - Email: daniel.thomas.murnane@cern.ch
1378
- - GitHub: https://github.com/ATLAS-ITk-ML/colliderml/issues
1379
 
1380
  ### Acknowledgments
1381
 
1382
  This work was supported by:
1383
- - ATLAS ITk ML Reconstruction project
1384
  - NERSC computing resources
1385
  - U.S. Department of Energy, Office of Science
 
1386
 
1387
  ---
1388
 
 
845
 
846
  ## Dataset Description
847
 
848
+ This dataset contains simulated high-energy physics collision events for Higgs boson production from gluon-gluon fusion with no pileup (single interaction per event) generated using the **Open Data Detector (ODD)** geometry within the **Key4hep** and **ACTS (A Common Tracking Software)** frameworks, representing a generic collider detector similar to those at the HL-LHC.
849
 
850
  ### Dataset Summary
851
 
 
989
  | `time` | list<float64> | Production time (ns) |
990
  | `num_tracker_hits` | list<int64> | Number of hits in tracker |
991
  | `num_calo_hits` | list<int64> | Number of hits in calorimeter |
992
+ | `vertex_primary` | list<int64> | Primary vertex flag (1 = hard scatter, 2,...,N = pileup) |
993
  | `parent_id` | list<float64> | ID of parent particle |
994
 
995
  **Typical event**: ~200-500 particles per event
 
1077
  train, val = train_test_split(train_val, test_size=0.176, random_state=42) # 0.176 * 0.85 ≈ 0.15
1078
  ```
1079
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1080
  ### Support
1081
 
1082
  For questions, issues, or feature requests:
1083
  - Email: daniel.thomas.murnane@cern.ch
1084
+ - You can also open a discussion in the HuggingFace community panel for this dataset.
1085
 
1086
  ### Acknowledgments
1087
 
1088
  This work was supported by:
 
1089
  - NERSC computing resources
1090
  - U.S. Department of Energy, Office of Science
1091
+ - Danish Data Science Academy (DDSA)
1092
 
1093
  ---
1094