File size: 4,409 Bytes
630793a 302994a 30118fd 630793a 30118fd 630793a 30118fd 630793a 30118fd 302994a 30118fd 302994a 30118fd 302994a 30118fd 302994a 30118fd 302994a 30118fd 630793a 302994a ef25578 302994a 30118fd 630793a 302994a 30118fd 302994a 30118fd 302994a 30118fd 302994a 30118fd 630793a 302994a 30118fd fa395e5 302994a | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 | ---
language:
- zh
license: apache-2.0
task_categories:
- question-answering
- tabular-classification
- text-generation
tags:
- data-analytics
- agents
- document-understanding
- benchmark
pretty_name: AIDABench
---
# Dataset Card for AIDABench
## Dataset Summary
**AIDABench** is a benchmark for evaluating AI systems on **end-to-end data analytics over real-world documents**. It contains **600+** diverse analytical tasks grounded in realistic scenarios and spans heterogeneous data sources such as **spreadsheets, databases, financial reports, and operational records**. Tasks are designed to be challenging, often requiring multi-step reasoning and tool use to complete reliably.

*Figure 1: Overview of the AIDABench evaluation framework.*
## Supported Tasks and Evaluation Targets
AIDABench focuses on practical document analytics workflows where a model/agent must read files, reason over structured data, and produce a final deliverable.
### Task Categories
The dataset is organized around three primary capability dimensions:
- **File Generation (43.3%)**
Data wrangling and transformation tasks such as filtering, normalization, deduplication, joins, and cross-sheet linkage, with outputs as generated files (e.g., spreadsheets).
- **Question Answering (QA) (37.5%)**
Analytical queries such as aggregation, averages, ranking, comparisons, and trend analysis, with outputs as final answers.
- **Data Visualization (19.2%)**
Chart creation/adaptation tasks (e.g., bar/line/pie) including style requirements and presentation constraints, with outputs as figures or chart files.

*Figure 2: Example evaluation scenarios for QA, Data Visualization, and File Generation.*
### Task Complexity
Tasks are stratified by the number of expert-level reasoning steps required:
- **Easy (29.5%)**: ≤ 6 steps
- **Medium (49.4%)**: 7–12 steps
- **Hard (21.1%)**: ≥ 13 steps
- **Cross-file Reasoning**: 27.4% of tasks require reasoning over multiple input files (up to 14 files).
### Data Formats
Most inputs are tabular files (xlsx/csv dominate), complemented by **DOCX** and **PDF** formats to support mixed-type document processing.
## Evaluation Framework
All models are evaluated under a unified **tool-augmented protocol**: the model receives task instructions and associated files, and can execute **arbitrary Python code** within a **sandboxed environment** to complete the task.
To align with task categories, AIDABench uses three dedicated **LLM-based evaluators**:
1. **QA Evaluator**
A binary judge that determines whether the produced answer matches the reference (under the benchmark’s scoring rules).
2. **Visualization Evaluator**
Scores both **correctness** and **readability** of generated visualizations.
3. **Spreadsheet File Evaluator**
Verifies generated spreadsheet outputs with a **coarse-to-fine** strategy, combining structural checks with sampled content validation and task-specific verification.

*Figure 3: The design of the three types of evaluators in AIDABench.*
## Baseline Performance
Results indicate that complex, tool-augmented document analytics remains challenging: the best-performing baseline model (**Claude-Sonnet-4.5**) achieves **59.43 pass@1** on AIDABench (see the paper for full settings, model list, and breakdowns).
## Intended Uses
AIDABench is intended for:
- Evaluating **agents** or **tool-using LLM systems** on realistic document analytics tasks
- Benchmarking end-to-end capabilities across **QA**, **file generation**, and **visualization**
- Diagnosing failure modes in multi-step, multi-file reasoning over business-like data
## Limitations
- The benchmark is designed for tool-augmented settings; purely text-only inference may underperform due to the need for code execution and file manipulation.
- Automated evaluation relies on LLM judges, which introduces additional compute cost and (small) scoring variance depending on settings.
## Citation
If you use this dataset, please cite the original paper:
```bibtex
@article{yang2026aidabench,
title={AIDABENCH: AI DATA ANALYTICS BENCHMARK},
author={Yang, Yibo and Lei, Fei and Sun, Yixuan and others},
journal={arXiv preprint},
year={2026}
}
```
---
|