AIDA / README.md
MichaelYang-lyx's picture
Update README.md
ef25578 verified
---
language:
- zh
license: apache-2.0
task_categories:
- question-answering
- tabular-classification
- text-generation
tags:
- data-analytics
- agents
- document-understanding
- benchmark
pretty_name: AIDABench
---
# Dataset Card for AIDABench
## Dataset Summary
**AIDABench** is a benchmark for evaluating AI systems on **end-to-end data analytics over real-world documents**. It contains **600+** diverse analytical tasks grounded in realistic scenarios and spans heterogeneous data sources such as **spreadsheets, databases, financial reports, and operational records**. Tasks are designed to be challenging, often requiring multi-step reasoning and tool use to complete reliably.
![Overview of AIDABench Framework](images/figure1_overview.png)
*Figure 1: Overview of the AIDABench evaluation framework.*
## Supported Tasks and Evaluation Targets
AIDABench focuses on practical document analytics workflows where a model/agent must read files, reason over structured data, and produce a final deliverable.
### Task Categories
The dataset is organized around three primary capability dimensions:
- **File Generation (43.3%)**
Data wrangling and transformation tasks such as filtering, normalization, deduplication, joins, and cross-sheet linkage, with outputs as generated files (e.g., spreadsheets).
- **Question Answering (QA) (37.5%)**
Analytical queries such as aggregation, averages, ranking, comparisons, and trend analysis, with outputs as final answers.
- **Data Visualization (19.2%)**
Chart creation/adaptation tasks (e.g., bar/line/pie) including style requirements and presentation constraints, with outputs as figures or chart files.
![Evaluation Scenarios](images/figure2_scenarios.png)
*Figure 2: Example evaluation scenarios for QA, Data Visualization, and File Generation.*
### Task Complexity
Tasks are stratified by the number of expert-level reasoning steps required:
- **Easy (29.5%)**: ≤ 6 steps
- **Medium (49.4%)**: 7–12 steps
- **Hard (21.1%)**: ≥ 13 steps
- **Cross-file Reasoning**: 27.4% of tasks require reasoning over multiple input files (up to 14 files).
### Data Formats
Most inputs are tabular files (xlsx/csv dominate), complemented by **DOCX** and **PDF** formats to support mixed-type document processing.
## Evaluation Framework
All models are evaluated under a unified **tool-augmented protocol**: the model receives task instructions and associated files, and can execute **arbitrary Python code** within a **sandboxed environment** to complete the task.
To align with task categories, AIDABench uses three dedicated **LLM-based evaluators**:
1. **QA Evaluator**
A binary judge that determines whether the produced answer matches the reference (under the benchmark’s scoring rules).
2. **Visualization Evaluator**
Scores both **correctness** and **readability** of generated visualizations.
3. **Spreadsheet File Evaluator**
Verifies generated spreadsheet outputs with a **coarse-to-fine** strategy, combining structural checks with sampled content validation and task-specific verification.
![Evaluator Design](images/figure3_evaluators.png)
*Figure 3: The design of the three types of evaluators in AIDABench.*
## Baseline Performance
Results indicate that complex, tool-augmented document analytics remains challenging: the best-performing baseline model (**Claude-Sonnet-4.5**) achieves **59.43 pass@1** on AIDABench (see the paper for full settings, model list, and breakdowns).
## Intended Uses
AIDABench is intended for:
- Evaluating **agents** or **tool-using LLM systems** on realistic document analytics tasks
- Benchmarking end-to-end capabilities across **QA**, **file generation**, and **visualization**
- Diagnosing failure modes in multi-step, multi-file reasoning over business-like data
## Limitations
- The benchmark is designed for tool-augmented settings; purely text-only inference may underperform due to the need for code execution and file manipulation.
- Automated evaluation relies on LLM judges, which introduces additional compute cost and (small) scoring variance depending on settings.
## Citation
If you use this dataset, please cite the original paper:
```bibtex
@article{yang2026aidabench,
title={AIDABENCH: AI DATA ANALYTICS BENCHMARK},
author={Yang, Yibo and Lei, Fei and Sun, Yixuan and others},
journal={arXiv preprint},
year={2026}
}
```
---