File size: 4,409 Bytes
630793a
 
302994a
30118fd
 
 
 
 
 
 
 
 
 
 
630793a
30118fd
630793a
30118fd
630793a
30118fd
302994a
30118fd
302994a
30118fd
302994a
30118fd
302994a
 
 
30118fd
 
 
302994a
 
 
 
 
 
 
 
 
 
 
 
 
 
30118fd
630793a
302994a
 
 
ef25578
 
 
302994a
30118fd
630793a
302994a
 
30118fd
 
 
302994a
 
 
30118fd
302994a
 
30118fd
302994a
 
 
 
 
 
 
 
 
30118fd
630793a
302994a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30118fd
 
 
 
 
 
 
 
 
 
 
 
fa395e5
302994a
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
---
language:
- zh
license: apache-2.0
task_categories:
- question-answering
- tabular-classification
- text-generation
tags:
- data-analytics
- agents
- document-understanding
- benchmark
pretty_name: AIDABench
---

# Dataset Card for AIDABench

## Dataset Summary

**AIDABench** is a benchmark for evaluating AI systems on **end-to-end data analytics over real-world documents**. It contains **600+** diverse analytical tasks grounded in realistic scenarios and spans heterogeneous data sources such as **spreadsheets, databases, financial reports, and operational records**. Tasks are designed to be challenging, often requiring multi-step reasoning and tool use to complete reliably.

![Overview of AIDABench Framework](images/figure1_overview.png)

*Figure 1: Overview of the AIDABench evaluation framework.*

## Supported Tasks and Evaluation Targets

AIDABench focuses on practical document analytics workflows where a model/agent must read files, reason over structured data, and produce a final deliverable.

### Task Categories

The dataset is organized around three primary capability dimensions:

- **File Generation (43.3%)**  
  Data wrangling and transformation tasks such as filtering, normalization, deduplication, joins, and cross-sheet linkage, with outputs as generated files (e.g., spreadsheets).

- **Question Answering (QA) (37.5%)**  
  Analytical queries such as aggregation, averages, ranking, comparisons, and trend analysis, with outputs as final answers.

- **Data Visualization (19.2%)**  
  Chart creation/adaptation tasks (e.g., bar/line/pie) including style requirements and presentation constraints, with outputs as figures or chart files.

![Evaluation Scenarios](images/figure2_scenarios.png)

*Figure 2: Example evaluation scenarios for QA, Data Visualization, and File Generation.*

### Task Complexity

Tasks are stratified by the number of expert-level reasoning steps required:

- **Easy (29.5%)**: ≤ 6 steps  
- **Medium (49.4%)**: 7–12 steps  
- **Hard (21.1%)**: ≥ 13 steps  
- **Cross-file Reasoning**: 27.4% of tasks require reasoning over multiple input files (up to 14 files).

### Data Formats

Most inputs are tabular files (xlsx/csv dominate), complemented by **DOCX** and **PDF** formats to support mixed-type document processing.

## Evaluation Framework

All models are evaluated under a unified **tool-augmented protocol**: the model receives task instructions and associated files, and can execute **arbitrary Python code** within a **sandboxed environment** to complete the task.

To align with task categories, AIDABench uses three dedicated **LLM-based evaluators**:

1. **QA Evaluator**  
   A binary judge that determines whether the produced answer matches the reference (under the benchmark’s scoring rules).

2. **Visualization Evaluator**  
   Scores both **correctness** and **readability** of generated visualizations.

3. **Spreadsheet File Evaluator**  
   Verifies generated spreadsheet outputs with a **coarse-to-fine** strategy, combining structural checks with sampled content validation and task-specific verification.

![Evaluator Design](images/figure3_evaluators.png)

*Figure 3: The design of the three types of evaluators in AIDABench.*

## Baseline Performance

Results indicate that complex, tool-augmented document analytics remains challenging: the best-performing baseline model (**Claude-Sonnet-4.5**) achieves **59.43 pass@1** on AIDABench (see the paper for full settings, model list, and breakdowns).

## Intended Uses

AIDABench is intended for:

- Evaluating **agents** or **tool-using LLM systems** on realistic document analytics tasks
- Benchmarking end-to-end capabilities across **QA**, **file generation**, and **visualization**
- Diagnosing failure modes in multi-step, multi-file reasoning over business-like data

## Limitations

- The benchmark is designed for tool-augmented settings; purely text-only inference may underperform due to the need for code execution and file manipulation.
- Automated evaluation relies on LLM judges, which introduces additional compute cost and (small) scoring variance depending on settings.

## Citation

If you use this dataset, please cite the original paper:

```bibtex
@article{yang2026aidabench,
  title={AIDABENCH: AI DATA ANALYTICS BENCHMARK},
  author={Yang, Yibo and Lei, Fei and Sun, Yixuan and others},
  journal={arXiv preprint},
  year={2026}
}
```
---