File size: 3,441 Bytes
00657c3 2b8baed d5ac2dd 1d66725 2b8baed 1a0139a 9e58469 adf575a 0e9fcfe 9e58469 9eb9344 9e58469 1d66725 4092b6c 9e58469 8bab4f7 9e58469 8bab4f7 9e58469 8bab4f7 9e58469 8bab4f7 9e58469 88559b9 1d66725 9e58469 9a0ec9b 9e58469 b5a79a9 9e58469 c3f22fe 9e58469 c3f22fe 9e58469 77975dc cd11190 77975dc cd11190 9e58469 3378f62 8bab4f7 3378f62 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 |
---
license: cc-by-3.0
tags:
- agent
- workflow
- multimodal
- spreadsheet
- pdf
- image
- code
- finance
- accouning
modalities:
- text
- spreadsheet
- pdf
- image
- code
configs:
- config_name: Finch_Dataset_All
data_files:
- split: test
path:
- finch_workflows_test.jsonl
---

# Finch: Benchmarking Finance & Accounting across Spreadsheet-Centric Enterprise Workflows
This repository contains the dataset for **Finch**, an enterprise-level benchmark for evaluating an agent’s ability to act like a skilled finance & accounting expert on real-world workflows.
* **Paper**: _to be added_
* **Project Page**: https://huggingface.co/datasets/FinWorkBench/Finch
* **Code**: https://github.com/FinWorkBench
---
## Dataset Description
Finch focuses on **messy and long-horizon finance & accounting workflows** that span:
> data entry/import, structuring/formatting, web search, cross-sheet/file retrieval, calculation, financial modeling, validation, translation, visualization, and reporting.
The workflows are derived from **real-world enterprise workspaces** (primarily Enron, as well as World Bank, Canadian/Australian government agencies, and other corporations), including:
- Enterprise **email threads** where collaborators naturally describe, discuss, and track workflows
- Large and messy **spreadsheets** with multimodal artifacts including text, tables, formulas, charts, pivots, images, etc
- Interlinked **PDFs and documents** that provide additional business context
We adopt a three-step workflow labeling process:
1. **Inducing workflow types** from real collaborative context in enterprise email threads.
2. **Deriving concrete workflow instances** by analyzing changes across spreadsheet versions.
3. **Conductin meticulous expert annotation** of task instructions, input files, and reference outputs, involving hundreds of hours of expert work.
This process yields **172 enterprise-grade workflows—primarily multi-task composite** — each with carefully written instructions and aligned input/reference files, capturing the intrinsic **compositional, messy, multimodal, and collaborative nature** of real-world finance & accounting work. In this release, we provide full annotations for the first 72 workflows, with the remaining 100 to be released in a subsequent update.
Experiment results show that even frontier agents solve fewer than 30% of the workflows, revealing a substantial performance gap for real-world enterprise scenarios.
---
## 📁 Dataset Structure
The instruction-tuning corpus is released in **JSONL** format.
Each line corresponds to one **workflow-centric example**:
```json
{
"id": "<workflow identifier>",
"instruction_en": "<English task instruction for a finance & accounting workflow>",
"source_files": ["<input file name>", "..."],
"source_files_urls": ["<input file download URL>", "..."],
"reference_outputs": {
"files": ["<reference output file name>"],
"text": "<textual reference output>"
},
"reference_file_urls": ["<reference output file download URL>"],
"task_type": "<task category (e.g., reporting, modeling)>",
"business_type": "<business domain (e.g., budgeting, trading)>"
}
```
---
## 📣 Feedback & Issues
If you find any issues with the dataset or have suggestions, please open a discussion in the **Community** tab — we value your feedback!
|