File size: 2,304 Bytes
00657c3 d5ac2dd 9e58469 9eb9344 9e58469 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 |
---
license: cc-by-3.0
task_categories:
- text-generation
language:
- en
tags:
- agent
- finance
- multimodal
- spreadsheet
- workflow
configs:
- config_name: Finch_Dataset_All
data_files:
- split: valid
path:
- Finch_Dataset_valid.jsonl
- split: test
path:
- Finch_Dataset_test.jsonl
---
# Finch

This repository contains the dataset for **Finch**, an enterprise-level benchmark for evaluating an agent’s ability to act like a skilled finance & accounting expert on real-world workflows.
* **Paper**: _to be added_
* **Project Page**: https://huggingface.co/datasets/FinWorkBench/Finch
* **Code**: https://github.com/FinWorkBench
---
## Dataset Description
Finch focuses on **composite finance & accounting workflows** that span:
> data entry/import, structuring/formatting, web search, cross-sheet/file retrieval, calculation, financial modeling, validation, translation, visualization, and reporting.
The workflows are derived from **real-world enterprise workspaces**, including:
- Large and messy **spreadsheets** with multimodal artifacts including text, tables, formulas, charts, pivots, images, etc
- Linked **PDFs and documents** that provide additional business context
We adopt a three-step workflow labeling process:
1. **Summarizing workflow types** supported by real collaborative enterprise email threads.
2. **Deriving concrete workflow instances** from versioned spreadsheets and related files.
3. **Meticulous expert annotation** of instructions and reference outputs.
This process yields **172 enterprise-grade single- and multi-task workflows** with carefully written instructions and aligned input/reference files, capturing the intrinsic **complexity, messiness, and multimodality** of real-world finance & accounting work.
---
## Dataset Structure
The instruction-tuning corpus is released in **JSONL** format and follows the standard **LLaMA Factory Alpaca** schema.
Each line is one **workflow-centric example**:
```json
{
"instruction": "Task instruction for a specific finance & accounting workflow (e.g., 'Reconcile the cash balance and explain the variance using the provided spreadsheet tabs.').",
"input": _to be added_,
"output": _to be added_
}
|