|
|
--- |
|
|
dataset_info: |
|
|
features: |
|
|
- name: prompt |
|
|
dtype: string |
|
|
- name: metadata |
|
|
dtype: string |
|
|
- name: task |
|
|
dtype: string |
|
|
- name: dataset |
|
|
dtype: string |
|
|
splits: |
|
|
- name: train |
|
|
num_bytes: 432885480 |
|
|
num_examples: 30647 |
|
|
download_size: 132698519 |
|
|
dataset_size: 432885480 |
|
|
configs: |
|
|
- config_name: default |
|
|
data_files: |
|
|
- split: train |
|
|
path: data/train-* |
|
|
task_categories: |
|
|
- table-question-answering |
|
|
--- |
|
|
|
|
|
# Dataset Card for MMTU |
|
|
|
|
|
## Dataset Summary |
|
|
|
|
|
MMTU: A Massive Multi-Task Table Understanding and Reasoning Benchmark |
|
|
by Junjie Xing, [Yeye He](https://www.microsoft.com/en-us/research/people/yeyehe/), Mengyu Zhou, Haoyu Dong, Shi Han, Lingjiao Chen, Dongmei Zhang, [Surajit Chaudhuri](https://www.microsoft.com/en-us/research/people/surajitc/), and [H. V. Jagadish](https://web.eecs.umich.edu/~jag/). |
|
|
|
|
|
[Paper](https://huggingface.co/papers/2506.05587) |
|
|
[Code](https://github.com/MMTU-Benchmark/MMTU) |
|
|
|
|
|
This is a large-scale benchmark designed to evaluate the table reasoning capabilities of large language models (LLMs). It consists of over 30,000 questions across 25 real-world table tasks, focusing on deep understanding, reasoning, and manipulation of tabular data. |
|
|
|
|
|
These tasks are curated from decades of computer science research and represent challenges encountered by expert users in real applications, making MMTU a rigorous test for LLMs aspiring to professional-level table understanding. |
|
|
|
|
|
A complete list of tasks: 'table-transform-by-relationalization', 'table-transform-by-output-schema', 'table-transform-by-output-table', 'Entity matching', 'Schema matching', 'Head value matching', 'data-imputation', 'error-detection', 'list-to-table', 'semantic-join', 'equi-join-detect', 'program-transform-by-example', 'formula-by-context', 'semantic-transform-by-example', 'arithmetic-relationship', 'functional-relationship', 'string-relationship', 'Needle-in-a-haystack-table', 'Needle-in-a-haystack-index', 'NL-2-SQL', 'Table Question Answering', 'Fact Verification', 'Column type annotation', 'Column property annotation', 'Cell entity annotation'. |
|
|
|
|
|
## Leaderboards |
|
|
|
|
|
| **Model Type** | **Model** | **MMTU Score** | |
|
|
|----------------|---------------------|----------------------| |
|
|
| Reasoning | o4-mini | **0.637 ± 0.01** | |
|
|
| Reasoning | Deepseek-R1 | 0.557 ± 0.01 | |
|
|
| Chat | Deepseek-V3 | 0.517 ± 0.01 | |
|
|
| Chat | GPT-4o | 0.490 ± 0.01 | |
|
|
| Chat | Llama-3.3-70B | 0.438 ± 0.01 | |
|
|
| Chat | Mistral-Large | 0.430 ± 0.01 | |
|
|
| Chat | Mistral-Small | 0.402 ± 0.01 | |
|
|
| Chat | GPT-4o-mini | 0.386 ± 0.01 | |
|
|
| Chat | Llama-3.1-8B | 0.259 ± 0.01 | |
|
|
|
|
|
## Language |
|
|
|
|
|
English |
|
|
|
|
|
## Data Structure |
|
|
|
|
|
### Data Fields |
|
|
|
|
|
- prompt: The prompt presented in the MMTU instance. |
|
|
- metadata: Supplementary information associated with the MMTU instance, typically used for evaluation purposes. |
|
|
- task: The specific subtask category within the MMTU framework to which the instance belongs. |
|
|
- dataset: The original source dataset from which the MMTU instance is derived. |
|
|
|
|
|
## Dataset Creation |
|
|
|
|
|
Please refer to Section 3.2 in the paper. |