File size: 4,184 Bytes
b94c634
 
 
 
 
bde7a67
135d9a9
58bbd53
171c86d
58bbd53
 
 
 
 
af052a1
58bbd53
e9fbf62
 
 
 
 
 
 
 
 
 
 
135d9a9
58bbd53
135d9a9
 
e9fbf62
 
 
 
 
 
 
 
 
 
 
 
 
58bbd53
135d9a9
 
e9fbf62
 
 
 
 
 
 
 
 
 
 
 
 
135d9a9
58bbd53
135d9a9
 
e9fbf62
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
135d9a9
 
58bbd53
135d9a9
 
 
58bbd53
135d9a9
 
e9fbf62
 
 
 
171c86d
 
 
 
 
 
 
 
 
e9fbf62
b94c634
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
---
task_categories:
- table-question-answering
---

# MMSci_Table


Dataset for the paper "[Does Table Source Matter? Benchmarking and Improving Multimodal Scientific Table Understanding and Reasoning](https://arxiv.org/abs/2501.13042)"
<p align="center">

<p>

<p align="center">
       📑 <a href="https://arxiv.org/abs/2501.13042">Paper</a> &nbsp&nbsp   <a href="https://github.com/Bernard-Yang/MMSci_Table">Github</a>  </a>  
</p>

# MMSci Dataset Collection

The MMSci dataset collection consists of three complementary datasets designed for scientific multimodal table understanding and reasoning: MMSci-Pre, MMSci-Ins, and MMSci-Eval.

## Dataset Summary

- **MMSci-Pre**: A domain-specific pre-training dataset containing 52K scientific table structure recognition samples
- **MMSci-Ins**: An instruction tuning dataset with 12K samples across three table-based tasks
- **MMSci-Eval**: A benchmark with 3,114 testing samples for numerical reasoning evaluation

## Framework Overview
![Framework Overview](./assets/model.jpg)
*Figure 1: Overview of the MMSci framework showing the four key stages: Table Image Generation, Dataset Construction, Table Structure Learning, and Visual Instruction Tuning.*

## Dataset Details

### MMSci-Pre
- **Size**: 52K samples
- **Format**: Table image-to-HTML pairs
- **Source**: Scientific papers from SciGen dataset
- **Purpose**: Table structure learning and alignment of visual features with textual representations
- **Features**:
  - High-quality HTML format tables
  - Rendered table images preserving structural integrity
  - Complex layouts and relationships from scientific papers
  - Focus on tables with significant numerical values

![MMSci-Pre Example](./assets/html1.jpg)
*Figure 2: Example from MMSci-Pre dataset showing the table image and its corresponding HTML representation.*

### MMSci-Ins
- **Size**: 12K samples
- **Format**: Instruction-following samples with reasoning steps
- **Tasks**:
  - Table Question Answering (TQA)
  - Table Fact Verification (TFV)
  - Table-to-Text Generation (T2T)
- **Features**:
  - Detailed step-by-step reasoning processes
  - Balanced distribution across three tasks
  - Each table paired with one TQA, TFV, and T2T task
  - Built upon scientific domain tables


![MMSci-Ins Example](./assets/input1.jpg)
*Figure 3: Example from MMSci-Ins dataset showing instruction-following samples across different tasks.*

### MMSci-Eval
- **Size**: 3,114 samples
- **Purpose**: Comprehensive evaluation of numerical reasoning capabilities
- **Features**:
  - Testing samples across TQA, TFV, and T2T tasks
  - Focus on numerical reasoning assessment
  - Based on SciGen dataset test set
  - Diverse reasoning types and complexity levels

## Dataset Creation

The datasets were created through a rigorous process:
1. Collection of raw tabular data from SciGen dataset
2. Transformation of textual tables into HTML format
3. Rendering of HTML tables into high-quality images
4. Generation of instruction-following samples with reasoning steps
5. Quality assurance through balanced task distribution

## Intended Uses

- Pre-training multimodal language models for table understanding
- Fine-tuning models for specific table-based tasks
- Evaluating numerical reasoning capabilities in scientific contexts
- Benchmarking table understanding and reasoning systems


#### Table Question Answering (TQA)
![TQA Example](./assets/tqa1.jpg)
*Figure 4: Example of a TQA task showing the question, reasoning steps, and answer.*

#### Table Fact Verification (TFV)
![TFV Example](./assets/tfv1.jpg)
*Figure 5: Example of a TFV task showing the statement, verification process, and conclusion.*

## Citation

If you found this repository or paper is helpful to you, please cite our paper.
```
@misc{yang2025doestablesourcematter,
      title={Does Table Source Matter? Benchmarking and Improving Multimodal Scientific Table Understanding and Reasoning}, 
      author={Bohao Yang and Yingji Zhang and Dong Liu and André Freitas and Chenghua Lin},
      year={2025},
      eprint={2501.13042},
      archivePrefix={arXiv},
      primaryClass={cs.CL},
      url={https://arxiv.org/abs/2501.13042}, 
}
```
}