yangbh217 commited on
Commit
e9fbf62
·
verified ·
1 Parent(s): 5a5f20f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +74 -3
README.md CHANGED
@@ -1,3 +1,74 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+ # Does Table Source Matter? Benchmarking and Improving Multimodal Scientific Table Understanding and Reasoning
5
+ Dataset for the paper "[Does Table Source Matter? Benchmarking and Improving Multimodal Scientific Table Understanding and Reasoning](s)"
6
+
7
+ # MMSci Dataset Collection
8
+
9
+ The MMSci dataset collection consists of three complementary datasets designed for scientific multimodal table understanding and reasoning: MMSci-Pre, MMSci-Ins, and MMSci-Eval.
10
+
11
+ ## Dataset Summary
12
+
13
+ - **MMSci-Pre**: A domain-specific pre-training dataset containing 52K scientific table structure recognition samples
14
+ - **MMSci-Ins**: An instruction tuning dataset with 12K samples across three table-based tasks
15
+ - **MMSci-Eval**: A benchmark with 3,114 testing samples for numerical reasoning evaluation
16
+
17
+ ## Dataset Details
18
+
19
+ ### MMSci-Pre
20
+ - **Size**: 52K samples
21
+ - **Format**: Table image-to-HTML pairs
22
+ - **Source**: Scientific papers from SciGen dataset
23
+ - **Purpose**: Table structure learning and alignment of visual features with textual representations
24
+ - **Features**:
25
+ - High-quality HTML format tables
26
+ - Rendered table images preserving structural integrity
27
+ - Complex layouts and relationships from scientific papers
28
+ - Focus on tables with significant numerical values
29
+
30
+ ### MMSci-Ins
31
+ - **Size**: 12K samples
32
+ - **Format**: Instruction-following samples with reasoning steps
33
+ - **Tasks**:
34
+ - Table Question Answering (TQA)
35
+ - Table Fact Verification (TFV)
36
+ - Table-to-Text Generation (T2T)
37
+ - **Features**:
38
+ - Detailed step-by-step reasoning processes
39
+ - Balanced distribution across three tasks
40
+ - Each table paired with one TQA, TFV, and T2T task
41
+ - Built upon scientific domain tables
42
+
43
+ ### MMSci-Eval
44
+ - **Size**: 3,114 samples
45
+ - **Purpose**: Comprehensive evaluation of numerical reasoning capabilities
46
+ - **Features**:
47
+ - Testing samples across TQA, TFV, and T2T tasks
48
+ - Focus on numerical reasoning assessment
49
+ - Based on SciGen dataset test set
50
+ - Diverse reasoning types and complexity levels
51
+
52
+ ## Dataset Creation
53
+
54
+ The datasets were created through a rigorous process:
55
+ 1. Collection of raw tabular data from SciGen dataset
56
+ 2. Transformation of textual tables into HTML format
57
+ 3. Rendering of HTML tables into high-quality images
58
+ 4. Generation of instruction-following samples with reasoning steps
59
+ 5. Quality assurance through balanced task distribution
60
+
61
+ ## Intended Uses
62
+
63
+ - Pre-training multimodal language models for table understanding
64
+ - Fine-tuning models for specific table-based tasks
65
+ - Evaluating numerical reasoning capabilities in scientific contexts
66
+ - Benchmarking table understanding and reasoning systems
67
+
68
+ ## Citation
69
+
70
+ If you found this repository or paper is helpful to you, please cite our paper.
71
+ ```
72
+
73
+ ```
74
+ }