ShuoChen99 commited on
Commit
11ee365
·
verified ·
1 Parent(s): bf08f6f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +196 -1
README.md CHANGED
@@ -9,4 +9,199 @@ tags:
9
  - MLLMs
10
  - in-context-learning
11
  - vision-language
12
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  - MLLMs
10
  - in-context-learning
11
  - vision-language
12
+ ---
13
+
14
+ # TrueMICL: True Multimodal In-Context Learning Dataset
15
+
16
+ A comprehensive multimodal dataset designed to evaluate and improve true multimodal in-context learning capabilities in Multimodal Large Language Models (MLLMs).
17
+
18
+ ## Table of Contents
19
+ - [Dataset Overview](#dataset-overview)
20
+ - [Dataset Structure](#dataset-structure)
21
+ - [Tasks and Domains](#tasks-and-domains)
22
+ - [Usage Examples](#usage-examples)
23
+ - [Data Collection Methodology](#data-collection-methodology)
24
+ - [Citation](#citation)
25
+ - [License](#license)
26
+ - [Contact](#contact)
27
+
28
+ ## Dataset Overview
29
+
30
+ TrueMICL addresses a critical limitation in current Multimodal Large Language Models: their tendency to neglect visual information in multimodal demonstrations, leading to superficial text imitation. This dataset is specifically designed to test **true** multimodal in-context learning by ensuring that:
31
+
32
+ - Tasks are unsolvable without visual context
33
+ - Novel image-text relationships are introduced
34
+ - Visual information is perceivable and critical
35
+ - Compatibility with language model backbones is maintained
36
+
37
+ ### Key Statistics
38
+ - **Total samples**: 867 evaluation samples + extensive training data
39
+ - **Task categories**: 4 major categories
40
+ - **Distinct tasks**: 7 different tasks
41
+ - **Domains**: Mathematical reasoning, pattern recognition, concept learning, visual question answering
42
+
43
+ ## Dataset Structure
44
+
45
+ The dataset is organized into task-specific directories, each containing:
46
+
47
+ ### File Organization
48
+ ```
49
+ dataset/
50
+ ├── classification/ # Character classification task
51
+ │ ├── img/ # Query and support images
52
+ │ ├── query.json # Test queries (656 samples)
53
+ │ └── support.json # Support examples (128 samples)
54
+ ├── clevr/ # CLEVR-based reasoning tasks
55
+ │ ├── material/ # Material-based images
56
+ │ ├── query/ # Query images
57
+ │ ├── shape/ # Shape-based images
58
+ │ ├── size/ # Size-based images
59
+ │ ├── support/ # Support images
60
+ │ ├── query.json # Main queries (368 samples)
61
+ │ ├── support.json # Support examples (1568 samples)
62
+ │ └── [query/support]_[material/shape/size].json # Task-specific splits
63
+ ├── clock/ # Clock reading and math
64
+ │ ├── img/ # Clock face images
65
+ │ ├── query.json # Test queries (900 samples)
66
+ │ └── support.json # Support examples (192 samples)
67
+ ├── operator_induction/ # Mathematical operator learning
68
+ │ ├── query.json # Test queries (900 samples)
69
+ │ ├── support.json # Support examples (192 samples)
70
+ │ └── processed_training_data.json # Training data (6300 samples)
71
+ ├── palindrome_dataset/ # Palindrome pattern recognition
72
+ │ ├── query.json # Test queries (800 samples)
73
+ │ ├── support.json # Support examples (256 samples)
74
+ │ └── training_data.json # Training data (3150 samples)
75
+ ├── shapes_count/ # Shape counting task
76
+ │ ├── query.json # Test queries (900 samples)
77
+ │ ├── support.json # Support examples (288 samples)
78
+ │ └── training_data.json # Training data (3150 samples)
79
+ ├── sudoku/ # Sudoku puzzle solving
80
+ │ ├── query.json # Test queries (800 samples)
81
+ │ └── support.json # Support examples (128 samples)
82
+ └── vqav2/ # Visual Question Answering v2
83
+ ├── query.json # Test queries (157,680 samples)
84
+ └── support.json # Support examples (2,662,542 samples)
85
+ ```
86
+
87
+ ### Data Format
88
+
89
+ Each JSON file contains structured data with the following schema:
90
+
91
+ **Query/Support Format**:
92
+ ```json
93
+ {
94
+ "id": "unique_identifier",
95
+ "image": ["path/to/image.png"],
96
+ "question": "Question text with multiple choice options",
97
+ "answer": "Correct answer"
98
+ }
99
+ ```
100
+
101
+ **VQA Format** (slightly different):
102
+ ```json
103
+ {
104
+ "image_id": 12345,
105
+ "question_id": 67890,
106
+ "question": "Question text",
107
+ "answer": "Answer text"
108
+ }
109
+ ```
110
+
111
+ ### Data Types and Columns
112
+
113
+ | Field | Type | Description |
114
+ |-------|------|-------------|
115
+ | `id` | string | Unique identifier for the sample |
116
+ | `image` | array | List of image file paths |
117
+ | `question` | string | Question or task description |
118
+ | `answer` | string | Ground truth answer |
119
+ | `image_id` | integer | Image identifier (VQA format) |
120
+ | `question_id` | integer | Question identifier (VQA format) |
121
+
122
+ ## Tasks and Domains
123
+
124
+ ### 1. Mathematical Reasoning
125
+ - **Operator Induction**: Learn novel mathematical operators from visual examples
126
+ - **Clock Math**: Time reading and calculation tasks
127
+
128
+ ### 2. Concept Binding
129
+ - **Character Classification**: Classify novel character types from visual examples
130
+ - **CLEVR Count**: Object counting and attribute reasoning
131
+
132
+ ### 3. Pattern Finding
133
+ - **Sudoku**: Complete Sudoku puzzles using visual pattern recognition
134
+ - **Palindrome**: Identify palindromic patterns in visual sequences
135
+
136
+ ### 4. Novel Concept Learning
137
+ - **Shapes Count**: Count specific shapes and understand spatial relationships
138
+ - **VQA**: General visual question answering requiring multimodal reasoning
139
+
140
+ ## Usage Examples
141
+
142
+ ### Basic Data Exploration
143
+
144
+ ```python
145
+ import json
146
+ import matplotlib.pyplot as plt
147
+ from PIL import Image
148
+
149
+ # Load and examine a sample
150
+ with open("classification/query.json", "r") as f:
151
+ data = json.load(f)
152
+
153
+ sample = data[0]
154
+ print(f"ID: {sample['id']}")
155
+ print(f"Question: {sample['question']}")
156
+ print(f"Answer: {sample['answer']}")
157
+
158
+ # Load and display the image
159
+ img_path = sample['image'][0]
160
+ img = Image.open(img_path)
161
+ plt.imshow(img)
162
+ plt.title(sample['question'])
163
+ plt.show()
164
+ ```
165
+
166
+ ### Task-Specific Loading
167
+
168
+ ```python
169
+ # Load CLEVR subtasks
170
+ clevr_tasks = ['material', 'shape', 'size']
171
+ for task in clevr_tasks:
172
+ with open(f"clevr/query_{task}.json", "r") as f:
173
+ task_data = json.load(f)
174
+ print(f"CLEVR {task}: {len(task_data)} samples")
175
+ ```
176
+
177
+ ## Data Collection Methodology
178
+
179
+ The dataset was constructed following rigorous criteria to ensure true multimodal learning:
180
+
181
+ 1. **Visual Dependency**: All tasks require visual information and cannot be solved through text-only reasoning
182
+ 2. **Novel Relationships**: Introduction of previously unseen image-text mappings
183
+ 3. **Perceptual Validity**: Visual elements are clearly perceivable and unambiguous
184
+ 4. **Model Compatibility**: Designed to work with standard language model architectures
185
+
186
+ ### Source Data
187
+ - **CLEVR**: Modified from the original CLEVR dataset for visual reasoning
188
+ - **VQAv2**: Subset of the Visual Question Answering v2 dataset
189
+ - **Synthetic Tasks**: Custom-generated tasks for operator induction, palindromes, and shape counting
190
+ - **Novel Concepts**: Artificially created character types and visual patterns
191
+
192
+ ## License
193
+
194
+ This dataset is released under the [MIT License](LICENSE). Please see the license file for detailed terms and conditions.
195
+
196
+
197
+ ## Contact
198
+
199
+ For questions, issues, or contributions regarding this dataset:
200
+
201
+ - **Project Website**: https://chenxshuo.github.io/true-micl-colm/
202
+ - **Paper**: Available at the project website
203
+ - **Issues**: Please report bugs or request features through the appropriate channels
204
+
205
+ ---
206
+
207
+ **Note**: This dataset is designed for research purposes to advance multimodal in-context learning. The novel tasks and visual concepts are specifically crafted to test true multimodal understanding rather than superficial pattern matching.