File size: 5,695 Bytes
4fb1a00
 
5745473
4fb1a00
5745473
 
4fb1a00
5745473
4fb1a00
5745473
 
 
 
 
 
4fb1a00
5745473
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4fb1a00
 
5745473
4fb1a00
 
 
5745473
4fb1a00
5745473
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4fb1a00
 
 
5745473
4fb1a00
 
 
 
5745473
4fb1a00
5745473
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
---
license: mit
pretty_name: MMIB Evaluation Dataset
task_categories:
- visual-question-answering
- multiple-choice
language:
- en
tags:
- vision
- language
- multimodal
- counterfactual
- mechanistic-interpretability
- synthetic
size_categories:
- n<1K
dataset_info:
  features:
  - name: original_image
    dtype: image
  - name: counterfactual1_image
    dtype: image
  - name: counterfactual2_image
    dtype: image
  - name: counterfactual1_type
    dtype: string
  - name: counterfactual2_type
    dtype: string
  - name: counterfactual1_description
    dtype: string
  - name: counterfactual2_description
    dtype: string
  - name: original_question
    dtype: string
  - name: counterfactual1_question
    dtype: string
  - name: counterfactual2_question
    dtype: string
  - name: original_question_difficulty
    dtype: string
  - name: counterfactual1_question_difficulty
    dtype: string
  - name: counterfactual2_question_difficulty
    dtype: string
  - name: original_image_answer_to_original_question
    dtype: string
  - name: original_image_answer_to_cf1_question
    dtype: string
  - name: original_image_answer_to_cf2_question
    dtype: string
  - name: cf1_image_answer_to_original_question
    dtype: string
  - name: cf1_image_answer_to_cf1_question
    dtype: string
  - name: cf1_image_answer_to_cf2_question
    dtype: string
  - name: cf2_image_answer_to_original_question
    dtype: string
  - name: cf2_image_answer_to_cf1_question
    dtype: string
  - name: cf2_image_answer_to_cf2_question
    dtype: string
configs:
- config_name: default
  data_files:
  - split: train
    path: data/train-*
---

# Multimodal Mechanistic Interpretability Benchmark (MMIB) Dataset



The **MMIB Dataset** is a highly controlled, synthetic vision-language dataset designed to rigorously evaluate mechanistic interpretability (MI) methods in Large Multimodal Models (VLMs). Built upon procedurally generated CLEVR-style assets, this dataset provides exact ground-truth causal pathways to test whether MI techniques (like causal tracing or interchange interventions) localize genuine cognitive circuits or merely identify descriptive correlations.

Unlike standard VQA benchmarks, MMIB uses strict **automated rejection sampling** to eliminate geometric ambiguity, ensuring every spatial and causal relationship is mathematically verifiable. 

## Dataset Structure & Interventions

This dataset is built on a structured intervention triplet for every base scene. Each row provides a complete $3 \times 3$ cross-modal evaluation matrix (3 images $\times$ 3 text queries), allowing researchers to systematically trace cross-modal information flow.

### 1. Semantic Counterfactuals (Causal Reasoning)
To evaluate the model's internal causal logic, we generate minimal counterfactual pairs where the intervention mathematically guarantees a change in the ground-truth answer ($y' \neq y$).
* **Image-Based CFs:** 10 targeted 3D scene graph edits (e.g., `change_color`, `change_position`, `relational_flip`) that alter the visual logic while keeping the question fixed.
* **Text-Based CFs:** Minimal deterministic mutations to the textual query (e.g., swapping "red" for "blue" or "left" for "right") that guarantee an answer flip on the fixed base image. 

### 2. Negative Counterfactuals (Diagnostic Stress Tests)
To control for basic visual fragility, we generate **Negative Counterfactuals** featuring 8 types of perceptual corruptions (e.g., `add_noise`, `change_lighting`, `apply_fisheye`). These interventions drastically alter the image distribution *without* changing the underlying 3D geometry or ground-truth answer ($y' = y$). They serve as an experimental baseline: if a model fails on these stress tests, its failure on semantic tasks indicates vulnerability to domain shifts rather than flawed causal logic.

## Using the Dataset

### Loading from Python

The dataset is hosted in standard Parquet format. You can load it directly into your mechanistic evaluation pipeline using the Hugging Face `datasets` library:

```python
from datasets import load_dataset

# Load the MMIB dataset
ds = load_dataset("scholo/MMB_dataset", split="train")

# Inspect the 3x3 evaluation matrix for the first scene
print("Base Question:", ds[0]['original_question'])
print("Base Image -> Base Question Answer:", ds[0]['original_image_answer_to_original_question'])
print("Semantic CF Image -> Base Question Answer:", ds[0]['cf1_image_answer_to_original_question'])
(No trust_remote_code=True is required.)Directory StructureMMB-Dataset/
├── README.md                           # This dataset card
├── .gitattributes                      # Git LFS configuration
├── data/                               # Dataset files (Parquet format)
│   └── train.parquet                   # Main benchmark matrix
├── Dataset/                            # Raw generation artifacts
│   ├── images/                         # Uncompressed PNG renders (720x720)
│   ├── scenes/                         # JSON 3D scene graphs and metadata
│   ├── image_mapping_with_questions.csv # Source mapping for the 3x3 grid
│   └── run_metadata.json               # Procedural generation engine parameters
Application & ProtocolFollowing the rigorous evaluation protocol established in the MMIB paper, interpretability metrics (such as Circuit Performance Ratio, Circuit-Model Distance, and Interchange Intervention Accuracy) should only be computed on samples where the target VLM correctly answers the base question ($a_b = y$). This behavioral filter ensures that the model possesses the causal circuit prior to mechanistic evaluation.LicenseMIT