--- license: mit pretty_name: MMIB Evaluation Dataset task_categories: - visual-question-answering - multiple-choice language: - en tags: - vision - language - multimodal - counterfactual - mechanistic-interpretability - synthetic size_categories: - n<1K dataset_info: features: - name: original_image dtype: image - name: counterfactual1_image dtype: image - name: counterfactual2_image dtype: image - name: counterfactual1_type dtype: string - name: counterfactual2_type dtype: string - name: counterfactual1_description dtype: string - name: counterfactual2_description dtype: string - name: original_question dtype: string - name: counterfactual1_question dtype: string - name: counterfactual2_question dtype: string - name: original_question_difficulty dtype: string - name: counterfactual1_question_difficulty dtype: string - name: counterfactual2_question_difficulty dtype: string - name: original_image_answer_to_original_question dtype: string - name: original_image_answer_to_cf1_question dtype: string - name: original_image_answer_to_cf2_question dtype: string - name: cf1_image_answer_to_original_question dtype: string - name: cf1_image_answer_to_cf1_question dtype: string - name: cf1_image_answer_to_cf2_question dtype: string - name: cf2_image_answer_to_original_question dtype: string - name: cf2_image_answer_to_cf1_question dtype: string - name: cf2_image_answer_to_cf2_question dtype: string configs: - config_name: default data_files: - split: train path: data/train-* --- # Multimodal Mechanistic Interpretability Benchmark (MMIB) Dataset The **MMIB Dataset** is a highly controlled, synthetic vision-language dataset designed to rigorously evaluate mechanistic interpretability (MI) methods in Large Multimodal Models (VLMs). Built upon procedurally generated CLEVR-style assets, this dataset provides exact ground-truth causal pathways to test whether MI techniques (like causal tracing or interchange interventions) localize genuine cognitive circuits or merely identify descriptive correlations. Unlike standard VQA benchmarks, MMIB uses strict **automated rejection sampling** to eliminate geometric ambiguity, ensuring every spatial and causal relationship is mathematically verifiable. ## Dataset Structure & Interventions This dataset is built on a structured intervention triplet for every base scene. Each row provides a complete $3 \times 3$ cross-modal evaluation matrix (3 images $\times$ 3 text queries), allowing researchers to systematically trace cross-modal information flow. ### 1. Semantic Counterfactuals (Causal Reasoning) To evaluate the model's internal causal logic, we generate minimal counterfactual pairs where the intervention mathematically guarantees a change in the ground-truth answer ($y' \neq y$). * **Image-Based CFs:** 10 targeted 3D scene graph edits (e.g., `change_color`, `change_position`, `relational_flip`) that alter the visual logic while keeping the question fixed. * **Text-Based CFs:** Minimal deterministic mutations to the textual query (e.g., swapping "red" for "blue" or "left" for "right") that guarantee an answer flip on the fixed base image. ### 2. Negative Counterfactuals (Diagnostic Stress Tests) To control for basic visual fragility, we generate **Negative Counterfactuals** featuring 8 types of perceptual corruptions (e.g., `add_noise`, `change_lighting`, `apply_fisheye`). These interventions drastically alter the image distribution *without* changing the underlying 3D geometry or ground-truth answer ($y' = y$). They serve as an experimental baseline: if a model fails on these stress tests, its failure on semantic tasks indicates vulnerability to domain shifts rather than flawed causal logic. ## Using the Dataset ### Loading from Python The dataset is hosted in standard Parquet format. You can load it directly into your mechanistic evaluation pipeline using the Hugging Face `datasets` library: ```python from datasets import load_dataset # Load the MMIB dataset ds = load_dataset("scholo/MMB_dataset", split="train") # Inspect the 3x3 evaluation matrix for the first scene print("Base Question:", ds[0]['original_question']) print("Base Image -> Base Question Answer:", ds[0]['original_image_answer_to_original_question']) print("Semantic CF Image -> Base Question Answer:", ds[0]['cf1_image_answer_to_original_question']) (No trust_remote_code=True is required.)Directory StructureMMB-Dataset/ ├── README.md # This dataset card ├── .gitattributes # Git LFS configuration ├── data/ # Dataset files (Parquet format) │ └── train.parquet # Main benchmark matrix ├── Dataset/ # Raw generation artifacts │ ├── images/ # Uncompressed PNG renders (720x720) │ ├── scenes/ # JSON 3D scene graphs and metadata │ ├── image_mapping_with_questions.csv # Source mapping for the 3x3 grid │ └── run_metadata.json # Procedural generation engine parameters Application & ProtocolFollowing the rigorous evaluation protocol established in the MMIB paper, interpretability metrics (such as Circuit Performance Ratio, Circuit-Model Distance, and Interchange Intervention Accuracy) should only be computed on samples where the target VLM correctly answers the base question ($a_b = y$). This behavioral filter ensures that the model possesses the causal circuit prior to mechanistic evaluation.LicenseMIT