--- license: mit language: - en pretty_name: TabMI-Bench tags: - mechanistic-interpretability - tabular-foundation-models - tabpfn - tabicl - tabdpt - iltm - benchmark - protocol-benchmark - neurips-2026 size_categories: - n<1K configs: - config_name: rd5_fullscale_aggregated data_files: rd5_fullscale_aggregated.json - config_name: tabdpt_probing_3seed data_files: tabdpt_probing_3seed.json - config_name: tabdpt_causal_3seed data_files: tabdpt_causal_3seed.json - config_name: nam_holdout data_files: nam_holdout.json - config_name: lofo_primary_endpoint data_files: lofo_primary_endpoint.json - config_name: c1c2_baselines data_files: c1c2_baselines.json - config_name: scale_10k_multiseed data_files: scale_10k_multiseed.json - config_name: tabpfn25_fullscale_aggregated data_files: tabpfn25_fullscale_aggregated.json --- # TabMI-Bench A protocol benchmark for **mechanistic interpretability (MI) of tabular foundation models (TFMs)**. NeurIPS 2026 Evaluations & Datasets Track submission. ## What's in this dataset This Hugging Face repository hosts the **frozen aggregated artifacts** that drive every numbered table and figure in the paper. Bundling these allows reviewers to verify the paper's key numerics without re-running 40 GPU-hours of experiments. | File | Source experiment | Used by | |---|---|---| | `rd5_fullscale_aggregated.json` | Phase 5 multi-seed core comparison (5 seeds × 3 models × 4 functions) | Table 1 (3 strategies), Figure 1 | | `tabdpt_probing_3seed.json` | TabDPT in-family holdout probing (3 seeds) | Table 1 TabDPT row, §4.3 | | `tabdpt_causal_3seed.json` | TabDPT noising-based causal tracing (3 seeds) | §4.4 TabDPT causal claim | | `nam_holdout.json` | NAM out-of-family holdout (5 seeds) | §4.3 NAM holdout boundary case | | `lofo_primary_endpoint.json` | Leave-one-function-out robustness on primary endpoint | LOFO appendix | | `c1c2_baselines.json` | Shuffled-label and random-target negative controls | §4.6, Tables 21–22 | | `scale_10k_multiseed.json` | N=10K probing scale validation (3 seeds) | Appendix G | | `tabpfn25_fullscale_aggregated.json` | TabPFN v2 vs v2.5 comparison (3 seeds) | Table 7, §G.1 | ## Code & full benchmark suite The hooks, synthetic probe generators, evaluation scripts, statistical analysis, figure generation, and tests are hosted at: **** To regenerate paper-facing tables and figures from this dataset *without GPU access*: ```bash git clone https://github.com/evaldataset/TabMI-Bench cd TabMI-Bench pip install -r requirements.txt make reproduce-paper-frozen ``` ## What is TabMI-Bench? TabMI-Bench provides: 1. **Hook-based activation extraction** for 5 TFMs spanning 3 architectural families (TabPFN v2/v2.5, TabICL v2, TabDPT, iLTM) plus NAM out-of-family holdout 2. **4 controlled synthetic probe families** (bilinear, sinusoidal, polynomial, mixed) with known ground-truth intermediary variables 3. **4-step evaluation protocol** (synthetic profile → causal validation → negative controls → real-world transfer) 4. **Evidence-coded MI applicability matrix** (8 techniques × 4 architectures with seed-count superscripts) 5. **A primary diagnostic finding**: whole-layer clean activation patching is *uninformative* on ICL-style TFMs due to deterministic cascading; corruption-based (noising) tracing is the informative alternative. Three descriptive reference computation profiles emerge as calibration baselines: - **Staged** (TabPFN): U-shaped intermediary recoverability with mid-layer concentration - **Distributed** (TabICL, TabDPT): uniformly high recoverability across layers - **Preprocessing-dominant** (iLTM): tree+PCA preprocessing performs the heavy lifting ## Croissant 1.0 metadata The repository includes `croissant.json` with **12 RAI fields** (data limitations, biases, sensitive information, use cases, social impact, synthetic data flag, source datasets, provenance, collection, maintenance plan, etc.). See `croissant.json` in the repository. ## Source datasets Real-world evaluation uses **public datasets only** (no new collection): | Dataset | Source | License | Use | |---|---|---|---| | California Housing | OpenML 8092 | CC0 | causal tracing + steering | | Diabetes | scikit-learn | BSD-3-Clause | causal tracing + steering | | Wine Quality | OpenML 287 | CC0 | steering | | Bike Sharing | OpenML 44063 | CC0 | steering | | Abalone, Boston, Energy, Breast Cancer, Iris, Adult, Credit-G | OpenML / scikit-learn | CC0 / BSD-3-Clause | causal tracing | ## Citation ```bibtex @inproceedings{anonymous2026tabmibench, title={TabMI-Bench: Evaluating Mechanistic Interpretability Methods Across Tabular Foundation Model Architectures}, author={Anonymous}, booktitle={Advances in Neural Information Processing Systems (NeurIPS) Evaluations \& Datasets Track}, year={2026} } ``` ## License MIT. See `LICENSE`.