TabMI-Bench / README.md
EvalData's picture
Upload README.md
cf89c11 verified
---
license: mit
language:
- en
pretty_name: TabMI-Bench
tags:
- mechanistic-interpretability
- tabular-foundation-models
- tabpfn
- tabicl
- tabdpt
- iltm
- benchmark
- protocol-benchmark
- neurips-2026
size_categories:
- n<1K
configs:
- config_name: rd5_fullscale_aggregated
data_files: rd5_fullscale_aggregated.json
- config_name: tabdpt_probing_3seed
data_files: tabdpt_probing_3seed.json
- config_name: tabdpt_causal_3seed
data_files: tabdpt_causal_3seed.json
- config_name: nam_holdout
data_files: nam_holdout.json
- config_name: lofo_primary_endpoint
data_files: lofo_primary_endpoint.json
- config_name: c1c2_baselines
data_files: c1c2_baselines.json
- config_name: scale_10k_multiseed
data_files: scale_10k_multiseed.json
- config_name: tabpfn25_fullscale_aggregated
data_files: tabpfn25_fullscale_aggregated.json
---
# TabMI-Bench
A protocol benchmark for **mechanistic interpretability (MI) of tabular foundation models (TFMs)**. NeurIPS 2026 Evaluations & Datasets Track submission.
## What's in this dataset
This Hugging Face repository hosts the **frozen aggregated artifacts** that drive every numbered table and figure in the paper. Bundling these allows reviewers to verify the paper's key numerics without re-running 40 GPU-hours of experiments.
| File | Source experiment | Used by |
|---|---|---|
| `rd5_fullscale_aggregated.json` | Phase 5 multi-seed core comparison (5 seeds × 3 models × 4 functions) | Table 1 (3 strategies), Figure 1 |
| `tabdpt_probing_3seed.json` | TabDPT in-family holdout probing (3 seeds) | Table 1 TabDPT row, §4.3 |
| `tabdpt_causal_3seed.json` | TabDPT noising-based causal tracing (3 seeds) | §4.4 TabDPT causal claim |
| `nam_holdout.json` | NAM out-of-family holdout (5 seeds) | §4.3 NAM holdout boundary case |
| `lofo_primary_endpoint.json` | Leave-one-function-out robustness on primary endpoint | LOFO appendix |
| `c1c2_baselines.json` | Shuffled-label and random-target negative controls | §4.6, Tables 21–22 |
| `scale_10k_multiseed.json` | N=10K probing scale validation (3 seeds) | Appendix G |
| `tabpfn25_fullscale_aggregated.json` | TabPFN v2 vs v2.5 comparison (3 seeds) | Table 7, §G.1 |
## Code & full benchmark suite
The hooks, synthetic probe generators, evaluation scripts, statistical analysis, figure generation, and tests are hosted at: **<https://github.com/evaldataset/TabMI-Bench>**
To regenerate paper-facing tables and figures from this dataset *without GPU access*:
```bash
git clone https://github.com/evaldataset/TabMI-Bench
cd TabMI-Bench
pip install -r requirements.txt
make reproduce-paper-frozen
```
## What is TabMI-Bench?
TabMI-Bench provides:
1. **Hook-based activation extraction** for 5 TFMs spanning 3 architectural families (TabPFN v2/v2.5, TabICL v2, TabDPT, iLTM) plus NAM out-of-family holdout
2. **4 controlled synthetic probe families** (bilinear, sinusoidal, polynomial, mixed) with known ground-truth intermediary variables
3. **4-step evaluation protocol** (synthetic profile → causal validation → negative controls → real-world transfer)
4. **Evidence-coded MI applicability matrix** (8 techniques × 4 architectures with seed-count superscripts)
5. **A primary diagnostic finding**: whole-layer clean activation patching is *uninformative* on ICL-style TFMs due to deterministic cascading; corruption-based (noising) tracing is the informative alternative.
Three descriptive reference computation profiles emerge as calibration baselines:
- **Staged** (TabPFN): U-shaped intermediary recoverability with mid-layer concentration
- **Distributed** (TabICL, TabDPT): uniformly high recoverability across layers
- **Preprocessing-dominant** (iLTM): tree+PCA preprocessing performs the heavy lifting
## Croissant 1.0 metadata
The repository includes `croissant.json` with **12 RAI fields** (data limitations, biases, sensitive information, use cases, social impact, synthetic data flag, source datasets, provenance, collection, maintenance plan, etc.). See `croissant.json` in the repository.
## Source datasets
Real-world evaluation uses **public datasets only** (no new collection):
| Dataset | Source | License | Use |
|---|---|---|---|
| California Housing | OpenML 8092 | CC0 | causal tracing + steering |
| Diabetes | scikit-learn | BSD-3-Clause | causal tracing + steering |
| Wine Quality | OpenML 287 | CC0 | steering |
| Bike Sharing | OpenML 44063 | CC0 | steering |
| Abalone, Boston, Energy, Breast Cancer, Iris, Adult, Credit-G | OpenML / scikit-learn | CC0 / BSD-3-Clause | causal tracing |
## Citation
```bibtex
@inproceedings{anonymous2026tabmibench,
title={TabMI-Bench: Evaluating Mechanistic Interpretability Methods Across Tabular Foundation Model Architectures},
author={Anonymous},
booktitle={Advances in Neural Information Processing Systems (NeurIPS) Evaluations \& Datasets Track},
year={2026}
}
```
## License
MIT. See `LICENSE`.