Dataset Viewer
Auto-converted to Parquet Duplicate
meta
dict
c1
dict
c2
dict
{ "quick_run": false, "seed": 42, "n_train": 100, "n_test": 50, "device": "cuda:0" }
{ "probing_baselines": { "tabpfn": { "layers": [ 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11 ], "real_alpha_r2": [ 0.42617684602737427, 0.6675007343292236, 0.7059307098388672, ...
{ "tabpfn": { "standard_recovery": [ 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1 ], "noising_sensitivity": [ 1, 0.02216839648543145, 0.031675854755258734, 0.011619469088686713, 0.03728389365322783, 0.1747077...

TabMI-Bench

A protocol benchmark for mechanistic interpretability (MI) of tabular foundation models (TFMs). NeurIPS 2026 Evaluations & Datasets Track submission.

What's in this dataset

This Hugging Face repository hosts the frozen aggregated artifacts that drive every numbered table and figure in the paper. Bundling these allows reviewers to verify the paper's key numerics without re-running 40 GPU-hours of experiments.

File Source experiment Used by
rd5_fullscale_aggregated.json Phase 5 multi-seed core comparison (5 seeds × 3 models × 4 functions) Table 1 (3 strategies), Figure 1
tabdpt_probing_3seed.json TabDPT in-family holdout probing (3 seeds) Table 1 TabDPT row, §4.3
tabdpt_causal_3seed.json TabDPT noising-based causal tracing (3 seeds) §4.4 TabDPT causal claim
nam_holdout.json NAM out-of-family holdout (5 seeds) §4.3 NAM holdout boundary case
lofo_primary_endpoint.json Leave-one-function-out robustness on primary endpoint LOFO appendix
c1c2_baselines.json Shuffled-label and random-target negative controls §4.6, Tables 21–22
scale_10k_multiseed.json N=10K probing scale validation (3 seeds) Appendix G
tabpfn25_fullscale_aggregated.json TabPFN v2 vs v2.5 comparison (3 seeds) Table 7, §G.1

Code & full benchmark suite

The hooks, synthetic probe generators, evaluation scripts, statistical analysis, figure generation, and tests are hosted at: https://github.com/evaldataset/TabMI-Bench

To regenerate paper-facing tables and figures from this dataset without GPU access:

git clone https://github.com/evaldataset/TabMI-Bench
cd TabMI-Bench
pip install -r requirements.txt
make reproduce-paper-frozen

What is TabMI-Bench?

TabMI-Bench provides:

  1. Hook-based activation extraction for 5 TFMs spanning 3 architectural families (TabPFN v2/v2.5, TabICL v2, TabDPT, iLTM) plus NAM out-of-family holdout
  2. 4 controlled synthetic probe families (bilinear, sinusoidal, polynomial, mixed) with known ground-truth intermediary variables
  3. 4-step evaluation protocol (synthetic profile → causal validation → negative controls → real-world transfer)
  4. Evidence-coded MI applicability matrix (8 techniques × 4 architectures with seed-count superscripts)
  5. A primary diagnostic finding: whole-layer clean activation patching is uninformative on ICL-style TFMs due to deterministic cascading; corruption-based (noising) tracing is the informative alternative.

Three descriptive reference computation profiles emerge as calibration baselines:

  • Staged (TabPFN): U-shaped intermediary recoverability with mid-layer concentration
  • Distributed (TabICL, TabDPT): uniformly high recoverability across layers
  • Preprocessing-dominant (iLTM): tree+PCA preprocessing performs the heavy lifting

Croissant 1.0 metadata

The repository includes croissant.json with 12 RAI fields (data limitations, biases, sensitive information, use cases, social impact, synthetic data flag, source datasets, provenance, collection, maintenance plan, etc.). See croissant.json in the repository.

Source datasets

Real-world evaluation uses public datasets only (no new collection):

Dataset Source License Use
California Housing OpenML 8092 CC0 causal tracing + steering
Diabetes scikit-learn BSD-3-Clause causal tracing + steering
Wine Quality OpenML 287 CC0 steering
Bike Sharing OpenML 44063 CC0 steering
Abalone, Boston, Energy, Breast Cancer, Iris, Adult, Credit-G OpenML / scikit-learn CC0 / BSD-3-Clause causal tracing

Citation

@inproceedings{anonymous2026tabmibench,
  title={TabMI-Bench: Evaluating Mechanistic Interpretability Methods Across Tabular Foundation Model Architectures},
  author={Anonymous},
  booktitle={Advances in Neural Information Processing Systems (NeurIPS) Evaluations \& Datasets Track},
  year={2026}
}

License

MIT. See LICENSE.

Downloads last month
40