Datasets:
Tasks:
Feature Extraction
Languages:
English
Size:
1K<n<10K
Tags:
representation-similarity
representation-convergence
cross-model-transport
benchmark
alignment
evaluation
License:
File size: 6,648 Bytes
285f7ca | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 | ---
license: apache-2.0
language:
- en
pretty_name: BCCT-Hub
tags:
- representation-similarity
- representation-convergence
- cross-model-transport
- benchmark
- alignment
- evaluation
- neurips-2026
size_categories:
- 1K<n<10K
task_categories:
- feature-extraction
configs:
- config_name: vision_atlas
data_files:
- split: pairs
path: atlas.json
- config_name: llm_atlas
data_files:
- split: pairs
path: llm_atlas.json
- config_name: audio_atlas
data_files:
- split: pairs
path: audio_atlas.json
- config_name: video_atlas
data_files:
- split: pairs
path: video_atlas.json
- config_name: meta_analysis
data_files:
- split: papers
path: meta_analysis.csv
---
# BCCT-Hub
A benchmark and evaluation dataset for measuring **representation convergence
and cross-model transport** across pretrained encoders. Released alongside the
NeurIPS 2026 Evaluations & Datasets Track submission *"BCCT-Hub: A Benchmark
and Toolkit for Measuring Representation Convergence Across Model Families."*
The dataset packages four pre-computed pairwise compatibility atlases, 41
pre-extracted feature tensors, statistical-analysis JSON outputs, an 88-paper
meta-analysis CSV, and a Croissant 1.0 metadata file with the five
Responsible-AI fields required by the NeurIPS 2026 E&D Track.
## Atlas summary
| Atlas | Pairs | Models | Source data |
|---|---|---|---|
| `atlas.json` (vision) | 190 | 20 vision encoders | CIFAR-100 test (5000 images) |
| `llm_atlas.json` | 36 | 9 base LLMs | WikiText-103 (2000 passages of 128 tokens) |
| `audio_atlas.json` (preliminary) | 15 | 6 audio encoders | LibriSpeech test-clean |
| `video_atlas.json` (exploratory) | 15 | 6 video encoders | STL-10 pseudo-clips |
| `meta_analysis.csv` | — | 88 papers tagged | Survey extraction |
Each pairwise record carries six BCCT quantities:
**R** (effective-rank bitrate proxy), **τ** (bidirectional Procrustes-vs-MLP
transport linearity score), **λ** (alignment locality), **TAI** (transport
asymmetry), **Δ** (bottleneck mismatch), and a **regime** label in
{Collapsed, Local-Only, Convergent, Divergent}.
## Headline empirical results
- **Family is the strongest predictor of transport success:** mixed-effects
random-intercept LMM with both models in each pair contributing as random
effects yields β_family = 0.20 (p < 10⁻¹⁶), β_Δ = −0.034 (LRT χ² = 55.1,
p < 10⁻¹³), β_objective = 0.042 (p = 0.023).
- **The bitrate–transport association survives dependence-aware resampling:**
block-bootstrap ρ = −0.70, 95 % CI [−0.82, −0.43].
- **S_local predicts cross-encoder k-NN retrieval:** raw ρ = 0.76 (p =
1.1 × 10⁻⁴; Benjamini–Hochberg-adjusted p = 1.7 × 10⁻³ across 15 retrieval
correlations, 9/15 significant).
## Repository layout
```
data/
├── atlas.json (vision; 190 pairs)
├── llm_atlas.json (language; 36 pairs)
├── audio_atlas.json (preliminary; 15 pairs)
├── video_atlas.json (exploratory; 15 pairs)
├── croissant.json (Croissant 1.0 + RAI fields)
├── meta_analysis.csv (88-paper survey)
├── bitrate_estimator_robustness.json
├── audio_bitrate_sweep.json
├── stl10_vs_cifar100_comparison.json
├── cifar100_{test,train}_labels.pt (paired CIFAR-100 labels)
├── experiments/ (mixed-effects, holdout, retrieval, stitching, ...)
├── features/ (vision: 20 encoders × 5000 CIFAR-100 test)
├── features_train/ (vision: 20 encoders × 5000 CIFAR-100 train, seed=42)
├── features_external/ (5 out-of-atlas encoders for the case study)
├── features_stl10/ (vision robustness check)
└── audio_features/ (6 audio encoders × LibriSpeech test-clean)
```
## Croissant 1.0 metadata
The dataset ships with a validated Croissant file
([`data/croissant.json`](data/croissant.json), 17 KB). It declares SHA-256
hashes for each atlas, a `conformsTo: http://mlcommons.org/croissant/1.0`
header, and the five Responsible-AI fields required by the NeurIPS 2026
Evaluations & Datasets Track (`rai:dataLimitations`, `rai:dataBiases`,
`rai:dataPersonalSensitiveInformation`, `rai:dataUseCases`,
`rai:dataSocialImpact`). The file passes the official validator:
```bash
pip install mlcroissant
mlcroissant validate --jsonld data/croissant.json
# → "Done." (zero errors)
```
## Intended uses
1. Compatibility screening between candidate model pairs prior to
model-stitching or knowledge-transfer experiments.
2. Reproducing and auditing the headline statistical findings of the paper.
3. Extending the atlas with new encoders by re-running the extraction pipeline
in the companion code repository.
4. Regression testing for representation-similarity research that builds on
CKA, mutual k-NN, Procrustes-based transport, or effective-rank proxies.
## Out-of-scope uses (not validated, not recommended)
- Deployment-time decisions about whether two production models are
interchangeable in a downstream application.
- Safety or fairness certification of any individual encoder.
- Inference about model training data, intellectual property, or copyright
provenance from feature-space similarity.
BCCT scores are diagnostic summaries for comparative research, **not deployment
guarantees**. See `data/croissant.json` `rai:dataSocialImpact` for the full
scope statement.
## Source-data licensing
The dataset redistributes only **deterministic numerical activations and
per-pair metric outputs**, never source images, audio, or text. Upstream
benchmarks remain under their original licenses:
- CIFAR-100: MIT-style permissive
- STL-10: permissive (Coates et al., 2011)
- WikiText-103: CC BY-SA 3.0
- LibriSpeech test-clean: CC BY 4.0
Our derivative numerical artifacts are released under **Apache-2.0** (see
`LICENSE`).
## Companion artifacts
- Code repository (toolkit + paper source): <https://github.com/evaldataset/BCCT-Hub>
- Anonymous mirror used during NeurIPS review:
<https://anonymous.4open.science/r/bcct-hub>
## Citation
If you use BCCT-Hub in your research, please cite the accompanying paper:
```bibtex
@misc{bcct_hub_2026,
title = {{BCCT-Hub}: A Benchmark and Toolkit for Measuring Representation
Convergence Across Model Families},
author = {Anonymous Authors},
year = {2026},
note = {NeurIPS 2026 Evaluations \& Datasets Track (under review)}
}
```
|