| # Baseline models |
|
|
| The 10 speech representations benchmarked in the accompanying paper. Each is loaded from a publicly available pretrained checkpoint; weights are downloaded by the corresponding extraction script in `code/`. |
|
|
| | Model | Paradigm | Training data | Dim | HF checkpoint | Extraction script | |
| |---|---|---|---|---|---| |
| | x-vector | Supervised classification | VoxCeleb 1+2 | 512 | `speechbrain/spkrec-xvect-voxceleb` | `extract_xvector.py` | |
| | ECAPA-TDNN | AAM-Softmax | VoxCeleb 1+2 | 192 | `speechbrain/spkrec-ecapa-voxceleb` | `extract_ecapa_tdnn.py` | |
| | RawNet3 | AAM-Softmax | VoxCeleb 1+2 | 192 | `espnet/voxcelebs12_rawnet3` | `extract_rawnet3_embeddings.py` | |
| | TitaNet (large) | AAM-Softmax | VoxCeleb 1+2, Fisher, SWB, LibriSpeech, NIST SRE | 192 | `nvidia/speakerverification_en_titanet_large` | `extract_titanet.py` | |
| | resemblyzer | GE2E loss, 3-layer LSTM | LibriSpeech + VoxCeleb 1+2 | 256 | bundled with `resemblyzer` package | `extract_resemblyzer.py` | |
| | wav2vec 2.0 | Contrastive masked prediction | LibriSpeech 960 h | 768 | `facebook/wav2vec2-base` | `extract_wav2vec2.py` | |
| | HuBERT | Masked prediction | LibriSpeech 960 h | 768 | `facebook/hubert-base-ls960` | `extract_hubert.py` | |
| | WavLM | Masked prediction + denoising | 94K h mixed | 768 | `microsoft/wavlm-base-plus` | `extract_wavlm.py` | |
| | XLS-R | Contrastive multilingual | 436K h, 128 languages | 1024 | `facebook/wav2vec2-xls-r-300m` | `extract_xlsr.py` | |
| | Whisper (encoder, base) | Multitask weakly supervised ASR | 680K h web audio | 512 | `openai/whisper-base` | `extract_whisper.py` | |
|
|
| ## Output format |
|
|
| Each script saves `<model>.npz` to `<VIPBENCH_ROOT>/data/embeddings/`. The file is a key-value store with 9,900 keys (audio basenames without `.wav`) mapping to 1-D `np.float32` arrays of shape `(embedding_dim,)`. |
|
|
| For SSL models (wav2vec 2.0, HuBERT, WavLM, XLS-R, Whisper), `extract_ssl_layers.py` additionally produces a per-layer mean-pooled bundle saved to `<VIPBENCH_ROOT>/data/embeddings/layers/<model>.npz`. Values there have shape `(num_layers, embedding_dim)`. |
|
|
| ## Pooling |
|
|
| Self-supervised models (wav2vec 2.0, HuBERT, WavLM, XLS-R) and Whisper produce frame-level outputs; we use **mean pooling across time** for the utterance-level embedding. Speaker-specialized models (x-vector, ECAPA-TDNN, RawNet3, TitaNet, resemblyzer) produce a single utterance-level vector by design and are passed through unchanged. |
|
|
| ## Best-layer protocol |
|
|
| For SSL models the final layer is known to underperform on speaker tasks (SUPERB, 2021). The notebook implements a **SUPERB-style nested speaker-CV best-layer protocol**: for each held-out speaker fold, the layer that maximizes Pearson r against `P(same)` on training speakers is selected and applied to the test speakers. No pair contributes to selecting its own layer. Per-layer Pearson r curves are reported in the appendix figure. |
|
|
| ## Licenses |
|
|
| Each baseline retains its original license. As of release time: |
|
|
| - speechbrain checkpoints: Apache 2.0 |
| - ESPnet checkpoints: Apache 2.0 |
| - NVIDIA NeMo: NVIDIA Source Code License (research permitted) |
| - Hugging Face hosted facebook/microsoft checkpoints: respective lab licenses (typically MIT / Apache 2.0) |
| - OpenAI Whisper: MIT |
| - resemblyzer: MIT |
|
|
| Verify the license at the linked checkpoint page before redistribution. The CC-BY-NC 4.0 license on VIPBench's audio + judgments + derived embeddings does not extend to these third-party model weights. |
|
|