File size: 3,450 Bytes
6351b36 5740783 6351b36 | 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 | # Baseline models
The 10 speech representations benchmarked in the accompanying paper. Each is loaded from a publicly available pretrained checkpoint; weights are downloaded by the corresponding extraction script in `code/`.
| Model | Paradigm | Training data | Dim | HF checkpoint | Extraction script |
|---|---|---|---|---|---|
| x-vector | Supervised classification | VoxCeleb 1+2 | 512 | `speechbrain/spkrec-xvect-voxceleb` | `extract_xvector.py` |
| ECAPA-TDNN | AAM-Softmax | VoxCeleb 1+2 | 192 | `speechbrain/spkrec-ecapa-voxceleb` | `extract_ecapa_tdnn.py` |
| RawNet3 | AAM-Softmax | VoxCeleb 1+2 | 192 | `espnet/voxcelebs12_rawnet3` | `extract_rawnet3_embeddings.py` |
| TitaNet (large) | AAM-Softmax | VoxCeleb 1+2, Fisher, SWB, LibriSpeech, NIST SRE | 192 | `nvidia/speakerverification_en_titanet_large` | `extract_titanet.py` |
| resemblyzer | GE2E loss, 3-layer LSTM | LibriSpeech + VoxCeleb 1+2 | 256 | bundled with `resemblyzer` package | `extract_resemblyzer.py` |
| wav2vec 2.0 | Contrastive masked prediction | LibriSpeech 960 h | 768 | `facebook/wav2vec2-base` | `extract_wav2vec2.py` |
| HuBERT | Masked prediction | LibriSpeech 960 h | 768 | `facebook/hubert-base-ls960` | `extract_hubert.py` |
| WavLM | Masked prediction + denoising | 94K h mixed | 768 | `microsoft/wavlm-base-plus` | `extract_wavlm.py` |
| XLS-R | Contrastive multilingual | 436K h, 128 languages | 1024 | `facebook/wav2vec2-xls-r-300m` | `extract_xlsr.py` |
| Whisper (encoder, base) | Multitask weakly supervised ASR | 680K h web audio | 512 | `openai/whisper-base` | `extract_whisper.py` |
## Output format
Each script saves `<model>.npz` to `<VIPBENCH_ROOT>/data/embeddings/`. The file is a key-value store with 9,900 keys (audio basenames without `.wav`) mapping to 1-D `np.float32` arrays of shape `(embedding_dim,)`.
For SSL models (wav2vec 2.0, HuBERT, WavLM, XLS-R, Whisper), `extract_ssl_layers.py` additionally produces a per-layer mean-pooled bundle saved to `<VIPBENCH_ROOT>/data/embeddings/layers/<model>.npz`. Values there have shape `(num_layers, embedding_dim)`.
## Pooling
Self-supervised models (wav2vec 2.0, HuBERT, WavLM, XLS-R) and Whisper produce frame-level outputs; we use **mean pooling across time** for the utterance-level embedding. Speaker-specialized models (x-vector, ECAPA-TDNN, RawNet3, TitaNet, resemblyzer) produce a single utterance-level vector by design and are passed through unchanged.
## Best-layer protocol
For SSL models the final layer is known to underperform on speaker tasks (SUPERB, 2021). The notebook implements a **SUPERB-style nested speaker-CV best-layer protocol**: for each held-out speaker fold, the layer that maximizes Pearson r against `P(same)` on training speakers is selected and applied to the test speakers. No pair contributes to selecting its own layer. Per-layer Pearson r curves are reported in the appendix figure.
## Licenses
Each baseline retains its original license. As of release time:
- speechbrain checkpoints: Apache 2.0
- ESPnet checkpoints: Apache 2.0
- NVIDIA NeMo: NVIDIA Source Code License (research permitted)
- Hugging Face hosted facebook/microsoft checkpoints: respective lab licenses (typically MIT / Apache 2.0)
- OpenAI Whisper: MIT
- resemblyzer: MIT
Verify the license at the linked checkpoint page before redistribution. The CC-BY-NC 4.0 license on VIPBench's audio + judgments + derived embeddings does not extend to these third-party model weights.
|