Dataset Viewer
Auto-converted to Parquet Duplicate
audio
audioduration (s)
0.03
6.1
subset
stringclasses
14 values
speaker
stringclasses
867 values
label
stringlengths
0
103
HS2
HS2-SPEAKER-26
guitar
HS2
HS2-SPEAKER-65
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-20
guitar
HS2
HS2-SPEAKER-172
guitar
HS2
HS2-SPEAKER-37
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-72
guitar
HS2
HS2-SPEAKER-43
guitar
HS2
HS2-SPEAKER-5
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-20
guitar
HS2
HS2-SPEAKER-65
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-88
guitar
HS2
HS2-SPEAKER-145
guitar
HS2
HS2-SPEAKER-19
guitar
HS2
HS2-SPEAKER-88
guitar
HS2
HS2-SPEAKER-25
guitar
HS2
HS2-SPEAKER-20
guitar
HS2
HS2-SPEAKER-2
guitar
HS2
HS2-SPEAKER-2
guitar
HS2
HS2-SPEAKER-40
guitar
HS2
HS2-SPEAKER-137
guitar
HS2
HS2-SPEAKER-10
guitar
HS2
HS2-SPEAKER-37
guitar
HS2
HS2-SPEAKER-7
guitar
HS2
HS2-SPEAKER-15
guitar
HS2
HS2-SPEAKER-26
guitar
HS2
HS2-SPEAKER-42
guitar
HS2
HS2-SPEAKER-12
guitar
HS2
HS2-SPEAKER-137
guitar
HS2
HS2-SPEAKER-37
guitar
HS2
HS2-SPEAKER-20
guitar
HS2
HS2-SPEAKER-59
guitar
HS2
HS2-SPEAKER-43
guitar
HS2
HS2-SPEAKER-37
guitar
HS2
HS2-SPEAKER-18
guitar
HS2
HS2-SPEAKER-20
guitar
HS2
HS2-SPEAKER-15
guitar
HS2
HS2-SPEAKER-3
guitar
HS2
HS2-SPEAKER-36
guitar
HS2
HS2-SPEAKER-15
guitar
HS2
HS2-SPEAKER-64
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-20
guitar
HS2
HS2-SPEAKER-51
guitar
HS2
HS2-SPEAKER-6
guitar
HS2
HS2-SPEAKER-88
guitar
HS2
HS2-SPEAKER-126
guitar
HS2
HS2-SPEAKER-10
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-2
guitar
HS2
HS2-SPEAKER-59
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-5
guitar
HS2
HS2-SPEAKER-17
guitar
HS2
HS2-SPEAKER-72
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-43
guitar
HS2
HS2-SPEAKER-126
guitar
HS2
HS2-SPEAKER-84
guitar
HS2
HS2-SPEAKER-25
guitar
HS2
HS2-SPEAKER-21
guitar
HS2
HS2-SPEAKER-138
guitar
HS2
HS2-SPEAKER-20
guitar
HS2
HS2-SPEAKER-52
guitar
HS2
HS2-SPEAKER-33
guitar
HS2
HS2-SPEAKER-172
guitar
HS2
HS2-SPEAKER-52
guitar
HS2
HS2-SPEAKER-20
guitar
HS2
HS2-SPEAKER-28
guitar
HS2
HS2-SPEAKER-20
guitar
HS2
HS2-SPEAKER-126
guitar
HS2
HS2-SPEAKER-28
guitar
HS2
HS2-SPEAKER-43
guitar
HS2
HS2-SPEAKER-12
guitar
HS2
HS2-SPEAKER-29
guitar
HS2
HS2-SPEAKER-33
guitar
HS2
HS2-SPEAKER-17
guitar
HS2
HS2-SPEAKER-127
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-22
guitar
HS2
HS2-SPEAKER-7
guitar
HS2
HS2-SPEAKER-1
guitar
HS2
HS2-SPEAKER-7
guitar
HS2
HS2-SPEAKER-65
guitar
HS2
HS2-SPEAKER-166
guitar
HS2
HS2-SPEAKER-43
guitar
HS2
HS2-SPEAKER-65
guitar
HS2
HS2-SPEAKER-56
guitar
HS2
HS2-SPEAKER-176
guitar
HS2
HS2-SPEAKER-6
guitar
HS2
HS2-SPEAKER-71
trumpet
HS2
HS2-SPEAKER-8
trumpet
HS2
HS2-SPEAKER-2
trumpet
HS2
HS2-SPEAKER-49
trumpet
HS2
HS2-SPEAKER-60
trumpet
HS2
HS2-SPEAKER-22
trumpet
HS2
HS2-SPEAKER-29
trumpet
End of preview. Expand in Data Studio

VocSim — Public Benchmark

GitHub Leaderboard License: CC BY 4.0

The public split of VocSim, a training-free benchmark for zero-shot content identity in single-source audio embeddings. VocSim probes the intrinsic geometric quality of frozen audio representations: do acoustically variable instances of the same content land near each other in embedding space, without any task-specific training?

Basha, M., Zai, A. T., Stoll, S., & Hahnloser, R. H. R. VocSim: A Training-free Benchmark for Zero-shot Content Identity in Single-source Audio. ICML 2026. arXiv:2512.10120

What's here

  • 114,641 clips across 15 public subsets, drawn from 19 source corpora.
  • Domains: human speech (phones, words, utterances), animal vocalizations (birdsong, otter calls), environmental sounds.
  • Conditions: clean to noisy, sub-100ms to multi-second, few to thousands of classes per subset.
  • All audio standardized to 16 kHz mono.
  • Single-source only — no overlapping speakers or simultaneous sources — so evaluation isolates content representation from source separation.

Four additional blind out-of-distribution subsets (low-resource speech in Shipibo-Conibo and Chintang) are held out for server-side evaluation via the leaderboard.

Schema

{
  "audio":  {"array": np.ndarray, "sampling_rate": 16000},
  "subset": "HW1",        # source-corpus tag (see paper for the full list)
  "speaker": "spk_042",   # speaker / animal / source ID, or "N/A"
  "label":  "hello",      # ground-truth class for similarity
}

Quick start

from datasets import load_dataset

ds = load_dataset("vocsim/public", split="train")
print(ds[0])

For end-to-end evaluation (feature extraction, distance computation, P@k / GSR), use the reference pipeline at github.com/vocsim/benchmark.

Companion datasets

Dataset Purpose
vocsim/avian-perception-benchmark Alignment of embeddings with zebra-finch perceptual judgments
vocsim/mouse-strain-classification-benchmark C57 vs DBA USV classification
vocsim/mouse-identity-classification-benchmark Individual-mouse identification from USVs

Licensing

Aggregation and metadata are released under CC BY 4.0. Each source corpus retains its original license; see Appendix A.1.1 of the paper for a per-source breakdown.

Citation

@inproceedings{basha2026vocsim,
  title     = {VocSim: A Training-free Benchmark for Zero-shot Content Identity in Single-source Audio},
  author    = {Basha, Maris and Zai, Anja T. and Stoll, Sabine and Hahnloser, Richard H. R.},
  booktitle = {Proceedings of the 43rd International Conference on Machine Learning (ICML)},
  year      = {2026},
  doi       = {10.48550/arXiv.2512.10120}
}
Downloads last month
53

Paper for vocsim/public