Dataset Viewer
Auto-converted to Parquet Duplicate
subject_id
stringlengths
13
24
predicted_prob
float64
0
1
NeuroVoz:HC:105
0.118294
NeuroVoz:HC:112
0.676388
NeuroVoz:HC:116
0.828416
NeuroVoz:HC:118
0.831112
NeuroVoz:HC:120
0.814477
NeuroVoz:HC:121
0.133899
NeuroVoz:HC:122
0.731584
NeuroVoz:HC:128
0.095173
NeuroVoz:HC:129
0.749635
NeuroVoz:HC:130
0.129021
NeuroVoz:HC:132
0.564955
NeuroVoz:HC:133
0.615277
NeuroVoz:HC:134
0.135217
NeuroVoz:HC:135
0.117521
NeuroVoz:HC:136
0.328517
NeuroVoz:HC:137
0.217027
NeuroVoz:HC:138
0.264049
NeuroVoz:HC:139
0.39201
NeuroVoz:HC:140
0.702393
NeuroVoz:HC:141
0.368273
NeuroVoz:HC:143
0.872137
NeuroVoz:HC:144
0.814349
NeuroVoz:HC:145
0.075465
NeuroVoz:HC:34
0.90393
NeuroVoz:HC:36
0.971847
NeuroVoz:HC:45
0.253888
NeuroVoz:HC:49
0.834753
NeuroVoz:HC:51
0.316468
NeuroVoz:HC:52
0.529112
NeuroVoz:HC:53
0.828976
NeuroVoz:HC:54
0.549193
NeuroVoz:HC:55
0.382404
NeuroVoz:HC:56
0.884134
NeuroVoz:HC:60
0.114878
NeuroVoz:HC:61
0.632887
NeuroVoz:HC:62
0.879942
NeuroVoz:HC:63
0.76601
NeuroVoz:HC:64
0.345057
NeuroVoz:HC:65
0.532474
NeuroVoz:HC:72
0.549336
NeuroVoz:HC:73
0.227787
NeuroVoz:HC:74
0.259934
NeuroVoz:HC:75
0.768321
NeuroVoz:HC:80
0.113304
NeuroVoz:HC:81
0.288782
NeuroVoz:HC:82
0.638511
NeuroVoz:HC:83
0.242379
NeuroVoz:HC:85
0.914955
NeuroVoz:HC:86
0.618744
NeuroVoz:HC:87
0.255171
NeuroVoz:PD:10
0.58391
NeuroVoz:PD:108
0.113627
NeuroVoz:PD:109
0.587838
NeuroVoz:PD:11
0.433313
NeuroVoz:PD:111
0.478341
NeuroVoz:PD:113
0.603856
NeuroVoz:PD:115
0.758922
NeuroVoz:PD:117
0.737209
NeuroVoz:PD:12
0.611853
NeuroVoz:PD:13
0.956736
NeuroVoz:PD:14
0.659844
NeuroVoz:PD:15
0.633432
NeuroVoz:PD:16
0.939236
NeuroVoz:PD:18
0.89052
NeuroVoz:PD:19
0.256821
NeuroVoz:PD:20
0.351269
NeuroVoz:PD:21
0.817232
NeuroVoz:PD:22
0.249312
NeuroVoz:PD:23
0.857449
NeuroVoz:PD:24
0.457029
NeuroVoz:PD:25
0.869647
NeuroVoz:PD:26
0.803683
NeuroVoz:PD:27
0.768059
NeuroVoz:PD:28
0.889104
NeuroVoz:PD:29
0.946767
NeuroVoz:PD:30
0.679248
NeuroVoz:PD:31
0.887661
NeuroVoz:PD:32
0.912119
NeuroVoz:PD:33
0.17206
NeuroVoz:PD:37
0.893481
NeuroVoz:PD:38
0.425604
NeuroVoz:PD:39
0.86011
NeuroVoz:PD:4
0.880078
NeuroVoz:PD:41
0.610759
NeuroVoz:PD:42
0.88436
NeuroVoz:PD:43
0.970393
NeuroVoz:PD:44
0.184903
NeuroVoz:PD:46
0.864482
NeuroVoz:PD:58
0.215353
NeuroVoz:PD:6
0.687329
NeuroVoz:PD:66
0.153725
NeuroVoz:PD:7
0.454815
NeuroVoz:PD:70
0.918391
NeuroVoz:PD:77
0.187076
NeuroVoz:PD:78
0.654217
NeuroVoz:PD:79
0.904788
NeuroVoz:PD:8
0.133714
NeuroVoz:PD:9
0.72393
NeuroVoz:HC:105
0.894785
NeuroVoz:HC:112
0.73321
End of preview. Expand in Data Studio

VoxClinBench

Anonymous submission for NeurIPS 2026 Evaluations & Datasets Track.

First cross-lingual, cross-disease clinical voice biomarker benchmark. Six corpora (Bridge2AI-Voice v3.0, NeuroVoz, SVD, DAIC-WOZ, E-DAIC, MODMA), four languages (en, es, de, zh), 22 ranked subject-disjoint classification tasks + 1 PHQ-8 regression scoping entry.

What this repo contains

  • splits/ — 6 partial subject-ID manifests (test subjects; train/val regenerable via voxbench.data.make_splits after upstream fetch).
  • predictions/66 per-seed probability CSVs (11 external-corpus seed-0 baselines + 55 B2AI Tier-2 5-seed × 11-disease canonical CSVs regenerated from our saved checkpoints). Each CSV: exactly two columns subject_id,predicted_prob (labels filtered out per DUA).
  • croissant.json — NeurIPS 2026 E&D Croissant metadata.
  • README.md, license.

Raw audio is not redistributed — use the voxbench fetch CLI from the companion GitHub mirror to get per-corpus URLs + credential check.

Companion GitHub repo (harness, training, baselines)

https://github.com/voice-bench-submission/voxclinbench

Two workflows supported:

# A. Submit your own model (pure numpy+scipy+sklearn, no torch)
pip install voxbench
voxbench eval --task <id> --predictions your.csv --labels upstream.csv

# B. Reproduce / retrain VoxClinBench-Base (heavy deps)
pip install "voxbench[train]"
python -m voxbench.train --help

Fetch corpora

Corpus License Access
bridge2ai PhysioNet credentialed hard login wall
daicwoz USC/ICT EULA HTTP public; EULA on download
edaic USC/ICT EULA (AVEC'19) HTTP public; EULA on download
svd CC BY 4.0 (Zenodo mirror) publicly downloadable
neurovoz CC BY-NC-ND 4.0 Zenodo access request
modma CC BY-NC 4.0 Lanzhou University form

License

MIT (code) / CC BY 4.0 (manifests, docs). Each upstream corpus retains its native license.

Downloads last month
20