subject_id stringlengths 13 24 | predicted_prob float64 0 1 |
|---|---|
NeuroVoz:HC:105 | 0.118294 |
NeuroVoz:HC:112 | 0.676388 |
NeuroVoz:HC:116 | 0.828416 |
NeuroVoz:HC:118 | 0.831112 |
NeuroVoz:HC:120 | 0.814477 |
NeuroVoz:HC:121 | 0.133899 |
NeuroVoz:HC:122 | 0.731584 |
NeuroVoz:HC:128 | 0.095173 |
NeuroVoz:HC:129 | 0.749635 |
NeuroVoz:HC:130 | 0.129021 |
NeuroVoz:HC:132 | 0.564955 |
NeuroVoz:HC:133 | 0.615277 |
NeuroVoz:HC:134 | 0.135217 |
NeuroVoz:HC:135 | 0.117521 |
NeuroVoz:HC:136 | 0.328517 |
NeuroVoz:HC:137 | 0.217027 |
NeuroVoz:HC:138 | 0.264049 |
NeuroVoz:HC:139 | 0.39201 |
NeuroVoz:HC:140 | 0.702393 |
NeuroVoz:HC:141 | 0.368273 |
NeuroVoz:HC:143 | 0.872137 |
NeuroVoz:HC:144 | 0.814349 |
NeuroVoz:HC:145 | 0.075465 |
NeuroVoz:HC:34 | 0.90393 |
NeuroVoz:HC:36 | 0.971847 |
NeuroVoz:HC:45 | 0.253888 |
NeuroVoz:HC:49 | 0.834753 |
NeuroVoz:HC:51 | 0.316468 |
NeuroVoz:HC:52 | 0.529112 |
NeuroVoz:HC:53 | 0.828976 |
NeuroVoz:HC:54 | 0.549193 |
NeuroVoz:HC:55 | 0.382404 |
NeuroVoz:HC:56 | 0.884134 |
NeuroVoz:HC:60 | 0.114878 |
NeuroVoz:HC:61 | 0.632887 |
NeuroVoz:HC:62 | 0.879942 |
NeuroVoz:HC:63 | 0.76601 |
NeuroVoz:HC:64 | 0.345057 |
NeuroVoz:HC:65 | 0.532474 |
NeuroVoz:HC:72 | 0.549336 |
NeuroVoz:HC:73 | 0.227787 |
NeuroVoz:HC:74 | 0.259934 |
NeuroVoz:HC:75 | 0.768321 |
NeuroVoz:HC:80 | 0.113304 |
NeuroVoz:HC:81 | 0.288782 |
NeuroVoz:HC:82 | 0.638511 |
NeuroVoz:HC:83 | 0.242379 |
NeuroVoz:HC:85 | 0.914955 |
NeuroVoz:HC:86 | 0.618744 |
NeuroVoz:HC:87 | 0.255171 |
NeuroVoz:PD:10 | 0.58391 |
NeuroVoz:PD:108 | 0.113627 |
NeuroVoz:PD:109 | 0.587838 |
NeuroVoz:PD:11 | 0.433313 |
NeuroVoz:PD:111 | 0.478341 |
NeuroVoz:PD:113 | 0.603856 |
NeuroVoz:PD:115 | 0.758922 |
NeuroVoz:PD:117 | 0.737209 |
NeuroVoz:PD:12 | 0.611853 |
NeuroVoz:PD:13 | 0.956736 |
NeuroVoz:PD:14 | 0.659844 |
NeuroVoz:PD:15 | 0.633432 |
NeuroVoz:PD:16 | 0.939236 |
NeuroVoz:PD:18 | 0.89052 |
NeuroVoz:PD:19 | 0.256821 |
NeuroVoz:PD:20 | 0.351269 |
NeuroVoz:PD:21 | 0.817232 |
NeuroVoz:PD:22 | 0.249312 |
NeuroVoz:PD:23 | 0.857449 |
NeuroVoz:PD:24 | 0.457029 |
NeuroVoz:PD:25 | 0.869647 |
NeuroVoz:PD:26 | 0.803683 |
NeuroVoz:PD:27 | 0.768059 |
NeuroVoz:PD:28 | 0.889104 |
NeuroVoz:PD:29 | 0.946767 |
NeuroVoz:PD:30 | 0.679248 |
NeuroVoz:PD:31 | 0.887661 |
NeuroVoz:PD:32 | 0.912119 |
NeuroVoz:PD:33 | 0.17206 |
NeuroVoz:PD:37 | 0.893481 |
NeuroVoz:PD:38 | 0.425604 |
NeuroVoz:PD:39 | 0.86011 |
NeuroVoz:PD:4 | 0.880078 |
NeuroVoz:PD:41 | 0.610759 |
NeuroVoz:PD:42 | 0.88436 |
NeuroVoz:PD:43 | 0.970393 |
NeuroVoz:PD:44 | 0.184903 |
NeuroVoz:PD:46 | 0.864482 |
NeuroVoz:PD:58 | 0.215353 |
NeuroVoz:PD:6 | 0.687329 |
NeuroVoz:PD:66 | 0.153725 |
NeuroVoz:PD:7 | 0.454815 |
NeuroVoz:PD:70 | 0.918391 |
NeuroVoz:PD:77 | 0.187076 |
NeuroVoz:PD:78 | 0.654217 |
NeuroVoz:PD:79 | 0.904788 |
NeuroVoz:PD:8 | 0.133714 |
NeuroVoz:PD:9 | 0.72393 |
NeuroVoz:HC:105 | 0.894785 |
NeuroVoz:HC:112 | 0.73321 |
End of preview. Expand in Data Studio
VoxClinBench
Anonymous submission for NeurIPS 2026 Evaluations & Datasets Track.
First cross-lingual, cross-disease clinical voice biomarker benchmark. Six corpora (Bridge2AI-Voice v3.0, NeuroVoz, SVD, DAIC-WOZ, E-DAIC, MODMA), four languages (en, es, de, zh), 22 ranked subject-disjoint classification tasks + 1 PHQ-8 regression scoping entry.
What this repo contains
splits/— 6 partial subject-ID manifests (test subjects; train/val regenerable viavoxbench.data.make_splitsafter upstream fetch).predictions/— 66 per-seed probability CSVs (11 external-corpus seed-0 baselines + 55 B2AI Tier-2 5-seed × 11-disease canonical CSVs regenerated from our saved checkpoints). Each CSV: exactly two columnssubject_id,predicted_prob(labels filtered out per DUA).croissant.json— NeurIPS 2026 E&D Croissant metadata.- README.md, license.
Raw audio is not redistributed — use the voxbench fetch CLI from
the companion GitHub mirror to get per-corpus URLs + credential check.
Companion GitHub repo (harness, training, baselines)
https://github.com/voice-bench-submission/voxclinbench
Two workflows supported:
# A. Submit your own model (pure numpy+scipy+sklearn, no torch)
pip install voxbench
voxbench eval --task <id> --predictions your.csv --labels upstream.csv
# B. Reproduce / retrain VoxClinBench-Base (heavy deps)
pip install "voxbench[train]"
python -m voxbench.train --help
Fetch corpora
| Corpus | License | Access |
|---|---|---|
bridge2ai |
PhysioNet credentialed | hard login wall |
daicwoz |
USC/ICT EULA | HTTP public; EULA on download |
edaic |
USC/ICT EULA (AVEC'19) | HTTP public; EULA on download |
svd |
CC BY 4.0 (Zenodo mirror) | publicly downloadable |
neurovoz |
CC BY-NC-ND 4.0 | Zenodo access request |
modma |
CC BY-NC 4.0 | Lanzhou University form |
License
MIT (code) / CC BY 4.0 (manifests, docs). Each upstream corpus retains its native license.
- Downloads last month
- 20