| # Data dictionary |
|
|
| Column-level schemas for every CSV in `data/` and `samples/`. |
|
|
| ## `speakers.csv` (100 rows) |
|
|
| | Column | Type | Description | Valid values | |
| |---|---|---|---| |
| | `id` | text | Speaker identifier. Joins to `stimuli.reference` and to `<id>R` keys in embedding files. | F01-F50, M01-M50 | |
| | `name` | text | Speaker name (public figure). | celebrity name | |
| | `group` | int | Sociophonetic group code. | 1-5 (see below) | |
| | `gender` | int | Speaker gender. | 1=male, 2=female | |
| | `age` | int | Speaker age bracket. | 1=under 45, 2=55 or older | |
|
|
| ### Sociophonetic group mapping |
|
|
| The 5 sociophonetic groups partition the 100 speakers into 20 speakers each: |
|
|
| | Code | Group | |
| |---|---| |
| | 1 | New York City English | |
| | 2 | Southern American English | |
| | 3 | African American English | |
| | 4 | Latino English | |
| | 5 | Asian American English | |
|
|
| The benchmark protocol uses `group` as a stratification variable for cross-validation folds (gender-balanced, speaker-level) and for fairness analyses. |
|
|
| ## `stimuli.csv` (9,800 rows) |
|
|
| | Column | Type | Description | |
| |---|---|---| |
| | `id` | text | Stimulus identifier (matches the comparison-audio basename). Key into embedding files. | |
| | `stimuli_type` | int | Stimulus type, 1-6. See `stimulus_types.md`. | |
| | `reference` | text | Reference speaker ID. Joins to `speakers.id`. | |
| | `comparison` | text | Comparison speaker ID for non-Type-6 pairs (NaN for Type 6). | |
| | `voice_clone` | int | 1 if comparison clip is an AI voice clone, 0 otherwise. | |
| | `correct_answer` | int | Metadata same/different label. 1=same speaker by metadata, 0=different. | |
| | `scale` | int | For Type 6 morphs, interpolation level in [0, 100]. 100 for non-morph pairs. | |
| | `num_response` | int | Number of listener judgments on this pair. | |
| | `same_vote` | int | Listeners who answered "same speaker". | |
| | `diff_vote` | int | Listeners who answered "different speaker". | |
| | `correct_vote` | int | Listeners whose answer matches the metadata label. | |
| | `incorrect_vote` | int | Listeners whose answer disagrees with the metadata label. | |
| | `accuracy` | float | `correct_vote / num_response`. | |
| | `group` | int | Reference speaker's sociophonetic group. | |
| | `gender` | int | Reference speaker's gender. | |
| | `age` | int | Reference speaker's age bracket. | |
|
|
| `P(same)` is computed downstream as `same_vote / num_response`. |
|
|
| ## `participant_responses.csv` (124,876 rows) |
| |
| | Column | Type | Description | |
| |---|---|---| |
| | `user_id` | int | Pseudonymized listener identifier. Tied to no external account. | |
| | `stimuli_id` | text | Stimulus identifier. Joins to `stimuli.id`. | |
| | `stimuli_type` | int | Stimulus type, 1-6. | |
| | `answer` | int | Listener's binary judgment. 1=same speaker, 0=different. | |
| | `correct` | int | 1 if `answer` matches `correct_answer` in `stimuli.csv`, 0 otherwise. | |
| | `know_speaker` | int | Listener-recognition probe. 1 if listener identified the reference speaker, 0 otherwise. May be missing for early-trial responses. | |
| | `age` | float | Listener age band (categorical, encoded as float). | |
| | `gender` | float | Listener gender. | |
| | `first_language` | float | Listener first-language flag. 0=English first, 1=other. | |
| | `num_stimuli_seen` | float | Cumulative stimulus count for this listener at the time of the response. | |
|
|
| ## `stimuli_interpol.csv` (8,100 rows) |
| |
| Per-stimulus metadata for Type 6 morphs. |
| |
| | Column | Type | Description | |
| |---|---|---| |
| | `id` | text | Stimulus identifier. Joins to `stimuli.id`. | |
| | `source` | text | Source speaker A (one endpoint of the morph trajectory). | |
| | `target` | text | Source speaker B (other endpoint). | |
| | `scale` | int | Interpolation level in [0, 100]. | |
| |
| (Other columns may be present and are described in their headers; the four above are the schema-stable subset used by the analysis notebook.) |
| |
| ## Embedding `.npz` files |
| |
| Each file in `data/embeddings/` is a key-value store: |
| |
| - **Keys** (`.files` attribute): audio basenames without `.wav`. 9,900 keys total: 100 references like `M01R`, `F03R` plus 9,800 comparisons like `1_F01`, `4_M12_M15B`, `6_F03_F09_50`. |
| - **Values**: 1-D `np.float32` arrays of shape `(embedding_dim,)`. Dim depends on the model (see `model_table.md`). |
| |
| Per-layer files (`data/embeddings/layers/`) use the same keys; values have shape `(num_layers, embedding_dim)` (mean-pooled across time per layer). |
| |
| ## How to pair reference and comparison |
| |
| Every voice pair in VIPBench is one row of `stimuli.csv`. The pairing rule is: |
| |
| | Asset | Reference clip | Comparison clip | |
| |---|---|---| |
| | CSV column | `reference` (e.g., `M01`) | `id` (e.g., `1_M01`, `4_M12_M15B`) | |
| | Audio file | `data/audio/reference/{reference}R.wav` | `data/audio/comparison/{id}.wav` | |
| | Embedding key | `{reference}R` (e.g., `M01R`) | `{id}` (e.g., `1_M01`) | |
|
|
| Equivalently: the reference clip's basename is the speaker ID with `R` appended; the comparison clip's basename is exactly the stimulus `id`. The 100 reference embeddings (`*R`) and 9,800 comparison embeddings (`id`) together make the 9,900 keys present in every `.npz`. |
|
|
| Reference example: scoring a model against `P(same)`. |
|
|
| ```python |
| import numpy as np, pandas as pd |
| from sklearn.metrics.pairwise import cosine_similarity |
| |
| stim = pd.read_csv('data/stimuli.csv') |
| emb = dict(np.load('data/embeddings/ecapa_tdnn.npz')) |
| |
| def cos_pair(row): |
| ref = emb[f'{row.reference}R'].reshape(1, -1) |
| cmp = emb[row.id].reshape(1, -1) |
| return cosine_similarity(ref, cmp)[0, 0] |
| |
| stim['cos'] = stim.apply(cos_pair, axis=1) |
| stim['p_same'] = stim['same_vote'] / stim['num_response'] |
| print(stim[['cos', 'p_same']].corr()) |
| ``` |
|
|
| The same pattern (with `librosa.load(...)` instead of dictionary lookup) loads the corresponding audio. |
|
|