whiser_parquet / README.md
jaeyeonkim99's picture
Update README.md
b78f6d0 verified
---
pretty_name: WHiSER
tags:
- audio
- speech
- emotion
task_categories:
- audio-classification
- other
task_ids:
- audio-emotion-recognition
language:
- en
license: other
size_categories:
- 1K<n<10K
dataset_summary: >
WHiSER is a multi-annotator speech emotion corpus providing categorical labels and
dimensional V-A-D (arousal, valence, dominance; SAM 1–7) ratings for short speech clips.
This repo contains raw WAV files and rich annotations in raw (labels.txt), consensus, and detailed formats (CSV/JSON).
papers:
- https://www.isca-archive.org/interspeech_2024/naini24_interspeech.pdf
extra_gated_prompt: "You agree to not use the dataset to conduct experiments that cause harm to human subjects."
extra_gated_fields:
Company: text
Country: country
Specific date: date_picker
I want to use this dataset for:
type: select
options:
- Research
- Education
- label: Other
value: other
I agree to use this dataset for non-commercial use ONLY: checkbox
---
# Dataset Description
## Motivation
- Enable both categorical emotion recognition and dimensional affect regression on speech clips with multiple crowd-sourced annotations per instance.
- Study annotator variability, label aggregation, and uncertainty.
## Composition
- Audio: WAV files in `wavs/`
- Annotations:
- Raw: `WHiSER/Labels/labels.txt` (per-file summary + multiple worker lines)
- Aggregated: `WHiSER/Labels/labels_consensus.{csv,json}`
- Per-annotator: `WHiSER/Labels/labels_detailed.{csv,json}`
# Supported Tasks
- Categorical speech emotion recognition (primary label)
- Dimensional affect regression (arousal/valence/dominance; 1–7 SAM scale)
- Research on annotation aggregation and inter-annotator disagreement
# Dataset Structure
## Files
- `wavs/`: raw audio clips (e.g., `006-040.1-2_13.wav`)
- `WHiSER/Labels/labels.txt`
- Header (one per file): `<filename>; <agg_code>; A:<float>; V:<float>; D:<float>;`
- Worker lines: `WORKER<ID>; <primary>; <secondary_csv>; A:<float>; V:<float>; D:<float>;`
- `WHiSER/Labels/labels_consensus.{csv,json}`
- Aggregated primary emotion (and secondary/attributes when applicable), gender, speaker, split (if defined)
- `WHiSER/Labels/labels_detailed.{csv,json}`
- Individual annotations per worker with primary, secondary (multi-label), and V-A-D
## Recommended Features
- `id`: string (e.g., `006-040.1-2_13`)
- `file`: string (e.g., `wavs/006-040.1-2_13.wav`)
- `audio`: Audio feature
- `agg_label_code`: string in `{A,S,H,U,F,D,C,N,O,X}`
- `agg_primary`: normalized primary label (e.g., `Happy`)
- `vad_mean`: `{ arousal: float32, valence: float32, dominance: float32 }` (1–7)
- `secondary`: sequence of strings (from the secondary label set)
- Optional metadata: `split`, `speaker`, `gender`
- `annotations` (from detailed): sequence of {
`worker_id`: string,
`primary`: string,
`secondary`: sequence<string>,
`vad`: { `arousal`: float32, `valence`: float32, `dominance`: float32 }
}
# Label Schema
## Primary consensus codes
- A: Angry
- S: Sad
- H: Happy
- U: Surprise
- F: Fear
- D: Disgust
- C: Contempt
- N: Neutral
- O: Other
- X: No agreement (no plurality winner)
## Secondary labels (multi-select)
- Angry, Sad, Happy, Amused, Neutral, Frustrated, Depressed, Surprise, Concerned,
Disgust, Disappointed, Excited, Confused, Annoyed, Fear, Contempt, Other
## Dimensional affect
- Self-Assessment Manikin (SAM) 1–7
- Valence (1 very negative; 7 very positive)
- Arousal (1 very calm; 7 very active)
- Dominance (1 very weak; 7 very strong)
# Example Instance
- `id`: `006-040.1-2_13`
- `file`: `wavs/006-040.1-2_13.wav`
- `agg_label_code`: `H`
- `vad_mean`: `{ arousal: 4.2, valence: 4.8, dominance: 5.0 }`
- `annotations` (subset):
- `{ worker_id: WORKER00014325, primary: Sad, secondary: [Sad, Concerned], vad: {A:2, V:4, D:4} }`
- `{ worker_id: WORKER00014332, primary: Happy, secondary: [Happy, Concerned], vad: {A:4, V:6, D:6} }`
# Usage
## Loading from Parquet
If you've converted the dataset to nested Parquet format using the provided script, you can load it with the Hugging Face `datasets` library:
```python
from datasets import Dataset
import pyarrow.parquet as pq
# Load the Parquet file
table = pq.read_table("parquet/whiser_nested.parquet")
ds = Dataset(table)
print(f"Dataset size: {len(ds)}")
print(f"Features: {ds.features}")
# Access a single example
example = ds[0]
print(f"ID: {example['id']}")
print(f"Consensus label: {example['agg_primary']}")
print(f"VAD mean: A={example['vad_mean']['arousal']:.2f}, V={example['vad_mean']['valence']:.2f}, D={example['vad_mean']['dominance']:.2f}")
print(f"Secondary labels: {example['secondary']}")
print(f"Number of annotators: {len(example['annotations'])}")
# Access audio bytes (embedded in Parquet)
audio_bytes = example['audio_bytes']
sample_rate = example['sample_rate']
print(f"Audio: {len(audio_bytes)} bytes, {sample_rate} Hz")
# Iterate through per-annotator annotations
for ann in example['annotations']:
worker = ann['worker_id']
primary = ann['primary']
vad = ann['vad']
print(f" Worker {worker}: primary={primary}, A={vad['arousal']}, V={vad['valence']}, D={vad['dominance']}")
```
## Loading Audio
To decode audio from the embedded bytes:
```python
import io
import soundfile as sf
example = ds[0]
audio_bytes = example['audio_bytes']
# Decode audio
audio_data, sr = sf.read(io.BytesIO(audio_bytes))
print(f"Audio shape: {audio_data.shape}, sample rate: {sr}")
```
# Reference
```bibtex
@inproceedings{Naini_2024_2,
author={A. {Reddy Naini} and L. Goncalves and M.A. Kohler and D. Robinson and E. Richerson and C. Busso},
title={{WHiSER}: {White House Tapes} Speech Emotion Recognition Corpus},
booktitle={Interspeech 2024},
volume={},
year={2024},
month={September},
address = {Kos Island, Greece},
pages={1595-1599},
doi={10.21437/Interspeech.2024-1227},
}
```