Datasets:
id string | text string | speaker_id int32 | style string | other_speaker_id int32 | other_style string | source_file_id string | channel int32 | start_s float32 | end_s float32 | codes list | n_frames int32 | k_codebooks int32 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
ex01-ex02_default_001__ch1_23.88-28.14 | What made you want to become an actor? Have you always been an actor? | 1 | default | 2 | default | ex01-ex02_default_001 | 1 | 23.879999 | 28.139999 | [
[
1049,
127,
1880,
1031,
842,
972,
1766,
1962,
914,
925,
1730,
1730,
170,
908,
1000,
997,
1069,
1620,
730,
620,
2,
298,
298,
1747,
1747,
1747,
1770,
1642,
889,
1696,
257,
1539,
317,
... | 54 | 32 |
ex01-ex02_default_001__ch1_55.12-61.83 | What do you think gave you the most confidence to take that step and get over the shyness, I guess? | 1 | default | 2 | default | ex01-ex02_default_001 | 1 | 55.119999 | 61.830002 | [
[
1049,
958,
1962,
914,
167,
557,
1478,
2001,
444,
232,
903,
107,
786,
1686,
1722,
1386,
1650,
1238,
623,
635,
1416,
917,
150,
150,
1292,
1658,
912,
1323,
397,
1865,
603,
807,
640,
... | 84 | 32 |
ex01-ex02_default_001__ch1_81.76-94.58 | Exactly. It's like taking a fear at a time. That kind of thing. And then, oh, if I can do this, I can also do that. And then building up. I totally get that. Yeah. So you're from Toronto? Toronto. | 1 | default | 2 | default | ex01-ex02_default_001 | 1 | 81.760002 | 94.580002 | [
[
1049,
1115,
1321,
199,
1364,
307,
1816,
42,
599,
1817,
603,
807,
512,
702,
801,
301,
390,
398,
510,
214,
214,
1827,
310,
1187,
1187,
392,
1736,
863,
103,
1476,
156,
1219,
578,
51... | 161 | 32 |
ex01-ex02_default_001__ch1_96.14-99.49 | Is it actually cold up there? | 1 | default | 2 | default | ex01-ex02_default_001 | 1 | 96.139999 | 99.489998 | [
[
1049,
624,
460,
856,
821,
1421,
146,
1856,
2020,
722,
1252,
1896,
190,
1347,
1713,
604,
897,
518,
502,
729,
396,
940,
919,
854,
398,
741,
741,
844,
1759,
19,
353,
353,
1481,
1481... | 42 | 32 |
ex01-ex02_default_001__ch1_130.75-141.90 | I was going to say because there's so many buildings up there that like, you know, the light reflects off the windows and it hits the street and it suns the asphalt, but garbage pile also works. | 1 | default | 2 | default | ex01-ex02_default_001 | 1 | 130.75 | 141.899994 | [
[
1698,
1494,
507,
382,
1012,
649,
930,
41,
909,
620,
2,
233,
1024,
947,
1527,
1900,
1087,
1390,
165,
474,
269,
715,
715,
1348,
584,
471,
9,
1324,
668,
1107,
1378,
1700,
1330,
1231... | 140 | 32 |
ex01-ex02_default_001__ch1_167.15-195.15 | Oh no, the smell must have been atrocious. Oy vey. Are you like... So what I know about Toronto is that there's a lot of wilderness outside of it, right? Like woods and forested areas and lakes and stuff like that. Did you ever find yourself out and about in the woods and exploring things | 1 | default | 2 | default | ex01-ex02_default_001 | 1 | 167.149994 | 195.149994 | [
[
1049,
1583,
2012,
1464,
536,
1958,
1770,
1770,
932,
953,
60,
207,
1706,
1706,
1627,
1277,
13,
766,
1161,
1264,
1761,
1341,
343,
429,
1227,
1615,
1522,
1403,
117,
547,
1636,
883,
646,... | 350 | 32 |
ex01-ex02_default_001__ch1_195.15-198.24 | nature and all that stuff. I love nature, so yeah. | 1 | default | 2 | default | ex01-ex02_default_001 | 1 | 195.149994 | 198.240005 | [
[
613,
1230,
1333,
442,
132,
1005,
1627,
561,
1756,
221,
432,
517,
1494,
288,
968,
303,
303,
1333,
442,
774,
1443,
1312,
421,
126,
803,
973,
1422,
1450,
1569,
969,
969,
598,
1098,
... | 39 | 32 |
ex01-ex02_default_001__ch2_0.00-23.76 | um hi my name is jackie or jacqueline hi um i am from toronto ontario canada i was raised there up until i decided to move here to the us in 2013 and i moved to la um yeah and i've been in hollywood ever since um | 2 | default | 1 | default | ex01-ex02_default_001 | 2 | 0 | 23.76 | [
[
1049,
1272,
953,
953,
580,
1788,
867,
226,
1142,
1205,
396,
758,
420,
1792,
1857,
999,
1628,
15,
658,
1987,
863,
1883,
1281,
610,
610,
1018,
1751,
908,
1899,
468,
1229,
705,
658,
... | 297 | 32 |
ex01-ex02_default_001__ch2_28.56-55.04 | "Um, I wanted to do acting and performing since I was a kid, but I was incredibly scared and not sup(...TRUNCATED) | 2 | default | 1 | default | ex01-ex02_default_001 | 2 | 28.559999 | 55.040001 | [[1049,1272,953,1387,805,805,1788,1770,867,1642,889,1759,1279,849,1140,1972,999,1628,1838,285,1242,1(...TRUNCATED) | 331 | 32 |
ex01-ex02_default_001__ch2_61.93-81.33 | "um being in a community of people with improv and sketch and doing all of that and knowing that um (...TRUNCATED) | 2 | default | 1 | default | ex01-ex02_default_001 | 2 | 61.93 | 81.330002 | [[1049,1272,1844,580,805,867,1142,79,79,79,79,353,353,353,1288,1342,727,1057,1057,1932,1437,117,510,(...TRUNCATED) | 243 | 32 |
Expresso — Mimi Codes (k = 32)
Pre-extracted Kyutai Mimi tokens (all 32 codebooks) for both the read and conversational subsets of Expresso. Source audio + transcripts live in shangeth/expresso; this dataset publishes the discrete-token version for training Mimi-based speech models without re-extracting.
⚠️ License: CC-BY-NC-4.0 — non-commercial use only.
Why Expresso for Wren?
Expresso is the most directly relevant dataset for speech disentanglement research — the same speakers utter similar content across different expressive styles (happy, sad, confused, whispered, animal, etc.). Useful for studying:
- Style vs content vs speaker identity
- How expressive variation is encoded across Mimi's 32 codebooks (semantic + acoustic hierarchy)
- Controllable style transfer at fixed speaker identity
Configs
read— ~11.6k mono utterances with human transcripts.conversational— ~15.9k mono per-utterance turns from stereo dialogues, transcribed with Whisper Large V3 Turbo (3.0% overall WER on a held-out human-transcribed set).
Why 32 codebooks?
Most published speech-codec datasets use only the first 8 of Mimi's 32 codebooks (1 semantic + 7 acoustic), which is enough for the original Moshi recipe. We publish all 32 so you can:
- Train models on more codebooks for higher resynthesis fidelity
- Study which codebooks carry which content (style/timbre/prosody)
- Slice
codes[:k]at load time to use any prefix
k_codebooks is stored per row so the schema works for both k=32 and any subset you slice.
Schemas
read
| Column | Type | Notes |
|---|---|---|
id |
string | e.g. ex01_confused_00001; longform: ex01_default_longform_00001 (full file in train) |
text |
string | human transcription (mixed case + punctuation); empty for chunked rows |
speaker_id |
int32 | 1–4 |
style |
string | default, confused, enunciated, happy, laughing, narration, sad, whisper |
substyle |
string | finer label (e.g. default_emphasis, default_essentials, default_longform) |
corpus |
string | base or longform |
start_s / end_s |
float32 | null for full-file rows |
codes |
int16[32][n_frames] |
Mimi codebook indices @ 12.5 fps |
n_frames |
int32 | |
k_codebooks |
int32 | 32 |
conversational
| Column | Type | Notes |
|---|---|---|
id |
string | e.g. ex01-ex02_default_001__ch1_23.88-28.14 |
text |
string | Whisper Large V3 Turbo transcript |
speaker_id |
int32 | this channel's speaker (1–4) |
style |
string | this channel's style |
other_speaker_id |
int32 | partner's speaker id |
other_style |
string | partner's style |
source_file_id |
string | the parent stereo file |
channel |
int32 | 1 or 2 |
start_s / end_s |
float32 | turn boundaries within source file |
codes |
int16[32][n_frames] |
Mimi codebook indices @ 12.5 fps |
n_frames |
int32 | |
k_codebooks |
int32 | 32 |
Extraction details
- Source audio:
shangeth/expresso(the official Expresso tar, segmented and built into HF format) - Codec:
kyutai/mimi@ 24 kHz, 12.5 fps, codebook size 2048 (fits int16) - Resampling: 48 kHz mono → 24 kHz before encoding
- Conversational text: machine-transcribed (Whisper Large V3 Turbo with anti-hallucination decoding)
Usage
from datasets import load_dataset
import torch
# Pick a config
read = load_dataset("shangeth/expresso-mimi-codes", "read", split="train")
conv = load_dataset("shangeth/expresso-mimi-codes", "conversational", split="train")
ex = read[0]
codes = torch.tensor(ex["codes"], dtype=torch.long) # [32, n_frames]
print(f"speaker={ex['speaker_id']} style={ex['style']} | {ex['text'][:60]}")
print(f"codes shape: {codes.shape} ({codes.shape[1]/12.5:.2f}s @ 12.5 fps)")
# Use only the first 8 codebooks (Moshi-style)
codes8 = codes[:8]
# Decode back to 24 kHz audio
from transformers import MimiModel
mimi = MimiModel.from_pretrained("kyutai/mimi").cuda().eval()
with torch.no_grad():
wav = mimi.decode(codes.unsqueeze(0).cuda()).audio_values[0].cpu() # [1, T] @ 24 kHz
Links
- Audio dataset:
shangeth/expresso - Extraction code: github.com/shangeth/wren-datasets
- Wren research: github.com/shangeth/wren
Citation
@inproceedings{nguyen2023expresso,
title = {Expresso: A Benchmark and Analysis of Discrete Expressive Speech Resynthesis},
author = {Nguyen, Tu Anh and Hsu, Wei-Ning and D'Avirro, Antony and Shi, Bowen and
Gat, Itai and Fazel-Zarani, Maryam and Remez, Tal and Copet, Jade and
Synnaeve, Gabriel and Hassid, Michael and Kreuk, Felix and Adi, Yossi and Dupoux, Emmanuel},
booktitle = {Interspeech},
year = {2023}
}
@misc{wren2026,
title = {Wren: A Family of Small Open-Weight Models for Unified Speech-Text Modelling},
author = {Shangeth Rajaa},
year = {2026},
url = {https://github.com/shangeth/wren}
}
License
CC-BY-NC-4.0 — non-commercial use only. See shangeth/expresso original_metadata/LICENSE.txt.
- Downloads last month
- 32