shangeth's picture
Add Links section (extraction code, Wren project, TTS models)
4520220 verified
metadata
license: cc-by-4.0
language:
  - en
task_categories:
  - text-to-speech
  - automatic-speech-recognition
tags:
  - mimi
  - neural-codec
  - speech-synthesis
  - speech-recognition
  - librispeech
  - audio-tokens
pretty_name: LibriSpeech Mimi Codes
size_categories:
  - 100K<n<1M

LibriSpeech — Mimi Codes

Pre-extracted Kyutai Mimi neural-codec tokens for the LibriSpeech corpus — multi-speaker English audiobook readings from the LibriVox project.

This dataset contains codes only, not audio. For waveforms, use any of the LibriSpeech mirrors (e.g. openslr/librispeech_asr); these codes let you skip the ~hours of GPU extraction needed to train Mimi-based speech models.

Schema

One row per utterance:

Column Type Notes
id string {speaker_id}-{chapter_id}-{utterance_id:04d}, e.g. 103-1240-0000
text string lowercased transcript
speaker_id int32 LibriSpeech speaker ID
codes int16[k=8][n_frames] Mimi codebook indices @ 12.5 fps
n_frames int32 = codes.shape[1]
k_codebooks int32 = 8

Extraction details

  • Codec: kyutai/mimi @ 24 kHz, 12.5 fps
  • Codebooks: all 8 extracted. Slice codes[:k] for fewer (Mimi's codebooks are ordered by importance; the first few capture most of the signal).
  • Codebook size: 2048 per codebook → values stored as int16
  • Transcripts: sourced from LibriSpeech's .trans.txt files, lowercased (the raw release is ALL-UPPER)

Splits

Each standard LibriSpeech split is a separate HF split (hyphens replaced with underscores):

HF Split Upstream Approx. rows Notes
train_clean_100 train-clean-100 ~28.5k clean read speech, ~100 h
train_clean_360 train-clean-360 ~104.0k clean read speech, ~360 h
train_other_500 train-other-500 ~148.7k noisier/accented, ~500 h
dev_clean dev-clean ~2.7k dev set, clean
dev_other dev-other ~2.9k dev set, noisier
test_clean test-clean ~2.6k test set, clean
test_other test-other ~2.9k test set, noisier

Splits are added incrementally — consult the "Files" tab or load_dataset(...).splits for the exact subset currently available.

Usage

from datasets import load_dataset
import torch

ds = load_dataset("shangeth/librispeech-mimi-codes", split="train_clean_100")

ex    = ds[0]
codes = torch.tensor(ex["codes"], dtype=torch.long)   # [8, n_frames]
print(f"{ex['id']} (speaker {ex['speaker_id']}) → {ex['text'][:60]}")
print("codes:", codes.shape, "duration:", codes.shape[1] / 12.5, "s")

# Use only the first 3 codebooks:
codes_3 = codes[:3]

Streaming (no full download):

ds = load_dataset("shangeth/librispeech-mimi-codes", split="train_clean_360", streaming=True)
for ex in ds.take(10):
    print(ex["id"], len(ex["codes"]), "codebooks")

Decode to audio with the Mimi decoder:

from transformers import MimiModel
mimi = MimiModel.from_pretrained("kyutai/mimi").cuda().eval()
with torch.no_grad():
    wav = mimi.decode(codes.unsqueeze(0).cuda()).audio_values[0].cpu()
# wav is [1, T] @ 24 kHz

License & Attribution

LibriSpeech is released under CC-BY-4.0. The derived Mimi codes inherit this license — attribution is required. Please cite both the original corpus and this dataset when redistributing.

Links

Citations

@misc{wren2026,
  title  = {Wren: A Family of Small Open-Weight Models for Unified Speech-Text Modelling},
  author = {Shangeth Rajaa},
  year   = {2026},
  url    = {https://github.com/shangeth/wren}
}

@inproceedings{panayotov2015librispeech,
  title     = {Librispeech: an ASR corpus based on public domain audio books},
  author    = {Panayotov, Vassil and Chen, Guoguo and Povey, Daniel and Khudanpur, Sanjeev},
  booktitle = {ICASSP},
  year      = {2015}
}

Related

Used to train the Wren series of speech-text multimodal models.