| | --- |
| | language: |
| | - en |
| | license: cc-by-4.0 |
| | configs: |
| | - config_name: default |
| | data_files: |
| | - split: emilia |
| | path: emilia/* |
| | - split: hifitts2 |
| | path: hifitts2/* |
| | splits: |
| | - name: emilia |
| | num_examples: 1693423 |
| | - name: hifitts2 |
| | num_examples: 1220574 |
| | pipeline_tag: text-to-speech |
| | tags: |
| | - voxtream |
| | - text-to-speech |
| | task_categories: |
| | - text-to-speech |
| | --- |
| | |
| | # Model Card for VoXtream training dataset |
| |
|
| | This repository contains a training dataset for [VoXtream](https://huggingface.co/herimor/voxtream) TTS model. |
| |
|
| | The dataset contains 9k hours: |
| |
|
| | - 4.5k hours sampled from [Emilia](https://huggingface.co/datasets/amphion/Emilia-Dataset) dataset. We applied additional diarization to remove multi-speaker utterances and discarded utterances with invalid automatic transcripts. We also used [NISQA](https://github.com/gabrielmittag/NISQA) model to remove low-quality utterances. |
| | - 4.5k hours sampled from [HiFiTTS2](https://huggingface.co/datasets/nvidia/hifitts-2) dataset (22 kHz subset). We selected only single-speaker utterances and filtered the dataset by the WER. |
| |
|
| | All utterances are 25-seconds long. For shorter audio clips we concatenated multiple utterances within the same speaker. Sampling rate: 24kHz. |
| |
|
| | ### Description |
| |
|
| | - **mimi_codes_16cb** - Tokens extracted by the [Mimi](https://huggingface.co/kyutai/mimi) audio codec (16 codebooks). |
| | - **phone_emb_indices** - Alignment of phoneme tokens to Mimi audio frames extracted by [MFA](https://montreal-forced-aligner.readthedocs.io). |
| | - **phone_tokens** - Phoneme tokens. |
| | - **sem_label_shifts** - Monotonic phoneme alignment labels. |
| | - **spk_templates** - Speaker templates for the first 3 seconds of audio extracted by [ReDimNet](https://github.com/IDRnD/redimnet) model. |
| |
|
| | ### Sources |
| |
|
| | - **Repository:** [repo](https://github.com/herimor/voxtream) |
| | - **Paper:** [paper](https://arxiv.org/pdf/2509.15969) |
| | - **Demo:** [demo](https://herimor.github.io/voxtream) |
| |
|
| | ## Get started |
| |
|
| | To download the dataset, use the following code: |
| |
|
| | ```bash |
| | from huggingface_hub import snapshot_download |
| | |
| | local_dir = snapshot_download('herimor/voxtream-train-9k', repo_type='dataset') |
| | ``` |
| |
|
| | Clone our [repo](https://github.com/herimor/voxtream) and follow the instructions in the README file. |
| |
|
| | ## Sample Usage |
| |
|
| | The following examples demonstrate how to use the VoXtream model (trained on this dataset) for output streaming and full streaming. |
| |
|
| |
|
| | ### Installation |
| | ```bash |
| | pip install voxtream |
| | ``` |
| |
|
| | ### Output streaming |
| | ```bash |
| | voxtream \ |
| | --prompt-audio assets/audio/male.wav \ |
| | --prompt-text "The liquor was first created as 'Brandy Milk', produced with milk, brandy and vanilla." \ |
| | --text "In general, however, some method is then needed to evaluate each approximation." \ |
| | --output "output_stream.wav" |
| | ``` |
| | * Note: Initial run may take some time to download model weights. |
| |
|
| | ### Full streaming |
| | ```bash |
| | voxtream \ |
| | --prompt-audio assets/audio/female.wav \ |
| | --prompt-text "Betty Cooper helps Archie with cleaning a store room, when Reggie attacks her." \ |
| | --text "Staff do not always do enough to prevent violence." \ |
| | --output "full_stream.wav" \ |
| | --full-stream |
| | ``` |
| |
|
| | ## Citation |
| |
|
| | ``` |
| | @article{torgashov2025voxtream, |
| | author = {Torgashov, Nikita and Henter, Gustav Eje and Skantze, Gabriel}, |
| | title = {Vo{X}tream: Full-Stream Text-to-Speech with Extremely Low Latency}, |
| | journal = {arXiv:2509.15969}, |
| | year = {2025} |
| | } |
| | ``` |