ArunKr's picture
Upload dataset
402d2ea verified
---
dataset_info:
features:
- name: text
dtype: string
- name: speaker
dtype: string
- name: nano_layer_1
list: int64
- name: nano_layer_2
list: int64
- name: nano_layer_3
list: int64
- name: nano_layer_4
list: int64
- name: encoded_len
dtype: int64
splits:
- name: train
num_bytes: 58048206
num_examples: 9949
download_size: 8076967
dataset_size: 58048206
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
license: mit
task_categories:
- text-to-speech
language:
- hi
---
# 🗣️ tts-quantized-dataset
This dataset contains **quantized Hindi Text-to-Speech (TTS)** samples generated using NVIDIA’s [`nemo-nano-codec-22khz-0.6kbps-12.5fps`](https://huggingface.co/nvidia/nemo-nano-codec-22khz-0.6kbps-12.5fps) neural audio codec.
It is designed for **training lightweight speech synthesis models**, such as token-based TTS models, audio language models, or text-to-codec models.
## 📚 Dataset Summary
| Field | Description |
|--------|--------------|
| **text** | The transcription (Hindi text) corresponding to each audio sample. |
| **speaker** | Speaker identity (derived from the file name). |
| **nano_layer_1–4** | Discrete audio tokens from four quantization layers of the NeMo Nano Codec model. |
| **encoded_len** | Length of each encoded token sequence. |
## 🧠 Source Dataset
- **Original dataset:** [`SayantanJoker/original_data_hindi_tts`](https://huggingface.co/datasets/SayantanJoker/original_data_hindi_tts)
- **Conversion tool:** [NVIDIA NeMo AudioCodecModel](https://docs.nvidia.com/nemo-framework/user-guide/docs/en/stable/tts/overview.html#audio-codec-model)
- **Sampling rate:** 22.05 kHz
- **Compression:** 0.6 kbps (4 quantization layers, 12.5 fps)
## ⚙️ Data Preparation Pipeline
The dataset was preprocessed using the following steps:
1. Load original Hindi TTS dataset from Hugging Face.
2. Resample all audio files to **22,050 Hz**.
3. Encode audio waveforms using `AudioCodecModel` from NVIDIA NeMo.
4. Store the resulting discrete codec tokens into 4 quantizer layers.
5. Save as Hugging Face `DatasetDict` and push to Hub.
The exact preprocessing script used:
```python
https://huggingface.co/ArunKr/tts-quantized-dataset/blob/main/data_prep.py
````
## 🧩 Example Entry
```json
{
"text": "यह एक परीक्षण वाक्य है।",
"speaker": "sample_001",
"nano_layer_1": [3, 18, 94, 105, 78, ...],
"nano_layer_2": [11, 45, 67, 53, 21, ...],
"nano_layer_3": [32, 98, 76, 12, 43, ...],
"nano_layer_4": [25, 14, 89, 64, 72, ...],
"encoded_len": 120
}
```
## 🧰 Usage Example
```python
from datasets import load_dataset
ds = load_dataset("ArunKr/tts-quantized-dataset", split="train")
print(ds[0]["text"])
# "यह एक परीक्षण वाक्य है।"
codes = [ds[0]["nano_layer_1"], ds[0]["nano_layer_2"], ds[0]["nano_layer_3"], ds[0]["nano_layer_4"]]
```
You can easily convert back to waveform using the same NeMo codec:
```python
from nemo.collections.tts.models import AudioCodecModel
import torch
codec = AudioCodecModel.from_pretrained("nvidia/nemo-nano-codec-22khz-0.6kbps-12.5fps").eval()
tokens = torch.tensor([codes])
audio = codec.decode(tokens)
```
## 📦 Dataset Structure
```
tts_quantized_dataset/
├── data_prep.py
├── README.md
├── data/
```
## 🧾 License
* **Source Dataset License:** As provided by the original dataset ([`SayantanJoker/original_data_hindi_tts`](https://huggingface.co/datasets/SayantanJoker/original_data_hindi_tts))
* **Generated Dataset License:** CC BY 4.0
(Attribution required for use and redistribution)
## 🙏 Acknowledgments
* [NVIDIA NeMo](https://github.com/NVIDIA/NeMo) for the neural codec model.
* [Hugging Face Datasets](https://huggingface.co/docs/datasets) for dataset hosting and preprocessing utilities.
* [SayantanJoker/original_data_hindi_tts](https://huggingface.co/datasets/SayantanJoker/original_data_hindi_tts) for providing the original Hindi speech dataset.
## 💬 Citation
If you use this dataset, please cite:
```bibtex
@dataset{arunkr_tts_quantized_dataset_2025,
author = {Arun Kumar Tiwary},
title = {Hindi TTS Quantized Dataset using NeMo Nano Codec},
year = {2025},
url = {https://huggingface.co/datasets/ArunKr/tts-quantized-dataset}
}
```