--- configs: - config_name: default data_files: - split: test.other path: data/test.other-* - split: validation.other path: data/validation.other-* - split: train.other.500 path: data/train.other.500-* - split: train.clean.100 path: data/train.clean.100-* - split: test.clean path: data/test.clean-* - split: train.clean.360 path: data/train.clean.360-* - split: validation.clean path: data/validation.clean-* dataset_info: features: - name: text dtype: string - name: id dtype: string - name: audio_codes sequence: sequence: int64 splits: - name: test.other num_bytes: 62049899 num_examples: 2939 - name: validation.other num_bytes: 59498714 num_examples: 2864 - name: train.other.500 num_bytes: 5761561617 num_examples: 148688 - name: train.clean.100 num_bytes: 1166450829 num_examples: 28539 - name: test.clean num_bytes: 62745230 num_examples: 2620 - name: train.clean.360 num_bytes: 4216515060 num_examples: 104014 - name: validation.clean num_bytes: 62578176 num_examples: 2703 download_size: 1801683161 dataset_size: 11391399525 --- # Dataset Card for "speech_tokenizer_16k" [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)