Datasets:
Upload folder using huggingface_hub
Browse files
README.md
CHANGED
|
@@ -20,10 +20,28 @@ size_categories:
|
|
| 20 |
|
| 21 |
This dataset contains audio segments with transcriptions from Peaky Blinders for Text-to-Speech training.
|
| 22 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 23 |
## Dataset Structure
|
| 24 |
-
- `
|
| 25 |
-
- `metadata
|
| 26 |
-
- `
|
| 27 |
|
| 28 |
## Usage
|
| 29 |
```python
|
|
@@ -31,26 +49,27 @@ import json
|
|
| 31 |
import os
|
| 32 |
|
| 33 |
# Load metadata
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
if file.startswith("tts_metadata_") and file.endswith(".json"):
|
| 37 |
-
with open(os.path.join(metadata_dir, file), 'r') as f:
|
| 38 |
-
data = json.load(f)
|
| 39 |
-
break
|
| 40 |
|
| 41 |
# Access segments
|
| 42 |
for segment in data:
|
| 43 |
print(f"Text: {segment['text']}")
|
| 44 |
print(f"Speaker: {segment['speaker_id']}")
|
| 45 |
print(f"Audio: {segment['audio_file']}")
|
|
|
|
|
|
|
| 46 |
```
|
| 47 |
|
| 48 |
## Dataset Details
|
| 49 |
- **Language**: English
|
| 50 |
-
- **Format**: WAV audio files (
|
| 51 |
- **Speakers**: Multiple speakers (A, B, C, D, E)
|
| 52 |
-
- **Duration**: Variable segment lengths
|
| 53 |
-
- **Quality**: High-quality audio with quality scores
|
|
|
|
|
|
|
|
|
|
| 54 |
|
| 55 |
## Metadata Format
|
| 56 |
Each segment contains:
|
|
@@ -60,6 +79,9 @@ Each segment contains:
|
|
| 60 |
- `duration`: Duration in seconds
|
| 61 |
- `quality_score`: Audio quality score
|
| 62 |
- `confidence_score`: Transcription confidence
|
|
|
|
|
|
|
|
|
|
| 63 |
|
| 64 |
## License
|
| 65 |
MIT License
|
|
@@ -68,7 +90,10 @@ MIT License
|
|
| 68 |
If you use this dataset, please cite:
|
| 69 |
@dataset{peaky_blinders_tts,
|
| 70 |
title={TTS Dataset - Peaky Blinders},
|
| 71 |
-
author={
|
| 72 |
year={2024},
|
| 73 |
url={https://huggingface.co/datasets/ahk-d/peaky-blinders-learning-purpose-only}
|
| 74 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
|
| 21 |
This dataset contains audio segments with transcriptions from Peaky Blinders for Text-to-Speech training.
|
| 22 |
|
| 23 |
+
## Dataset Generation
|
| 24 |
+
|
| 25 |
+
This dataset was generated using the [TTS-Dataset-Maker](https://github.com/ahk-d/TTS-Dataset-Maker) pipeline, which provides:
|
| 26 |
+
|
| 27 |
+
- **Silero VAD-based silence removal** - Removes long silences while preserving natural speech gaps
|
| 28 |
+
- **DeepFilterNet denoising** - CPU-optimized audio denoising with gentle attenuation (15dB)
|
| 29 |
+
- **AssemblyAI transcription** - High-quality speech-to-text with speaker diarization
|
| 30 |
+
- **Existing transcript support** - Uses pre-existing transcripts to skip transcription
|
| 31 |
+
- **Precise segmentation** - Exact audio segment extraction matching transcript timings
|
| 32 |
+
- **Unique file naming** - Source-aware segment names to prevent collisions
|
| 33 |
+
|
| 34 |
+
### Processing Pipeline
|
| 35 |
+
1. Audio preprocessing with DeepFilterNet denoising (atten_lim_db=15.0 for gentle noise reduction)
|
| 36 |
+
2. Voice Activity Detection using Silero VAD to remove long silences
|
| 37 |
+
3. Speaker diarization and transcription (or loading existing transcripts)
|
| 38 |
+
4. Precise audio segmentation based on transcript timestamps
|
| 39 |
+
5. Quality validation and filtering of segments
|
| 40 |
+
|
| 41 |
## Dataset Structure
|
| 42 |
+
- `audio/`: Audio files (WAV format, 44.1kHz)
|
| 43 |
+
- `metadata.json`: Complete segment metadata for Label Studio
|
| 44 |
+
- `transcripts/`: Per-file transcript JSON files
|
| 45 |
|
| 46 |
## Usage
|
| 47 |
```python
|
|
|
|
| 49 |
import os
|
| 50 |
|
| 51 |
# Load metadata
|
| 52 |
+
with open("metadata.json", 'r') as f:
|
| 53 |
+
data = json.load(f)
|
|
|
|
|
|
|
|
|
|
|
|
|
| 54 |
|
| 55 |
# Access segments
|
| 56 |
for segment in data:
|
| 57 |
print(f"Text: {segment['text']}")
|
| 58 |
print(f"Speaker: {segment['speaker_id']}")
|
| 59 |
print(f"Audio: {segment['audio_file']}")
|
| 60 |
+
print(f"Duration: {segment['duration']}s")
|
| 61 |
+
print(f"Quality Score: {segment['quality_score']}")
|
| 62 |
```
|
| 63 |
|
| 64 |
## Dataset Details
|
| 65 |
- **Language**: English
|
| 66 |
+
- **Format**: WAV audio files (44.1kHz, 16-bit PCM)
|
| 67 |
- **Speakers**: Multiple speakers (A, B, C, D, E)
|
| 68 |
+
- **Duration**: Variable segment lengths (filtered for >1s duration)
|
| 69 |
+
- **Quality**: High-quality audio with quality scores and confidence metrics
|
| 70 |
+
- **Total Segments**: 356 segments
|
| 71 |
+
- **Total Duration**: ~89 minutes
|
| 72 |
+
- **Processing**: Denoised with DeepFilterNet, silence-removed with Silero VAD
|
| 73 |
|
| 74 |
## Metadata Format
|
| 75 |
Each segment contains:
|
|
|
|
| 79 |
- `duration`: Duration in seconds
|
| 80 |
- `quality_score`: Audio quality score
|
| 81 |
- `confidence_score`: Transcription confidence
|
| 82 |
+
- `start_time`: Start time in milliseconds
|
| 83 |
+
- `end_time`: End time in milliseconds
|
| 84 |
+
- `source_file`: Original source file path
|
| 85 |
|
| 86 |
## License
|
| 87 |
MIT License
|
|
|
|
| 90 |
If you use this dataset, please cite:
|
| 91 |
@dataset{peaky_blinders_tts,
|
| 92 |
title={TTS Dataset - Peaky Blinders},
|
| 93 |
+
author={ahk-d},
|
| 94 |
year={2024},
|
| 95 |
url={https://huggingface.co/datasets/ahk-d/peaky-blinders-learning-purpose-only}
|
| 96 |
+
}
|
| 97 |
+
|
| 98 |
+
## Generation Tool
|
| 99 |
+
Generated using [TTS-Dataset-Maker](https://github.com/ahk-d/TTS-Dataset-Maker) - A comprehensive pipeline for creating high-quality TTS datasets from audio files with automatic transcription, denoising, and segmentation.
|