--- dataset_info: features: - name: id dtype: string - name: duration dtype: float64 - name: audio dtype: audio - name: text dtype: string language: - de task_categories: - automatic-speech-recognition source_datasets: - i4ds/spc_r license: cc-by-4.0 --- # i4ds/spc_r_segmented Diarized and segmented speech dataset derived from [i4ds/spc_r](https://huggingface.co/datasets/i4ds/spc_r). ## Description Each row is a merged speech segment belonging to a single speaker. The source audio and SRT subtitles from `i4ds/spc_r` were processed with the following pipeline: 1. **Diarization** -- [pyannote/speaker-diarization-3.1](https://huggingface.co/pyannote/speaker-diarization-3.1) assigned speaker labels to each SRT segment based on temporal overlap. 2. **Merging** -- Consecutive SRT segments from the same speaker were merged when the silence gap between them was below a threshold (default 1.0s) and the resulting duration stayed within bounds (default 10--20s). 3. **Slicing** -- The merged time ranges were used to slice the original audio waveform. Each segment is encoded as FLAC. ## Columns | Column | Type | Description | |------------|---------|--------------------------------------------------| | `id` | string | Unique identifier (`row{NNNNN}_seg{NNN}`) | | `duration` | float64 | Segment duration in seconds | | `audio` | audio | FLAC audio for the segment | | `text` | string | Merged transcript text from the SRT segments | ## Usage ```python from datasets import load_dataset ds = load_dataset("i4ds/spc_r_segmented") print(ds["train"][0]) ```