--- language: - he license: mit task_categories: - automatic-speech-recognition tags: - speech-recognition - lyrics - hebrew - music pretty_name: Caspi STT Benchmark size_categories: - n<1K --- # Caspi STT Benchmark Hebrew speech-to-text (STT) evaluation dataset built from **Mati Caspi** songs: YouTube audio plus reference lyrics, packaged for Hugging Face. ## Dataset description - **Audio**: 16 kHz mono WAV segments (one row per track or segment). - **Text**: Reference transcript (lyrics or song title when lyrics are missing). - **Metadata**: `id`, `youtube_id`, `title`, `song_name`. Intended for STT benchmarking: compare model transcriptions to `text` (e.g. WER/CER). ## How to use ```python from datasets import load_dataset, Audio ds = load_dataset("ozlabs/caspi", split="train") ds = ds.cast_column("audio", Audio(sampling_rate=16_000)) # Example row ex = ds[0] # ex["audio"] → decoded array; ex["text"] → reference transcript ``` ## Source - Audio: extracted from YouTube (playlists) at 16 kHz mono. - Lyrics: provided manually or via Shazam; ## License MIT (or as specified in the repo). Audio and lyrics are used for research/evaluation; respect YouTube and lyric providers’ terms.