--- license: apache-2.0 task_categories: - automatic-speech-recognition language: - uz --- # Speech-to-Text Evaluation Dataset ## Dataset Overview This dataset is designed for evaluating Uzbek speech-to-text (STT) models on real-world conversational speech data. The audio samples were collected from various open Telegram groups, capturing natural voice messages in diverse acoustic conditions and speaking styles. ### Key Statistics - **Total Samples**: 745 audio files - **Total Duration**: 1 hour 40 minutes (~100 minutes) - **Average Duration**: ~8 seconds per sample - **Source**: Voice messages from various open Telegram groups - **Transcriptions**: Manually annotated ## Dataset Structure The dataset is saved as a `datasets.Dataset` object in Arrow format, containing the following fields: - `name`: Name of audio file - `audio`: Audio file data (dict with `array`, and `sampling_rate`) - `transcription`: Ground truth text transcription (manually annotated) ## Loading the Dataset ### Installation To use this dataset, you need to install the Hugging Face `datasets` library: ```bash pip install datasets ``` ### Basic Loading ```python from datasets import load_dataset # Load the dataset from the Arrow files dataset = load_dataset("OvozifyLabs/asr_evaluate_set") # View dataset information print(dataset) print(f"Number of samples: {len(dataset)}") ``` ## Data Characteristics ### Audio Properties - **Source Domain**: Conversational voice messages from Telegram - **Variability**: Multiple speakers, diverse acoustic environments - **Recording Conditions**: Real-world - **Language**: Uzbek ### Transcription Details - **Annotation Method**: Manual transcription - **Quality**: Human-verified ground truth labels - **Convention**: punctuation removed, lowercased ## Use Cases This dataset is suitable for: - Evaluating speech-to-text model performance on conversational speech - Benchmarking ASR systems on real-world voice messages - Testing model robustness to varied acoustic conditions - Comparing different STT models