--- dataset_info: features: - name: audio dtype: audio: sampling_rate: 16000 - name: text dtype: string splits: - name: train num_bytes: 6224132402.741253 num_examples: 33607 - name: test num_bytes: 185203512.7434241 num_examples: 1000 download_size: 5848400127 dataset_size: 6409335915.484677 configs: - config_name: default data_files: - split: train path: data/train-* - split: test path: data/test-* --- # Ar-ASR ## Dataset Description This dataset is designed for Automatic Speech Recognition (ASR), focusing on Arabic speech with precise transcriptions including tashkeel (diacritics). It contains 33,607 audio samples from multiple sources: Microsoft Edge TTS API, Common Voice (validated Arabic subset), individual contributions, and manually transcribed YouTube videos (we also added the dataset [ClArTTS](https://huggingface.co/datasets/AtharvA7k/ClArTTS)). The dataset is paired with aligned Arabic text transcriptions and is intended for training and evaluating ASR models, such as OpenAI's Whisper, with an emphasis on accurate recognition of Arabic pronunciation and diacritics. - **Dataset Size**: 33,607 samples - **Audio**: 16 kHz - **Text**: Arabic transcriptions with tashkeel - **Language**: Modern Standard Arabic (MSA) ## Dataset Structure The dataset is hosted on Hugging Face and consists of two columns: - **audio**: Audio samples (arrays, 16 kHz sampling rate) - **text**: Arabic text transcriptions with tashkeel, aligned with the audio ### Example ```json { "audio": {"array": [...], "sampling_rate": 16000}, "text": "ثَلَاثَةٌ فِي المِئَةِ مِنَ المَاءِ العَذْبِ فِي الأَنْهَارِ وَالبُحَيْرَاتِ وَفِي الغِلَافِ الجَوِّيّ" } ``` ## Usage This dataset is ideal for: - Training Arabic ASR models - Evaluating transcription accuracy with tashkeel ### Loading the Dataset ```python from datasets import load_dataset dataset = load_dataset("CUAIStudents/Ar-ASR") ``` ### Training with Whisper The audio is pre-resampled to 16 kHz for Whisper compatibility: ```python from transformers import WhisperProcessor, WhisperForConditionalGeneration processor = WhisperProcessor.from_pretrained("openai/whisper-tiny") model = WhisperForConditionalGeneration.from_pretrained("openai/whisper-tiny") sample = dataset["train"][0] inputs = processor(sample["audio"]["array"], sampling_rate=16000, return_tensors="pt") ``` ## Limitations - **Quality**: Downsampling to 16 kHz may reduce high-frequency details, but speech remains clear - **Scope**: Includes synthetic `edge_tts` voices, Common Voice validated Arabic subset, individual contributions, and manually transcribed YouTube videos, which may vary in recording quality