--- dataset_info: features: - name: input_features sequence: sequence: float32 - name: labels sequence: int64 splits: - name: test num_bytes: 10227942104 num_examples: 6656 download_size: 2056491222 dataset_size: 10227942104 configs: - config_name: default data_files: - split: test path: data/test-* --- The input_features are nothing but the values generated after passing the dataset's audio array through a whisper processor's feature extraction and the field 'labels' consists of the tokenized(using whisper tokenizer) ground truths. The following is the link for what I did with the sarvah dataset and how I trained it on whisper-large-v3-turbo. The training steps for whisper-large-v3 are same. https://colab.research.google.com/drive/1oD0v7MWZ9WJqk7tZYThwgTUM85PTEhMN?usp=sharing