End of training
Browse files- README.md +84 -0
- generation_config.json +127 -0
README.md
ADDED
|
@@ -0,0 +1,84 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
library_name: transformers
|
| 3 |
+
language:
|
| 4 |
+
- ar
|
| 5 |
+
license: apache-2.0
|
| 6 |
+
base_model: tarteel-ai/whisper-tiny-ar-quran
|
| 7 |
+
tags:
|
| 8 |
+
- generated_from_trainer
|
| 9 |
+
datasets:
|
| 10 |
+
- numan98/synth-incorrect-verses
|
| 11 |
+
metrics:
|
| 12 |
+
- wer
|
| 13 |
+
model-index:
|
| 14 |
+
- name: Nextayah Tiny Whisper Finetuned
|
| 15 |
+
results:
|
| 16 |
+
- task:
|
| 17 |
+
name: Automatic Speech Recognition
|
| 18 |
+
type: automatic-speech-recognition
|
| 19 |
+
dataset:
|
| 20 |
+
name: Synthetic Incorrect Verses
|
| 21 |
+
type: numan98/synth-incorrect-verses
|
| 22 |
+
config: default
|
| 23 |
+
split: None
|
| 24 |
+
args: 'split: test'
|
| 25 |
+
metrics:
|
| 26 |
+
- name: Wer
|
| 27 |
+
type: wer
|
| 28 |
+
value: 22.098421541318476
|
| 29 |
+
---
|
| 30 |
+
|
| 31 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
| 32 |
+
should probably proofread and complete it, then remove this comment. -->
|
| 33 |
+
|
| 34 |
+
# Nextayah Tiny Whisper Finetuned
|
| 35 |
+
|
| 36 |
+
This model is a fine-tuned version of [tarteel-ai/whisper-tiny-ar-quran](https://huggingface.co/tarteel-ai/whisper-tiny-ar-quran) on the Synthetic Incorrect Verses dataset.
|
| 37 |
+
It achieves the following results on the evaluation set:
|
| 38 |
+
- Loss: 0.1079
|
| 39 |
+
- Wer: 22.0984
|
| 40 |
+
|
| 41 |
+
## Model description
|
| 42 |
+
|
| 43 |
+
More information needed
|
| 44 |
+
|
| 45 |
+
## Intended uses & limitations
|
| 46 |
+
|
| 47 |
+
More information needed
|
| 48 |
+
|
| 49 |
+
## Training and evaluation data
|
| 50 |
+
|
| 51 |
+
More information needed
|
| 52 |
+
|
| 53 |
+
## Training procedure
|
| 54 |
+
|
| 55 |
+
### Training hyperparameters
|
| 56 |
+
|
| 57 |
+
The following hyperparameters were used during training:
|
| 58 |
+
- learning_rate: 5e-06
|
| 59 |
+
- train_batch_size: 8
|
| 60 |
+
- eval_batch_size: 8
|
| 61 |
+
- seed: 42
|
| 62 |
+
- gradient_accumulation_steps: 2
|
| 63 |
+
- total_train_batch_size: 16
|
| 64 |
+
- optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
| 65 |
+
- lr_scheduler_type: linear
|
| 66 |
+
- lr_scheduler_warmup_steps: 500
|
| 67 |
+
- training_steps: 1500
|
| 68 |
+
- mixed_precision_training: Native AMP
|
| 69 |
+
|
| 70 |
+
### Training results
|
| 71 |
+
|
| 72 |
+
| Training Loss | Epoch | Step | Validation Loss | Wer |
|
| 73 |
+
|:-------------:|:-------:|:----:|:---------------:|:-------:|
|
| 74 |
+
| 0.0666 | 8.7788 | 500 | 0.1351 | 27.3909 |
|
| 75 |
+
| 0.0073 | 17.5487 | 1000 | 0.1090 | 23.3983 |
|
| 76 |
+
| 0.0029 | 26.3186 | 1500 | 0.1079 | 22.0984 |
|
| 77 |
+
|
| 78 |
+
|
| 79 |
+
### Framework versions
|
| 80 |
+
|
| 81 |
+
- Transformers 4.48.0
|
| 82 |
+
- Pytorch 2.5.1+cu121
|
| 83 |
+
- Datasets 3.2.0
|
| 84 |
+
- Tokenizers 0.21.0
|
generation_config.json
ADDED
|
@@ -0,0 +1,127 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"begin_suppress_tokens": [
|
| 3 |
+
220,
|
| 4 |
+
50257
|
| 5 |
+
],
|
| 6 |
+
"bos_token_id": 50257,
|
| 7 |
+
"decoder_start_token_id": 50258,
|
| 8 |
+
"eos_token_id": 50257,
|
| 9 |
+
"lang_to_id": {
|
| 10 |
+
"afrikaans": 68,
|
| 11 |
+
"albanian": 58,
|
| 12 |
+
"amharic": 75,
|
| 13 |
+
"arabic": 13,
|
| 14 |
+
"armenian": 53,
|
| 15 |
+
"assamese": 91,
|
| 16 |
+
"azerbaijani": 45,
|
| 17 |
+
"bashkir": 96,
|
| 18 |
+
"basque": 51,
|
| 19 |
+
"belarusian": 71,
|
| 20 |
+
"bengali": 43,
|
| 21 |
+
"bosnian": 56,
|
| 22 |
+
"breton": 50,
|
| 23 |
+
"bulgarian": 33,
|
| 24 |
+
"burmese": 100,
|
| 25 |
+
"cantonese": 99,
|
| 26 |
+
"castilian": 110,
|
| 27 |
+
"catalan": 11,
|
| 28 |
+
"chinese": 1,
|
| 29 |
+
"croatian": 32,
|
| 30 |
+
"czech": 24,
|
| 31 |
+
"danish": 26,
|
| 32 |
+
"dutch": 12,
|
| 33 |
+
"english": 0,
|
| 34 |
+
"estonian": 48,
|
| 35 |
+
"faroese": 79,
|
| 36 |
+
"finnish": 18,
|
| 37 |
+
"flemish": 102,
|
| 38 |
+
"french": 6,
|
| 39 |
+
"galician": 60,
|
| 40 |
+
"georgian": 70,
|
| 41 |
+
"german": 2,
|
| 42 |
+
"greek": 22,
|
| 43 |
+
"gujarati": 74,
|
| 44 |
+
"haitian": 103,
|
| 45 |
+
"haitian creole": 80,
|
| 46 |
+
"hausa": 95,
|
| 47 |
+
"hawaiian": 93,
|
| 48 |
+
"hebrew": 20,
|
| 49 |
+
"hindi": 17,
|
| 50 |
+
"hungarian": 27,
|
| 51 |
+
"icelandic": 52,
|
| 52 |
+
"indonesian": 16,
|
| 53 |
+
"italian": 15,
|
| 54 |
+
"japanese": 7,
|
| 55 |
+
"javanese": 97,
|
| 56 |
+
"kannada": 47,
|
| 57 |
+
"kazakh": 57,
|
| 58 |
+
"khmer": 64,
|
| 59 |
+
"korean": 5,
|
| 60 |
+
"lao": 77,
|
| 61 |
+
"latin": 35,
|
| 62 |
+
"latvian": 42,
|
| 63 |
+
"letzeburgesch": 104,
|
| 64 |
+
"lingala": 94,
|
| 65 |
+
"lithuanian": 34,
|
| 66 |
+
"luxembourgish": 86,
|
| 67 |
+
"macedonian": 49,
|
| 68 |
+
"malagasy": 90,
|
| 69 |
+
"malay": 23,
|
| 70 |
+
"malayalam": 37,
|
| 71 |
+
"maltese": 84,
|
| 72 |
+
"mandarin": 111,
|
| 73 |
+
"maori": 36,
|
| 74 |
+
"marathi": 61,
|
| 75 |
+
"moldavian": 107,
|
| 76 |
+
"moldovan": 108,
|
| 77 |
+
"mongolian": 55,
|
| 78 |
+
"myanmar": 87,
|
| 79 |
+
"nepali": 54,
|
| 80 |
+
"norwegian": 29,
|
| 81 |
+
"nynorsk": 83,
|
| 82 |
+
"occitan": 69,
|
| 83 |
+
"panjabi": 106,
|
| 84 |
+
"pashto": 81,
|
| 85 |
+
"persian": 41,
|
| 86 |
+
"polish": 10,
|
| 87 |
+
"portuguese": 8,
|
| 88 |
+
"punjabi": 62,
|
| 89 |
+
"pushto": 105,
|
| 90 |
+
"romanian": 25,
|
| 91 |
+
"russian": 4,
|
| 92 |
+
"sanskrit": 85,
|
| 93 |
+
"serbian": 44,
|
| 94 |
+
"shona": 65,
|
| 95 |
+
"sindhi": 73,
|
| 96 |
+
"sinhala": 63,
|
| 97 |
+
"sinhalese": 109,
|
| 98 |
+
"slovak": 39,
|
| 99 |
+
"slovenian": 46,
|
| 100 |
+
"somali": 67,
|
| 101 |
+
"spanish": 3,
|
| 102 |
+
"sundanese": 98,
|
| 103 |
+
"swahili": 59,
|
| 104 |
+
"swedish": 14,
|
| 105 |
+
"tagalog": 89,
|
| 106 |
+
"tajik": 72,
|
| 107 |
+
"tamil": 28,
|
| 108 |
+
"tatar": 92,
|
| 109 |
+
"telugu": 40,
|
| 110 |
+
"thai": 30,
|
| 111 |
+
"tibetan": 88,
|
| 112 |
+
"turkish": 9,
|
| 113 |
+
"turkmen": 82,
|
| 114 |
+
"ukrainian": 21,
|
| 115 |
+
"urdu": 31,
|
| 116 |
+
"uzbek": 78,
|
| 117 |
+
"valencian": 101,
|
| 118 |
+
"vietnamese": 19,
|
| 119 |
+
"welsh": 38,
|
| 120 |
+
"yiddish": 76,
|
| 121 |
+
"yoruba": 66
|
| 122 |
+
},
|
| 123 |
+
"max_length": 448,
|
| 124 |
+
"pad_token_id": 50257,
|
| 125 |
+
"transformers_version": "4.48.0",
|
| 126 |
+
"use_cache": false
|
| 127 |
+
}
|