carmi commited on
Commit
701eed4
·
verified ·
1 Parent(s): 84fae40

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -1
README.md CHANGED
@@ -16,6 +16,8 @@ license: apache-2.0
16
 
17
  **Fine-tuned Whisper model for the Levantine Dialect (Israeli-Arabic)**
18
 
 
 
19
  ## Model Description
20
 
21
  This model is a fine-tuned version of [Whisper Larg v3](https://github.com/openai/whisper) tailored specifically for transcribing Levantine Arabic, focusing on the Israeli dialect. It is designed to improve automatic speech recognition (ASR) performance for this particular variant of Arabic.
@@ -35,7 +37,7 @@ The dataset used for training and fine-tuning this model consists of approximate
35
  - **Annotation**: Human-transcribed and annotated for high accuracy.
36
 
37
  ## How to Use
38
-
39
  The model is compatible with 16kHz audio input. Ensure your files are at the same sample rate for optimal results. You can load the model as follows:
40
 
41
  Will save a .vtt file with transcriptions and timestamps in audio_dir:
 
16
 
17
  **Fine-tuned Whisper model for the Levantine Dialect (Israeli-Arabic)**
18
 
19
+ Thanks to [ivrit.ai](https://github.com/ivrit-ai/ivrit.ai/tree/master) for sharing fine-tuning code scripts!
20
+
21
  ## Model Description
22
 
23
  This model is a fine-tuned version of [Whisper Larg v3](https://github.com/openai/whisper) tailored specifically for transcribing Levantine Arabic, focusing on the Israeli dialect. It is designed to improve automatic speech recognition (ASR) performance for this particular variant of Arabic.
 
37
  - **Annotation**: Human-transcribed and annotated for high accuracy.
38
 
39
  ## How to Use
40
+ The finetuned model was converted using [faster-whisper](https://github.com/SYSTRAN/faster-whisper) package to run up to 4 times faster than openai/whisper.
41
  The model is compatible with 16kHz audio input. Ensure your files are at the same sample rate for optimal results. You can load the model as follows:
42
 
43
  Will save a .vtt file with transcriptions and timestamps in audio_dir: