mohammed's picture
Update README.md
e14fb0d verified
---
library_name: transformers
tags: []
---
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
This is a quantized model of the original version mohammed/whisper-small-arabic-cv-11
- **Developed by:** Mohammed Bakheet
- **Funded by [optional]:** Kalam Technology
- **Language(s) (NLP):** Arabic, English
## Uses
This a quantized model that reads arabic voice and transcribes/translate it into english
### Direct Use
First, install the following packages using the following commands:
pip install -U optimum[exporters,onnxruntime] transformers
pip install huggingface_hub
```python
# uncomment the following installation if you are using a notebook:
#!pip install -U optimum[exporters,onnxruntime] transformers
#!pip install huggingface_hub
# import the required packages
from optimum.onnxruntime import ORTModelForSpeechSeq2Seq
from transformers import WhisperTokenizerFast, WhisperFeatureExtractor, pipeline
# set model name/id
model_name = 'mohammed/quantized-whisper-small' # folder name
model = ORTModelForSpeechSeq2Seq.from_pretrained(model_name, export=False)
tokenizer = WhisperTokenizerFast.from_pretrained(model_name)
feature_extractor = WhisperFeatureExtractor.from_pretrained(model_name)
forced_decoder_ids = tokenizer.get_decoder_prompt_ids(language="ar", task="transcribe")
pipe = pipeline('automatic-speech-recognition',
model=model,
tokenizer=tokenizer,
feature_extractor=feature_extractor,
model_kwargs={"forced_decoder_ids": forced_decoder_ids})
# the file to be transcribed
pipe('Recording.mp3')
```
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
The model does a direct translation of Arabic speech, and doesn't do a direct transcription, we are still working on that.
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
```python
First, install the following packages using the following commands:
pip install -U optimum[exporters,onnxruntime] transformers
pip install huggingface_hub
from optimum.onnxruntime import ORTModelForSpeechSeq2Seq
from transformers import WhisperTokenizerFast, WhisperFeatureExtractor, pipeline
model_name = 'mohammed/quantized-whisper-small' # folder name
model = ORTModelForSpeechSeq2Seq.from_pretrained(model_name, export=False)
tokenizer = WhisperTokenizerFast.from_pretrained(model_name)
feature_extractor = WhisperFeatureExtractor.from_pretrained(model_name)
forced_decoder_ids = tokenizer.get_decoder_prompt_ids(language="ar", task="transcribe")
pipe = pipeline('automatic-speech-recognition',
model=model,
tokenizer=tokenizer,
feature_extractor=feature_extractor,
model_kwargs={"forced_decoder_ids": forced_decoder_ids})
# the file to be transcribed
pipe('Recording.mp3')
```
### Training Data
Please refer to the original model at "mohammed/whisper-small-arabic-cv-11"
### Training Procedure
Please refer to the original model at "mohammed/whisper-small-arabic-cv-11"
#### Preprocessing [optional]
Please refer to the original model at "mohammed/whisper-small-arabic-cv-11"
#### Training Hyperparameters
- **Training regime:** Please refer to the original model at "mohammed/whisper-small-arabic-cv-11"