Commit
·
24792f9
1
Parent(s):
bbd7ef8
Update README.md
Browse files
README.md
CHANGED
|
@@ -12,202 +12,78 @@ model-index:
|
|
| 12 |
split: test
|
| 13 |
metrics:
|
| 14 |
- type: wer
|
| 15 |
-
value: 10.
|
| 16 |
name: WER
|
| 17 |
---
|
| 18 |
|
| 19 |
# Model Card for Model ID
|
| 20 |
|
| 21 |
-
|
|
|
|
|
|
|
|
|
|
| 22 |
|
|
|
|
|
|
|
| 23 |
|
| 24 |
|
| 25 |
## Model Details
|
| 26 |
|
| 27 |
-
|
|
|
|
| 28 |
|
| 29 |
-
|
|
|
|
|
|
|
|
|
|
| 30 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 31 |
|
| 32 |
-
|
| 33 |
-
- **
|
| 34 |
-
- **Shared by [optional]:** [More Information Needed]
|
| 35 |
-
- **Model type:** [More Information Needed]
|
| 36 |
-
- **Language(s) (NLP):** [More Information Needed]
|
| 37 |
-
- **License:** [More Information Needed]
|
| 38 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
| 39 |
-
|
| 40 |
-
### Model Sources [optional]
|
| 41 |
-
|
| 42 |
-
<!-- Provide the basic links for the model. -->
|
| 43 |
-
|
| 44 |
-
- **Repository:** [More Information Needed]
|
| 45 |
-
- **Paper [optional]:** [More Information Needed]
|
| 46 |
-
- **Demo [optional]:** [More Information Needed]
|
| 47 |
|
| 48 |
## Uses
|
| 49 |
|
| 50 |
-
|
| 51 |
-
|
| 52 |
-
### Direct Use
|
| 53 |
-
|
| 54 |
-
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 55 |
-
|
| 56 |
-
[More Information Needed]
|
| 57 |
-
|
| 58 |
-
### Downstream Use [optional]
|
| 59 |
-
|
| 60 |
-
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 61 |
-
|
| 62 |
-
[More Information Needed]
|
| 63 |
|
| 64 |
### Out-of-Scope Use
|
| 65 |
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
[More Information Needed]
|
| 69 |
-
|
| 70 |
-
## Bias, Risks, and Limitations
|
| 71 |
-
|
| 72 |
-
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
| 73 |
-
|
| 74 |
-
[More Information Needed]
|
| 75 |
|
| 76 |
-
|
| 77 |
|
| 78 |
-
|
| 79 |
|
| 80 |
-
|
| 81 |
-
|
| 82 |
-
## How to Get Started with the Model
|
| 83 |
-
|
| 84 |
-
Use the code below to get started with the model.
|
| 85 |
-
|
| 86 |
-
[More Information Needed]
|
| 87 |
|
| 88 |
## Training Details
|
| 89 |
|
| 90 |
-
### Training Data
|
| 91 |
-
|
| 92 |
-
<!-- This should link to a Data Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 93 |
-
|
| 94 |
-
[More Information Needed]
|
| 95 |
-
|
| 96 |
### Training Procedure
|
| 97 |
|
| 98 |
-
|
| 99 |
-
|
| 100 |
-
#### Preprocessing [optional]
|
| 101 |
-
|
| 102 |
-
[More Information Needed]
|
| 103 |
-
|
| 104 |
|
| 105 |
#### Training Hyperparameters
|
| 106 |
|
| 107 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 108 |
|
| 109 |
-
#### Speeds, Sizes, Times [optional]
|
| 110 |
-
|
| 111 |
-
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 112 |
-
|
| 113 |
-
[More Information Needed]
|
| 114 |
|
| 115 |
## Evaluation
|
| 116 |
|
| 117 |
-
|
| 118 |
-
|
| 119 |
-
### Testing Data, Factors & Metrics
|
| 120 |
-
|
| 121 |
-
#### Testing Data
|
| 122 |
-
|
| 123 |
-
<!-- This should link to a Data Card if possible. -->
|
| 124 |
-
|
| 125 |
-
[More Information Needed]
|
| 126 |
-
|
| 127 |
-
#### Factors
|
| 128 |
-
|
| 129 |
-
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 130 |
-
|
| 131 |
-
[More Information Needed]
|
| 132 |
-
|
| 133 |
-
#### Metrics
|
| 134 |
-
|
| 135 |
-
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
| 136 |
-
|
| 137 |
-
[More Information Needed]
|
| 138 |
-
|
| 139 |
-
### Results
|
| 140 |
-
|
| 141 |
-
[More Information Needed]
|
| 142 |
-
|
| 143 |
-
#### Summary
|
| 144 |
-
|
| 145 |
-
|
| 146 |
-
|
| 147 |
-
## Model Examination [optional]
|
| 148 |
-
|
| 149 |
-
<!-- Relevant interpretability work for the model goes here -->
|
| 150 |
-
|
| 151 |
-
[More Information Needed]
|
| 152 |
-
|
| 153 |
-
## Environmental Impact
|
| 154 |
-
|
| 155 |
-
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 156 |
-
|
| 157 |
-
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 158 |
-
|
| 159 |
-
- **Hardware Type:** [More Information Needed]
|
| 160 |
-
- **Hours used:** [More Information Needed]
|
| 161 |
-
- **Cloud Provider:** [More Information Needed]
|
| 162 |
-
- **Compute Region:** [More Information Needed]
|
| 163 |
-
- **Carbon Emitted:** [More Information Needed]
|
| 164 |
-
|
| 165 |
-
## Technical Specifications [optional]
|
| 166 |
-
|
| 167 |
-
### Model Architecture and Objective
|
| 168 |
-
|
| 169 |
-
[More Information Needed]
|
| 170 |
-
|
| 171 |
-
### Compute Infrastructure
|
| 172 |
-
|
| 173 |
-
[More Information Needed]
|
| 174 |
-
|
| 175 |
-
#### Hardware
|
| 176 |
-
|
| 177 |
-
[More Information Needed]
|
| 178 |
-
|
| 179 |
-
#### Software
|
| 180 |
-
|
| 181 |
-
[More Information Needed]
|
| 182 |
-
|
| 183 |
-
## Citation [optional]
|
| 184 |
-
|
| 185 |
-
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 186 |
-
|
| 187 |
-
**BibTeX:**
|
| 188 |
-
|
| 189 |
-
[More Information Needed]
|
| 190 |
-
|
| 191 |
-
**APA:**
|
| 192 |
-
|
| 193 |
-
[More Information Needed]
|
| 194 |
-
|
| 195 |
-
## Glossary [optional]
|
| 196 |
-
|
| 197 |
-
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 198 |
-
|
| 199 |
-
[More Information Needed]
|
| 200 |
-
|
| 201 |
-
## More Information [optional]
|
| 202 |
-
|
| 203 |
-
[More Information Needed]
|
| 204 |
-
|
| 205 |
-
## Model Card Authors [optional]
|
| 206 |
-
|
| 207 |
-
[More Information Needed]
|
| 208 |
|
| 209 |
-
## Model Card Contact
|
| 210 |
|
| 211 |
-
|
|
|
|
|
|
|
| 212 |
|
| 213 |
|
|
|
|
| 12 |
split: test
|
| 13 |
metrics:
|
| 14 |
- type: wer
|
| 15 |
+
value: 10.30
|
| 16 |
name: WER
|
| 17 |
---
|
| 18 |
|
| 19 |
# Model Card for Model ID
|
| 20 |
|
| 21 |
+
This is a finetune for Whisper Small. A finetune to achieve better results on Whisper Small for Portuguese.
|
| 22 |
+
Whisper is a pre-trained model for automatic speech recognition (ASR) and speech translation. Trained on 680k hours
|
| 23 |
+
of labelled data, Whisper models demonstrate a strong ability to generalise to many datasets and domains **without** the need
|
| 24 |
+
for fine-tuning.
|
| 25 |
|
| 26 |
+
Whisper was proposed in the paper [Robust Speech Recognition via Large-Scale Weak Supervision](https://arxiv.org/abs/2212.04356)
|
| 27 |
+
by Alec Radford et al from OpenAI. The original code repository can be found [here](https://github.com/openai/whisper).
|
| 28 |
|
| 29 |
|
| 30 |
## Model Details
|
| 31 |
|
| 32 |
+
Whisper is a Transformer based encoder-decoder model, also referred to as a _sequence-to-sequence_ model.
|
| 33 |
+
It was trained on 680k hours of labelled speech data annotated using large-scale weak supervision.
|
| 34 |
|
| 35 |
+
The models were trained on either English-only data or multilingual data. The English-only models were trained
|
| 36 |
+
on the task of speech recognition. The multilingual models were trained on both speech recognition and speech
|
| 37 |
+
translation. For speech recognition, the model predicts transcriptions in the *same* language as the audio.
|
| 38 |
+
For speech translation, the model predicts transcriptions to a *different* language to the audio.
|
| 39 |
|
| 40 |
+
Whisper checkpoints come in five configurations of varying model sizes.
|
| 41 |
+
The smallest four are trained on either English-only or multilingual data.
|
| 42 |
+
The largest checkpoints are multilingual only. All ten of the pre-trained checkpoints
|
| 43 |
+
are available on the [Hugging Face Hub](https://huggingface.co/models?search=openai/whisper).
|
| 44 |
|
| 45 |
+
- **Developed by:** [ArtificialGuyBr](https://twitter.com/@artificialguybr)
|
| 46 |
+
- **Shared by:** [ArtificialGuyBr](https://twitter.com/@artificialguybr)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 47 |
|
| 48 |
## Uses
|
| 49 |
|
| 50 |
+
This repository contains a fine-tuned version of the Whisper ASR (Automatic Speech Recognition) system developed by OpenAI. The model has been specifically fine-tuned to improve performance in portuguese language.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 51 |
|
| 52 |
### Out-of-Scope Use
|
| 53 |
|
| 54 |
+
While this model is powerful and versatile, it's important to understand its limitations and inappropriate uses:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 55 |
|
| 56 |
+
1. **Misuse and Malicious Use**: This model should not be used for any illegal activities, including but not limited to eavesdropping, illegal surveillance, or any other form of privacy invasion. It's also not intended for the creation or spread of misinformation, hate speech, or harmful content.
|
| 57 |
|
| 58 |
+
2. **Non-Portuguese Languages**: While this model has been fine-tuned for Portuguese, it may not perform well with other languages. It's not recommended for transcribing multilingual content where languages other than Portuguese are spoken.
|
| 59 |
|
| 60 |
+
3. **Low-Quality Audio**: The model's performance can be significantly affected by the quality of the input audio. It may not work well with low-quality audio, background noise, or speakers who are far away from the microphone.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 61 |
|
| 62 |
## Training Details
|
| 63 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 64 |
### Training Procedure
|
| 65 |
|
| 66 |
+
Trained using the code from HF Whisper Event.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 67 |
|
| 68 |
#### Training Hyperparameters
|
| 69 |
|
| 70 |
+
The following hyperparameters were used during training:
|
| 71 |
+
- learning_rate: 1e-05
|
| 72 |
+
- train_batch_size: 64
|
| 73 |
+
- eval_batch_size: 32
|
| 74 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 75 |
+
- lr_scheduler_type: linear
|
| 76 |
+
- lr_scheduler_warmup_steps: 500
|
| 77 |
+
- training_steps: 5000
|
| 78 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 79 |
|
| 80 |
## Evaluation
|
| 81 |
|
| 82 |
+
Wer on CV13.0: 10.3
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 83 |
|
|
|
|
| 84 |
|
| 85 |
+
- **Hardware Type:** 1XA100 80GB
|
| 86 |
+
- **Hours used:** 8 Hours.
|
| 87 |
+
- **Cloud Provider:** [Redmond.ai](https://redmond.ai)
|
| 88 |
|
| 89 |
|