Update README.md
Browse files
README.md
CHANGED
|
@@ -157,4 +157,66 @@ pipeline_tag: automatic-speech-recognition
|
|
| 157 |
license: apache-2.0
|
| 158 |
base_model:
|
| 159 |
- openai/whisper-tiny
|
| 160 |
-
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 157 |
license: apache-2.0
|
| 158 |
base_model:
|
| 159 |
- openai/whisper-tiny
|
| 160 |
+
---
|
| 161 |
+
|
| 162 |
+
# Open4bits / Whisper Tiny FP16
|
| 163 |
+
|
| 164 |
+
This repository provides the **Whisper Tiny model converted to FP16 (float16) precision**, published by Open4bits to enable highly efficient inference with minimal memory usage.
|
| 165 |
+
|
| 166 |
+
The underlying Whisper model and architecture are **owned by OpenAI**. This repository contains only a precision-converted version of the original model weights.
|
| 167 |
+
|
| 168 |
+
The model is designed for fast, lightweight multilingual speech-to-text tasks and is well suited for resource-constrained environments.
|
| 169 |
+
|
| 170 |
+
---
|
| 171 |
+
|
| 172 |
+
## Model Overview
|
| 173 |
+
|
| 174 |
+
Whisper is a sequence-to-sequence transformer model developed by OpenAI for automatic speech recognition and speech translation.
|
| 175 |
+
This release uses the **Tiny** variant, prioritizing speed and low memory usage while preserving the original architecture.
|
| 176 |
+
|
| 177 |
+
---
|
| 178 |
+
|
| 179 |
+
## Model Details
|
| 180 |
+
|
| 181 |
+
- **Architecture:** Whisper Tiny
|
| 182 |
+
- **Parameters:** ~37.85 million
|
| 183 |
+
- **Precision:** float16 (FP16)
|
| 184 |
+
- **Task:** Automatic Speech Recognition (ASR)
|
| 185 |
+
- **Languages:** Multilingual
|
| 186 |
+
- **Weight tying:** Preserved
|
| 187 |
+
- **Compatibility:** Hugging Face Transformers, PyTorch
|
| 188 |
+
|
| 189 |
+
Compared to larger Whisper variants, this model offers significantly faster inference and lower VRAM requirements, with reduced accuracy in some scenarios.
|
| 190 |
+
|
| 191 |
+
---
|
| 192 |
+
|
| 193 |
+
## Intended Use
|
| 194 |
+
|
| 195 |
+
This model is intended for:
|
| 196 |
+
- Fast speech-to-text transcription
|
| 197 |
+
- Lightweight and real-time ASR applications
|
| 198 |
+
- Edge or low-resource deployments
|
| 199 |
+
- Research and prototyping
|
| 200 |
+
|
| 201 |
+
---
|
| 202 |
+
|
| 203 |
+
## Limitations
|
| 204 |
+
|
| 205 |
+
* Lower transcription accuracy compared to larger Whisper variants
|
| 206 |
+
* Performance depends on audio quality, language, and accent
|
| 207 |
+
* Not fine-tuned for domain-specific or noisy audio
|
| 208 |
+
|
| 209 |
+
---
|
| 210 |
+
|
| 211 |
+
## License
|
| 212 |
+
|
| 213 |
+
This model is released under the **Apache License 2.0**.
|
| 214 |
+
The original Whisper model and associated intellectual property are owned by OpenAI.
|
| 215 |
+
|
| 216 |
+
---
|
| 217 |
+
|
| 218 |
+
## Support
|
| 219 |
+
|
| 220 |
+
If you find this model useful, please consider supporting the project.
|
| 221 |
+
Your support helps us continue releasing and maintaining high-quality open models.
|
| 222 |
+
Support us with a heart.
|