--- language: - en - zh - de - es - ru - ko - fr - ja - pt - tr - pl - ca - nl - ar - sv - it - id - hi - fi - vi - he - uk - el - ms - cs - ro - da - hu - ta - 'no' - th - ur - hr - bg - lt - la - mi - ml - cy - sk - te - fa - lv - bn - sr - az - sl - kn - et - mk - br - eu - is - hy - ne - mn - bs - kk - sq - sw - gl - mr - pa - si - km - sn - yo - so - af - oc - ka - be - tg - sd - gu - am - yi - lo - uz - fo - ht - ps - tk - nn - mt - sa - lb - my - bo - tl - mg - as - tt - haw - ln - ha - ba - jw - su tags: - audio - automatic-speech-recognition - hf-asr-leaderboard - open4bits widget: - example_title: Librispeech sample 1 src: https://cdn-media.huggingface.co/speech_samples/sample1.flac - example_title: Librispeech sample 2 src: https://cdn-media.huggingface.co/speech_samples/sample2.flac model-index: - name: whisper-tiny results: - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: LibriSpeech (clean) type: librispeech_asr config: clean split: test args: language: en metrics: - name: Test WER type: wer value: 7.54 - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: LibriSpeech (other) type: librispeech_asr config: other split: test args: language: en metrics: - name: Test WER type: wer value: 17.15 - task: name: Automatic Speech Recognition type: automatic-speech-recognition dataset: name: Common Voice 11.0 type: mozilla-foundation/common_voice_11_0 config: hi split: test args: language: hi metrics: - name: Test WER type: wer value: 141 pipeline_tag: automatic-speech-recognition license: apache-2.0 base_model: - openai/whisper-tiny --- # Open4bits / Whisper Tiny FP16 This repository provides the **Whisper Tiny model converted to FP16 (float16) precision**, published by Open4bits to enable highly efficient inference with minimal memory usage. The underlying Whisper model and architecture are **owned by OpenAI**. This repository contains only a precision-converted version of the original model weights. The model is designed for fast, lightweight multilingual speech-to-text tasks and is well suited for resource-constrained environments. --- ## Model Overview Whisper is a sequence-to-sequence transformer model developed by OpenAI for automatic speech recognition and speech translation. This release uses the **Tiny** variant, prioritizing speed and low memory usage while preserving the original architecture. --- ## Model Details - **Architecture:** Whisper Tiny - **Parameters:** ~37.85 million - **Precision:** float16 (FP16) - **Task:** Automatic Speech Recognition (ASR) - **Languages:** Multilingual - **Weight tying:** Preserved - **Compatibility:** Hugging Face Transformers, PyTorch Compared to larger Whisper variants, this model offers significantly faster inference and lower VRAM requirements, with reduced accuracy in some scenarios. --- ## Intended Use This model is intended for: - Fast speech-to-text transcription - Lightweight and real-time ASR applications - Edge or low-resource deployments - Research and prototyping --- ## Limitations * Lower transcription accuracy compared to larger Whisper variants * Performance depends on audio quality, language, and accent * Not fine-tuned for domain-specific or noisy audio --- ## License This model is released under the **Apache License 2.0**. The original Whisper model and associated intellectual property are owned by OpenAI. --- ## Support If you find this model useful, please consider supporting the project. Your support helps us continue releasing and maintaining high-quality open models. Support us with a heart.