Whisper Large V3 Turbo (CTranslate2) β Optimized for Faster-Whisper & WhisperX
This repository contains a CTranslate2 (CT2) optimized version of the whisper-large-v3-turbo model finetuned by SadeghK.
It is designed for high-speed inference, low-latency ASR, and full WhisperX compatibility (ASR + alignment + diarization).
π Model Overview
This is a converted version of the original SadeghK/whisper-large-v3-turbo model into CTranslate2 format, which enables:
- β Faster inference (up to 4Γ vs PyTorch)
- β Lower memory usage (supports float16 / int8 / int8_float16)
- β Full compatibility with faster-whisper
- β Full compatibility with WhisperX for:
- ASR transcription
- Word-level alignment
- (optional) speaker diarization
All weights in this repository are ready-to-use, no additional conversion required.
π¬ Usage with WhisperX (ASR + alignment)
import whisperx
device = "cuda"
# ASR
asr_model = whisperx.load_model(
"SadeghK/whisper-large-v3-turbo-ct2",
device=device,
compute_type="float16"
)
result = asr_model.transcribe("audio.wav")
# Alignment (example for Persian)
align_model, metadata = whisperx.load_align_model("fa", device)
aligned = whisperx.align(result["segments"], align_model, metadata, "audio.wav", device)
π Repository Structure
whisper-large-v3-turbo-ct2/
β
βββ config.json
βββ model.bin
βββ preprocessor_config.json
βββ tokenizer.json
βββ vocabulary.json
- Downloads last month
- 34
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for SadeghK/whisper-large-v3-turbo-ct2
Base model
openai/whisper-large-v3
Finetuned
openai/whisper-large-v3-turbo
Finetuned
SadeghK/whisper-large-v3-turbo