Whisper Large V3 Turbo (CTranslate2) β€” Optimized for Faster-Whisper & WhisperX

This repository contains a CTranslate2 (CT2) optimized version of the whisper-large-v3-turbo model finetuned by SadeghK.
It is designed for high-speed inference, low-latency ASR, and full WhisperX compatibility (ASR + alignment + diarization).


πŸš€ Model Overview

This is a converted version of the original SadeghK/whisper-large-v3-turbo model into CTranslate2 format, which enables:

  • βœ” Faster inference (up to 4Γ— vs PyTorch)
  • βœ” Lower memory usage (supports float16 / int8 / int8_float16)
  • βœ” Full compatibility with faster-whisper
  • βœ” Full compatibility with WhisperX for:
    • ASR transcription
    • Word-level alignment
    • (optional) speaker diarization

All weights in this repository are ready-to-use, no additional conversion required.


πŸ”¬ Usage with WhisperX (ASR + alignment)

import whisperx

device = "cuda"

# ASR
asr_model = whisperx.load_model(
    "SadeghK/whisper-large-v3-turbo-ct2",
    device=device,
    compute_type="float16"
)

result = asr_model.transcribe("audio.wav")

# Alignment (example for Persian)
align_model, metadata = whisperx.load_align_model("fa", device)
aligned = whisperx.align(result["segments"], align_model, metadata, "audio.wav", device)

πŸ“ Repository Structure

whisper-large-v3-turbo-ct2/
β”‚
β”œβ”€β”€ config.json
β”œβ”€β”€ model.bin
β”œβ”€β”€ preprocessor_config.json
β”œβ”€β”€ tokenizer.json
└── vocabulary.json
Downloads last month
34
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for SadeghK/whisper-large-v3-turbo-ct2

Finetuned
(1)
this model