Faster Whisper Small Turkish (INT8 Quantized)

This model is an optimized CTranslate2 (INT8) conversion of the robust Turkish ASR model ogulcanakca/whisper-small-tr.

Performance Benchmarks

The model was benchmarked against the original Hugging Face Transformers implementation on an NVIDIA A100 GPU.

Model Format Precision Inference Time (Avg) Speedup Factor
Original PyTorch FP16 10.35 sec 1x (Baseline)
Faster-Whisper INT8 0.54 sec 19.2x Faster

Note: Benchmarks were conducted on a standard Common Voice audio sample.

Usage

To use this model, you need the faster-whisper library.

pip install faster-whisper
from faster_whisper import WhisperModel

model_id = "ogulcanakca/faster-whisper-small-tr"

# Run on GPU with INT8
model = WhisperModel(model_id, device="cuda", compute_type="int8")

# or Run on CPU with INT8 (High performance on CPU too!)
# model = WhisperModel(model_id, device="cpu", compute_type="int8")

segments, info = model.transcribe("audio.mp3", beam_size=5, language="tr")

print(f"Detected language '{info.language}' with probability {info.language_probability}")

for segment in segments:
    print(f"[{segment.start:.2f}s -> {segment.end:.2f}s] {segment.text}")

Model Details

  • Base Model: ogulcanakca/whisper-small-tr (Fine-tuned on Common Voice 23.0 with JIT Augmentation)
  • Quantization: 8-bit Integer (INT8)
  • Backend: CTranslate2
  • Objective: Low-latency real-time streaming and high-throughput batch processing.
Downloads last month
53
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ogulcanakca/faster-whisper-small-tr

Quantized
(1)
this model

Space using ogulcanakca/faster-whisper-small-tr 1

Evaluation results