Whisper Large V2 (eole)

This is openai/whisper-large-v2 converted to eole format using eole convert --model_dir openai/whisper-large-v2.

No weights were modified — this is a format conversion only.

Model details

Original model openai/whisper-large-v2
Parameters 1.55B
Encoder layers 32
Decoder layers 32
Hidden size 1280
Attention heads 20
Mel bins 80
Vocab size 51,865
License Apache 2.0

Usage

pip install eole[wer]

Transcribe

eole predict \
  -config eval_config.yaml \
  -model_path whisper-large-v2-eole \
  -src audio_files.txt \
  -output transcriptions.txt \
  -language en \
  -task transcribe \
  -gpu_ranks 0

Evaluation

All evaluations use beam size 5.

Benchmark WER
LibriSpeech test-clean 2.44%

Conversion

eole convert --model_dir openai/whisper-large-v2 --output whisper-large-v2-eole

Citation

@misc{radford2023robust,
      title={Robust Speech Recognition via Large-Scale Weak Supervision},
      author={Alec Radford and Jong Wook Kim and Tao Xu and Greg Brockman and Christine McLeavey and Ilya Sutskever},
      year={2023},
      eprint={2212.04356},
      archivePrefix={arXiv},
      primaryClass={eess.AS}
}
Downloads last month
-
Safetensors
Model size
2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for davidmeikle/whisper-large-v2-eole

Finetuned
(249)
this model

Paper for davidmeikle/whisper-large-v2-eole