Whisper Fine-tuning Experiment: jdd_topic1_batch2-whisper-base

Model Description

This model contains a complete Whisper fine-tuning experiment including:

  • Training checkpoints (SpeechBrain format)
  • Final model (Transformers format)
  • Test results and evaluation metrics
  • Training history visualizations

Model Information

  • Base Model: openai/whisper-base
  • Framework: SpeechBrain v1.0.3
  • Training Dataset: noflm/jdd_topic1_batch2
  • Language: Japanese (ja)
  • Task: Automatic Speech Recognition (ASR)

Test Results

  • WER: 12.17%
  • CER: 9.08%
  • Test Loss: 0.0814

Contents

β”œβ”€β”€ checkpoints/          # Training checkpoints
β”‚   β”œβ”€β”€ CKPT+epoch_*/    # Per-epoch checkpoints
β”‚   β”œβ”€β”€ CKPT+BEST_WER/   # Best WER checkpoint
β”‚   └── CKPT+FINAL/      # Final checkpoint
β”œβ”€β”€ final_model/          # Transformers-compatible model
β”‚   β”œβ”€β”€ config.json      # Model configuration
β”‚   β”œβ”€β”€ model.safetensors # Model weights
β”‚   β”œβ”€β”€ preprocessor_config.json
β”‚   β”œβ”€β”€ tokenizer_config.json
β”‚   └── ...
β”œβ”€β”€ test_results.json     # Test metrics
β”œβ”€β”€ detailed_metrics.json # Detailed training history
β”œβ”€β”€ training_history_speechbrain.png  # Training curves
└── training_report_speechbrain.txt   # Summary report

Usage

Load checkpoint (SpeechBrain format)

import torch
checkpoint = torch.load('checkpoints/CKPT+BEST_WER/model.ckpt')

Load final model (Transformers format)

from transformers import WhisperForConditionalGeneration, WhisperProcessor

model = WhisperForConditionalGeneration.from_pretrained("./final_model")
processor = WhisperProcessor.from_pretrained("./final_model")

Citation

If you use this experiment data, please cite the original Whisper paper:

@article{radford2022robust,
  title={Robust speech recognition via large-scale weak supervision},
  author={Radford, Alec and Kim, Jong Wook and Xu, Tao and Brockman, Greg and McLeavey, Christine and Sutskever, Ilya},
  journal={arXiv preprint arXiv:2212.04356},
  year={2022}
}
Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for noflm/whisper-ft-jdd-topic1-batch2-base

Finetuned
(618)
this model

Dataset used to train noflm/whisper-ft-jdd-topic1-batch2-base

Paper for noflm/whisper-ft-jdd-topic1-batch2-base