Itbanque/ScreenTalk_JA2ZH
Updated β’ 4
How to use Itbanque/whisper-ja-zh-tiny with Transformers:
# Use a pipeline as a high-level helper
from transformers import pipeline
pipe = pipeline("automatic-speech-recognition", model="Itbanque/whisper-ja-zh-tiny") # Load model directly
from transformers import AutoProcessor, AutoModelForSpeechSeq2Seq
processor = AutoProcessor.from_pretrained("Itbanque/whisper-ja-zh-tiny")
model = AutoModelForSpeechSeq2Seq.from_pretrained("Itbanque/whisper-ja-zh-tiny")A fine-tuned OpenAI Whisper tiny model on Japanese-to-Chinese speech translation, trained on a subset of the DataLabX/ScreenTalk_JA2ZH dataset.
openai/whisper-tinySeq2SeqTrainertrain_batch_size: 96
eval_batch_size: 64
learning_rate: 3e-4
warmup_steps: 1000
num_train_epochs: 20
gradient_accumulation_steps: 1
save_steps: 1000
eval_steps: 1000
logging_steps: 1000
fp16: true
eval_strategy: step
early_stopping: enabled (patience=5)
Best checkpoint auto-loaded via
load_best_model_at_end=Trueusingeval_bleuas the metric.
Final run metrics (test set):
loss: 2.3245
bleu: 0.6095
Repository includes:
config.json, generation_config.json, preprocessor_config.jsontokenizer_config.json, vocab.json, merges.txt, etc.training_20250610-194336.logruns/from transformers import WhisperProcessor, WhisperForConditionalGeneration
processor = WhisperProcessor.from_pretrained("fj11/whisper-ja-zh-tiny")
model = WhisperForConditionalGeneration.from_pretrained("fj11/whisper-ja-zh-tiny")
For business inquiries or collaboration, visit https://www.itbanque.com or reach out via Hugging Face.
CC BY-NC-SA 4.0 (Non-commercial, Attribution, ShareAlike)