hubert-large-ll60k-gui

This model is a fine-tuned version of facebook/hubert-large-ll60k on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 3.0032
  • Cer: 1.0

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0003
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 16
  • optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 100
  • num_epochs: 10
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Cer
12.9219 0.4329 100 2.9691 1.0
5.9613 0.8658 200 2.9701 1.0
5.9540 1.2987 300 2.9695 1.0
5.9360 1.7316 400 2.9586 1.0
6.1205 2.1645 500 2.9350 1.0
5.9243 2.5974 600 2.9020 1.0
5.9661 3.0303 700 2.9314 1.0
5.9673 3.4632 800 2.9456 1.0
5.9804 3.8961 900 2.9916 1.0
5.9797 4.3290 1000 2.9439 1.0
5.9855 4.7619 1100 2.9363 1.0
6.0039 5.1948 1200 2.9703 1.0
5.9174 5.6277 1300 2.9385 1.0
5.9848 6.0606 1400 2.9370 1.0
6.0087 6.4935 1500 2.9646 1.0
6.0090 6.9264 1600 2.9668 1.0
6.0276 7.3593 1700 2.9554 1.0
6.0094 7.7922 1800 2.9626 1.0
5.9875 8.2251 1900 2.9753 1.0
6.0609 8.6580 2000 2.9931 1.0
6.0833 9.0909 2100 2.9994 1.0
6.0989 9.5238 2200 2.9971 1.0
6.0445 9.9567 2300 3.0032 1.0

Framework versions

  • Transformers 5.1.0
  • Pytorch 2.9.1+cu128
  • Datasets 3.6.0
  • Tokenizers 0.22.2
Downloads last month
-
Safetensors
Model size
0.3B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ctaguchi/hubert-large-ll60k-gui

Finetuned
(19)
this model