BhBHT5New

This model is a fine-tuned version of Sabbir772/BhBHT5New on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1106
  • Bleu: 53.6682
  • Chrf: 76.5616

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 2

Training results

Training Loss Epoch Step Validation Loss Bleu Chrf
0.799 0.2851 400 1.1026 53.4427 76.3587
0.768 0.5702 800 1.0923 54.0598 76.5200
0.7882 0.8553 1200 1.1048 53.8004 76.4712
0.7507 1.1404 1600 1.1046 53.4656 76.4324
0.7357 1.4255 2000 1.1049 52.9521 76.1502
0.7166 1.7106 2400 1.1104 53.7297 76.6875
0.7087 1.9957 2800 1.1106 53.6682 76.5616

Framework versions

  • Transformers 4.57.1
  • Pytorch 2.8.0+cu126
  • Datasets 4.0.0
  • Tokenizers 0.22.1
Downloads last month
-
Safetensors
Model size
0.2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Sabbir772/BhBHT5New

Unable to build the model tree, the base model loops to the model itself. Learn more.