BartBana-deleteOriginal

This model is a fine-tuned version of IAmSkyDra/BARTBana_v4 on an unknown dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7018
  • Sacrebleu: 8.1356
  • Chrf++: 20.8127
  • Bertscore F1: 0.7507

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 32
  • eval_batch_size: 32
  • seed: 42
  • optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
  • lr_scheduler_type: linear
  • num_epochs: 3
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Sacrebleu Chrf++ Bertscore F1
0.7528 0.1000 2774 0.6202 3.8674 15.5174 0.7285
0.5889 0.1999 5548 0.6139 4.6054 16.4393 0.7325
0.4936 0.2999 8322 0.6120 5.4362 17.7869 0.7403
0.4208 0.3999 11096 0.6272 5.6644 17.9396 0.7393
0.3632 0.4998 13870 0.6338 5.9257 18.4708 0.7442
0.3166 0.5998 16644 0.6397 6.2413 18.7404 0.7449
0.282 0.6998 19418 0.6516 6.4587 18.9960 0.7467
0.2548 0.7997 22192 0.6615 6.6110 18.9584 0.7449
0.2298 0.8997 24966 0.6724 7.0878 19.5834 0.7456
0.2129 0.9997 27740 0.6713 7.3080 19.8304 0.7474
0.1862 1.0996 30514 0.6847 7.4609 20.1005 0.7493
0.1749 1.1996 33288 0.6923 7.4167 19.9609 0.7475
0.1658 1.2996 36062 0.6889 7.5104 20.2113 0.7499
0.1566 1.3995 38836 0.6898 7.7014 20.3499 0.7492
0.1485 1.4995 41610 0.6941 7.7123 20.1927 0.7496
0.1424 1.5995 44384 0.6968 7.8026 20.3893 0.7484
0.1374 1.6994 47158 0.7047 7.8897 20.4496 0.7487
0.1321 1.7994 49932 0.7002 8.0187 20.6740 0.7505
0.1278 1.8994 52706 0.7018 8.1356 20.8127 0.7507
0.1237 1.9994 55480 0.7031 8.1342 20.8311 0.7508
0.1118 2.0993 58254 0.7171 8.0924 20.7461 0.7509
0.1095 2.1993 61028 0.7181 8.1214 20.7349 0.7500

Framework versions

  • Transformers 4.57.1
  • Pytorch 2.9.0+cu128
  • Datasets 4.4.1
  • Tokenizers 0.22.1
Downloads last month
-
Safetensors
Model size
0.4B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for FiveC/BartBana-deleteOriginal

Finetuned
(10)
this model