T5_small_sum_30_epoch

This model is a fine-tuned version of t5-small on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.9724
  • Rouge1: 0.4391
  • Rouge2: 0.2715
  • Rougel: 0.4056
  • Rougelsum: 0.4053
  • Gen Len: 17.5469

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 30
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
No log 1.0 400 2.3489 0.3892 0.2234 0.3514 0.3512 18.2775
2.8157 2.0 800 2.2388 0.4043 0.2353 0.3675 0.3672 17.8419
2.5008 3.0 1200 2.1871 0.4146 0.2446 0.3782 0.378 17.8312
2.4062 4.0 1600 2.1500 0.416 0.2475 0.3808 0.3806 17.7606
2.3515 5.0 2000 2.1213 0.4182 0.2479 0.3821 0.3817 17.83
2.3515 6.0 2400 2.0984 0.4236 0.2531 0.3889 0.3886 17.7031
2.2997 7.0 2800 2.0788 0.4245 0.2555 0.3906 0.3905 17.6712
2.2606 8.0 3200 2.0643 0.4271 0.2569 0.3922 0.3921 17.6825
2.2363 9.0 3600 2.0530 0.4291 0.2581 0.394 0.3939 17.6062
2.2016 10.0 4000 2.0378 0.4315 0.2618 0.3958 0.3957 17.5869
2.2016 11.0 4400 2.0287 0.4326 0.2629 0.3982 0.398 17.5612
2.1758 12.0 4800 2.0241 0.4328 0.2634 0.398 0.3978 17.5962
2.1502 13.0 5200 2.0145 0.4341 0.2651 0.3995 0.3994 17.56
2.1444 14.0 5600 2.0094 0.4346 0.2659 0.3994 0.3995 17.5831
2.1183 15.0 6000 2.0039 0.4351 0.2678 0.4008 0.4006 17.5812
2.1183 16.0 6400 1.9987 0.4343 0.2667 0.3998 0.3997 17.5225
2.1133 17.0 6800 1.9967 0.4342 0.2674 0.4004 0.4005 17.5544
2.0918 18.0 7200 1.9900 0.4357 0.2681 0.4014 0.4013 17.5419
2.0739 19.0 7600 1.9879 0.4365 0.2686 0.4029 0.4026 17.5469
2.0733 20.0 8000 1.9831 0.4378 0.2699 0.403 0.4029 17.5481
2.0733 21.0 8400 1.9818 0.4378 0.2705 0.4037 0.4037 17.5319
2.0657 22.0 8800 1.9791 0.4375 0.2703 0.4037 0.4037 17.5225
2.0412 23.0 9200 1.9792 0.4363 0.27 0.4026 0.4023 17.5581
2.0514 24.0 9600 1.9765 0.4381 0.2703 0.4041 0.4039 17.5262
2.047 25.0 10000 1.9764 0.4396 0.2716 0.4056 0.4055 17.5525
2.047 26.0 10400 1.9744 0.4388 0.2716 0.4054 0.4051 17.5675
2.0279 27.0 10800 1.9733 0.4397 0.2715 0.4057 0.4054 17.5494
2.0503 28.0 11200 1.9730 0.4391 0.2711 0.4055 0.4052 17.5456
2.0278 29.0 11600 1.9726 0.439 0.2712 0.4056 0.4053 17.5388
2.0322 30.0 12000 1.9724 0.4391 0.2715 0.4056 0.4053 17.5469

Framework versions

  • Transformers 4.41.1
  • Pytorch 1.13.1+cu117
  • Datasets 2.19.1
  • Tokenizers 0.19.1
Downloads last month
8
Safetensors
Model size
60.5M params
Tensor type
F32
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for zera09/T5_small_sum_30_epoch

Base model

google-t5/t5-small
Finetuned
(2213)
this model

Evaluation results