flan-t5-base-samsum

This model is a fine-tuned version of google/flan-t5-base on the samsum dataset. It achieves the following results on the evaluation set:

  • Loss: 1.3743
  • Rouge1: 47.2194
  • Rouge2: 23.4461
  • Rougel: 39.8137
  • Rougelsum: 43.453
  • Gen Len: 17.2466

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 5

Training results

Training Loss Epoch Step Validation Loss Rouge1 Rouge2 Rougel Rougelsum Gen Len
1.4528 1.0 1842 1.3867 46.6599 22.7188 39.0092 42.8594 17.3993
1.3385 2.0 3684 1.3751 46.8467 23.2155 39.1262 43.0446 17.3431
1.2711 3.0 5526 1.3745 47.2411 23.4594 39.6138 43.4988 17.4493
1.2284 4.0 7368 1.3743 47.2194 23.4461 39.8137 43.453 17.2466
1.2085 5.0 9210 1.3759 47.3946 23.7172 39.8245 43.6654 17.2967

Framework versions

  • Transformers 4.36.1
  • Pytorch 2.1.0+cu121
  • Datasets 2.15.0
  • Tokenizers 0.15.0
Downloads last month
2
Safetensors
Model size
0.2B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for rizvi-rahil786/flan-t5-base-samsum

Finetuned
(904)
this model

Evaluation results