weny22's picture
End of training
a6418ac verified
---
base_model: weny22/sum_model_t5_saved
tags:
- generated_from_trainer
metrics:
- rouge
model-index:
- name: extract_long_text_balanced_data
results: []
---
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# extract_long_text_balanced_data
This model is a fine-tuned version of [weny22/sum_model_t5_saved](https://huggingface.co/weny22/sum_model_t5_saved) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 2.3331
- Rouge1: 0.2094
- Rouge2: 0.0794
- Rougel: 0.1697
- Rougelsum: 0.1696
- Gen Len: 18.9853
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.002
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 20
### Training results
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|
| No log | 1.0 | 119 | 2.3318 | 0.18 | 0.0552 | 0.1439 | 0.1439 | 18.97 |
| No log | 2.0 | 238 | 2.1980 | 0.1899 | 0.06 | 0.1503 | 0.1503 | 18.972 |
| No log | 3.0 | 357 | 2.1448 | 0.1952 | 0.0646 | 0.1542 | 0.1541 | 18.9993 |
| No log | 4.0 | 476 | 2.1372 | 0.1983 | 0.0683 | 0.1574 | 0.1574 | 18.9453 |
| 2.8015 | 5.0 | 595 | 2.1142 | 0.2 | 0.0725 | 0.1611 | 0.1611 | 18.9933 |
| 2.8015 | 6.0 | 714 | 2.0970 | 0.2027 | 0.0757 | 0.1629 | 0.1629 | 18.9987 |
| 2.8015 | 7.0 | 833 | 2.1187 | 0.2027 | 0.0755 | 0.1637 | 0.1634 | 18.968 |
| 2.8015 | 8.0 | 952 | 2.1222 | 0.2013 | 0.0737 | 0.1619 | 0.1618 | 18.9753 |
| 2.02 | 9.0 | 1071 | 2.1316 | 0.2021 | 0.0764 | 0.1648 | 0.1647 | 18.9667 |
| 2.02 | 10.0 | 1190 | 2.1455 | 0.2109 | 0.0784 | 0.169 | 0.1689 | 18.982 |
| 2.02 | 11.0 | 1309 | 2.1580 | 0.2065 | 0.0781 | 0.167 | 0.1669 | 18.968 |
| 2.02 | 12.0 | 1428 | 2.1792 | 0.2088 | 0.0788 | 0.1693 | 0.169 | 18.9767 |
| 1.683 | 13.0 | 1547 | 2.1958 | 0.2085 | 0.0781 | 0.1689 | 0.1689 | 18.9913 |
| 1.683 | 14.0 | 1666 | 2.2436 | 0.2082 | 0.0785 | 0.1693 | 0.1692 | 18.978 |
| 1.683 | 15.0 | 1785 | 2.2480 | 0.2075 | 0.0797 | 0.1678 | 0.1678 | 18.9853 |
| 1.683 | 16.0 | 1904 | 2.2714 | 0.208 | 0.0789 | 0.1686 | 0.1685 | 18.9887 |
| 1.45 | 17.0 | 2023 | 2.2771 | 0.2091 | 0.0787 | 0.1693 | 0.1691 | 18.98 |
| 1.45 | 18.0 | 2142 | 2.2913 | 0.2103 | 0.0792 | 0.17 | 0.1698 | 18.9873 |
| 1.45 | 19.0 | 2261 | 2.3163 | 0.2094 | 0.0792 | 0.1699 | 0.1697 | 18.9893 |
| 1.45 | 20.0 | 2380 | 2.3331 | 0.2094 | 0.0794 | 0.1697 | 0.1696 | 18.9853 |
### Framework versions
- Transformers 4.38.2
- Pytorch 2.1.2+cu121
- Datasets 2.18.0
- Tokenizers 0.15.2