--- base_model: weny22/sum_model_t5_saved tags: - generated_from_trainer metrics: - rouge model-index: - name: extract_long_text_balanced_data results: [] --- # extract_long_text_balanced_data This model is a fine-tuned version of [weny22/sum_model_t5_saved](https://huggingface.co/weny22/sum_model_t5_saved) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.3331 - Rouge1: 0.2094 - Rouge2: 0.0794 - Rougel: 0.1697 - Rougelsum: 0.1696 - Gen Len: 18.9853 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.002 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 20 ### Training results | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:| | No log | 1.0 | 119 | 2.3318 | 0.18 | 0.0552 | 0.1439 | 0.1439 | 18.97 | | No log | 2.0 | 238 | 2.1980 | 0.1899 | 0.06 | 0.1503 | 0.1503 | 18.972 | | No log | 3.0 | 357 | 2.1448 | 0.1952 | 0.0646 | 0.1542 | 0.1541 | 18.9993 | | No log | 4.0 | 476 | 2.1372 | 0.1983 | 0.0683 | 0.1574 | 0.1574 | 18.9453 | | 2.8015 | 5.0 | 595 | 2.1142 | 0.2 | 0.0725 | 0.1611 | 0.1611 | 18.9933 | | 2.8015 | 6.0 | 714 | 2.0970 | 0.2027 | 0.0757 | 0.1629 | 0.1629 | 18.9987 | | 2.8015 | 7.0 | 833 | 2.1187 | 0.2027 | 0.0755 | 0.1637 | 0.1634 | 18.968 | | 2.8015 | 8.0 | 952 | 2.1222 | 0.2013 | 0.0737 | 0.1619 | 0.1618 | 18.9753 | | 2.02 | 9.0 | 1071 | 2.1316 | 0.2021 | 0.0764 | 0.1648 | 0.1647 | 18.9667 | | 2.02 | 10.0 | 1190 | 2.1455 | 0.2109 | 0.0784 | 0.169 | 0.1689 | 18.982 | | 2.02 | 11.0 | 1309 | 2.1580 | 0.2065 | 0.0781 | 0.167 | 0.1669 | 18.968 | | 2.02 | 12.0 | 1428 | 2.1792 | 0.2088 | 0.0788 | 0.1693 | 0.169 | 18.9767 | | 1.683 | 13.0 | 1547 | 2.1958 | 0.2085 | 0.0781 | 0.1689 | 0.1689 | 18.9913 | | 1.683 | 14.0 | 1666 | 2.2436 | 0.2082 | 0.0785 | 0.1693 | 0.1692 | 18.978 | | 1.683 | 15.0 | 1785 | 2.2480 | 0.2075 | 0.0797 | 0.1678 | 0.1678 | 18.9853 | | 1.683 | 16.0 | 1904 | 2.2714 | 0.208 | 0.0789 | 0.1686 | 0.1685 | 18.9887 | | 1.45 | 17.0 | 2023 | 2.2771 | 0.2091 | 0.0787 | 0.1693 | 0.1691 | 18.98 | | 1.45 | 18.0 | 2142 | 2.2913 | 0.2103 | 0.0792 | 0.17 | 0.1698 | 18.9873 | | 1.45 | 19.0 | 2261 | 2.3163 | 0.2094 | 0.0792 | 0.1699 | 0.1697 | 18.9893 | | 1.45 | 20.0 | 2380 | 2.3331 | 0.2094 | 0.0794 | 0.1697 | 0.1696 | 18.9853 | ### Framework versions - Transformers 4.38.2 - Pytorch 2.1.2+cu121 - Datasets 2.18.0 - Tokenizers 0.15.2