b420da0d4015d59dc273e790e147248c
This model is a fine-tuned version of google-t5/t5-base on the Helsinki-NLP/opus_books [es-fr] dataset. It achieves the following results on the evaluation set:
- Loss: 1.2265
- Data Size: 1.0
- Epoch Runtime: 342.0698
- Bleu: 12.4692
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- distributed_type: multi-GPU
- num_devices: 4
- total_train_batch_size: 32
- total_eval_batch_size: 32
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
- lr_scheduler_type: constant
- num_epochs: 50
Training results
| Training Loss | Epoch | Step | Validation Loss | Data Size | Epoch Runtime | Bleu |
|---|---|---|---|---|---|---|
| No log | 0 | 0 | 2.5940 | 0 | 25.4424 | 5.6836 |
| No log | 1 | 1407 | 2.3556 | 0.0078 | 27.9662 | 7.2555 |
| No log | 2 | 2814 | 2.1227 | 0.0156 | 31.0816 | 6.9116 |
| 0.0588 | 3 | 4221 | 1.9865 | 0.0312 | 37.5414 | 6.4007 |
| 2.1763 | 4 | 5628 | 1.8951 | 0.0625 | 45.2317 | 5.8319 |
| 2.0593 | 5 | 7035 | 1.7950 | 0.125 | 64.8266 | 6.6318 |
| 1.8833 | 6 | 8442 | 1.6827 | 0.25 | 103.5913 | 7.5779 |
| 1.793 | 7 | 9849 | 1.5686 | 0.5 | 185.0851 | 8.4030 |
| 1.6179 | 8.0 | 11256 | 1.4512 | 1.0 | 341.8349 | 9.5631 |
| 1.5134 | 9.0 | 12663 | 1.3847 | 1.0 | 371.6036 | 10.3144 |
| 1.4902 | 10.0 | 14070 | 1.3443 | 1.0 | 347.1290 | 10.8170 |
| 1.3874 | 11.0 | 15477 | 1.3189 | 1.0 | 387.9105 | 11.0464 |
| 1.3334 | 12.0 | 16884 | 1.2923 | 1.0 | 342.7037 | 11.4063 |
| 1.3085 | 13.0 | 18291 | 1.2760 | 1.0 | 352.7639 | 11.5290 |
| 1.2906 | 14.0 | 19698 | 1.2660 | 1.0 | 369.5020 | 11.6876 |
| 1.2321 | 15.0 | 21105 | 1.2499 | 1.0 | 341.5451 | 11.7774 |
| 1.217 | 16.0 | 22512 | 1.2437 | 1.0 | 348.1670 | 12.0046 |
| 1.1782 | 17.0 | 23919 | 1.2357 | 1.0 | 354.7813 | 11.9847 |
| 1.1637 | 18.0 | 25326 | 1.2354 | 1.0 | 324.6648 | 11.9945 |
| 1.1299 | 19.0 | 26733 | 1.2271 | 1.0 | 349.8055 | 12.0771 |
| 1.1066 | 20.0 | 28140 | 1.2211 | 1.0 | 341.9852 | 12.2001 |
| 1.0848 | 21.0 | 29547 | 1.2183 | 1.0 | 369.7122 | 12.1532 |
| 1.0629 | 22.0 | 30954 | 1.2167 | 1.0 | 352.1940 | 12.3542 |
| 1.0584 | 23.0 | 32361 | 1.2174 | 1.0 | 334.0109 | 12.4420 |
| 1.0196 | 24.0 | 33768 | 1.2174 | 1.0 | 362.3988 | 12.3138 |
| 1.0138 | 25.0 | 35175 | 1.2188 | 1.0 | 415.4663 | 12.2772 |
| 0.9972 | 26.0 | 36582 | 1.2161 | 1.0 | 342.5789 | 12.3754 |
| 0.9592 | 27.0 | 37989 | 1.2182 | 1.0 | 361.4101 | 12.3343 |
| 0.9458 | 28.0 | 39396 | 1.2189 | 1.0 | 338.6802 | 12.5021 |
| 0.9623 | 29.0 | 40803 | 1.2195 | 1.0 | 400.4501 | 12.4644 |
| 0.9322 | 30.0 | 42210 | 1.2265 | 1.0 | 342.0698 | 12.4692 |
Framework versions
- Transformers 4.57.0
- Pytorch 2.8.0+cu128
- Datasets 4.2.0
- Tokenizers 0.22.1
- Downloads last month
- 1
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
Model tree for contemmcm/b420da0d4015d59dc273e790e147248c
Base model
google-t5/t5-base