lorenzoscottb/PreDA-t5-small
Browse files- README.md +26 -25
- tokenizer.json +2 -11
README.md
CHANGED
|
@@ -18,11 +18,11 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 18 |
|
| 19 |
This model is a fine-tuned version of [google-t5/t5-small](https://huggingface.co/google-t5/t5-small) on the None dataset.
|
| 20 |
It achieves the following results on the evaluation set:
|
| 21 |
-
- Loss: 1.
|
| 22 |
-
- Rouge1: 0.
|
| 23 |
-
- Rouge2: 0.
|
| 24 |
-
- Rougel: 0.
|
| 25 |
-
- Rougelsum: 0.
|
| 26 |
|
| 27 |
## Model description
|
| 28 |
|
|
@@ -47,6 +47,7 @@ The following hyperparameters were used during training:
|
|
| 47 |
- seed: 42
|
| 48 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 49 |
- lr_scheduler_type: linear
|
|
|
|
| 50 |
- num_epochs: 20
|
| 51 |
- mixed_precision_training: Native AMP
|
| 52 |
- label_smoothing_factor: 0.1
|
|
@@ -55,26 +56,26 @@ The following hyperparameters were used during training:
|
|
| 55 |
|
| 56 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
|
| 57 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|
|
| 58 |
-
| 2.
|
| 59 |
-
| 2.
|
| 60 |
-
| 1.
|
| 61 |
-
| 1.
|
| 62 |
-
| 1.
|
| 63 |
-
| 1.
|
| 64 |
-
| 1.8316 | 7.0 | 238 | 1.
|
| 65 |
-
| 1.
|
| 66 |
-
| 1.
|
| 67 |
-
| 1.
|
| 68 |
-
| 1.
|
| 69 |
-
| 1.
|
| 70 |
-
| 1.
|
| 71 |
-
| 1.
|
| 72 |
-
| 1.
|
| 73 |
-
| 1.
|
| 74 |
-
| 1.
|
| 75 |
-
| 1.
|
| 76 |
-
| 1.
|
| 77 |
-
| 1.
|
| 78 |
|
| 79 |
|
| 80 |
### Framework versions
|
|
|
|
| 18 |
|
| 19 |
This model is a fine-tuned version of [google-t5/t5-small](https://huggingface.co/google-t5/t5-small) on the None dataset.
|
| 20 |
It achieves the following results on the evaluation set:
|
| 21 |
+
- Loss: 1.6467
|
| 22 |
+
- Rouge1: 0.8305
|
| 23 |
+
- Rouge2: 0.7257
|
| 24 |
+
- Rougel: 0.8080
|
| 25 |
+
- Rougelsum: 0.8080
|
| 26 |
|
| 27 |
## Model description
|
| 28 |
|
|
|
|
| 47 |
- seed: 42
|
| 48 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 49 |
- lr_scheduler_type: linear
|
| 50 |
+
- lr_scheduler_warmup_steps: 10
|
| 51 |
- num_epochs: 20
|
| 52 |
- mixed_precision_training: Native AMP
|
| 53 |
- label_smoothing_factor: 0.1
|
|
|
|
| 56 |
|
| 57 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum |
|
| 58 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|
|
| 59 |
+
| 2.5334 | 1.0 | 34 | 2.2047 | 0.3655 | 0.1949 | 0.3570 | 0.3570 |
|
| 60 |
+
| 2.0989 | 2.0 | 68 | 2.0026 | 0.5321 | 0.3606 | 0.5169 | 0.5168 |
|
| 61 |
+
| 1.9755 | 3.0 | 102 | 1.9020 | 0.5873 | 0.4139 | 0.5639 | 0.5646 |
|
| 62 |
+
| 1.9445 | 4.0 | 136 | 1.8645 | 0.5968 | 0.4271 | 0.5800 | 0.5805 |
|
| 63 |
+
| 1.8995 | 5.0 | 170 | 1.8282 | 0.6438 | 0.4882 | 0.6216 | 0.6216 |
|
| 64 |
+
| 1.8616 | 6.0 | 204 | 1.7978 | 0.6675 | 0.5107 | 0.6473 | 0.6473 |
|
| 65 |
+
| 1.8316 | 7.0 | 238 | 1.7784 | 0.6890 | 0.5369 | 0.6638 | 0.6636 |
|
| 66 |
+
| 1.8049 | 8.0 | 272 | 1.7542 | 0.7191 | 0.5761 | 0.6934 | 0.6937 |
|
| 67 |
+
| 1.7977 | 9.0 | 306 | 1.7373 | 0.7322 | 0.5953 | 0.7049 | 0.7052 |
|
| 68 |
+
| 1.7642 | 10.0 | 340 | 1.7219 | 0.7545 | 0.6213 | 0.7248 | 0.7252 |
|
| 69 |
+
| 1.7562 | 11.0 | 374 | 1.7072 | 0.7664 | 0.6389 | 0.7418 | 0.7423 |
|
| 70 |
+
| 1.7437 | 12.0 | 408 | 1.6961 | 0.7777 | 0.6519 | 0.7494 | 0.7496 |
|
| 71 |
+
| 1.7271 | 13.0 | 442 | 1.6838 | 0.7893 | 0.6715 | 0.7636 | 0.7638 |
|
| 72 |
+
| 1.7238 | 14.0 | 476 | 1.6765 | 0.7946 | 0.6759 | 0.7701 | 0.7703 |
|
| 73 |
+
| 1.7151 | 15.0 | 510 | 1.6706 | 0.8065 | 0.6918 | 0.7830 | 0.7833 |
|
| 74 |
+
| 1.6997 | 16.0 | 544 | 1.6605 | 0.8143 | 0.7006 | 0.7889 | 0.7892 |
|
| 75 |
+
| 1.6937 | 17.0 | 578 | 1.6552 | 0.8202 | 0.7100 | 0.7965 | 0.7968 |
|
| 76 |
+
| 1.6919 | 18.0 | 612 | 1.6505 | 0.8238 | 0.7176 | 0.8019 | 0.8019 |
|
| 77 |
+
| 1.6826 | 19.0 | 646 | 1.6493 | 0.8262 | 0.7210 | 0.8039 | 0.8040 |
|
| 78 |
+
| 1.6811 | 20.0 | 680 | 1.6467 | 0.8305 | 0.7257 | 0.8080 | 0.8080 |
|
| 79 |
|
| 80 |
|
| 81 |
### Framework versions
|
tokenizer.json
CHANGED
|
@@ -2,20 +2,11 @@
|
|
| 2 |
"version": "1.0",
|
| 3 |
"truncation": {
|
| 4 |
"direction": "Right",
|
| 5 |
-
"max_length":
|
| 6 |
"strategy": "LongestFirst",
|
| 7 |
"stride": 0
|
| 8 |
},
|
| 9 |
-
"padding":
|
| 10 |
-
"strategy": {
|
| 11 |
-
"Fixed": 128
|
| 12 |
-
},
|
| 13 |
-
"direction": "Right",
|
| 14 |
-
"pad_to_multiple_of": null,
|
| 15 |
-
"pad_id": 0,
|
| 16 |
-
"pad_type_id": 0,
|
| 17 |
-
"pad_token": "<pad>"
|
| 18 |
-
},
|
| 19 |
"added_tokens": [
|
| 20 |
{
|
| 21 |
"id": 0,
|
|
|
|
| 2 |
"version": "1.0",
|
| 3 |
"truncation": {
|
| 4 |
"direction": "Right",
|
| 5 |
+
"max_length": 512,
|
| 6 |
"strategy": "LongestFirst",
|
| 7 |
"stride": 0
|
| 8 |
},
|
| 9 |
+
"padding": null,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 10 |
"added_tokens": [
|
| 11 |
{
|
| 12 |
"id": 0,
|