End of training
Browse files- README.md +25 -25
- model.safetensors +1 -1
- runs/Jul14_11-23-08_tardis/events.out.tfevents.1752484990.tardis.133679.0 +3 -0
- tokenizer.json +16 -2
- training_args.bin +1 -1
README.md
CHANGED
|
@@ -22,21 +22,21 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 22 |
|
| 23 |
This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
|
| 24 |
It achieves the following results on the evaluation set:
|
| 25 |
-
- Loss:
|
| 26 |
-
- Rouge1: 0.
|
| 27 |
-
- Rouge2: 0.
|
| 28 |
-
- Rougel: 0.
|
| 29 |
-
- Rougelsum: 0.
|
| 30 |
- Gen Len: 20.0
|
| 31 |
-
- Bleu: 0.
|
| 32 |
-
- Precisions: 0.
|
| 33 |
-
- Brevity Penalty: 0.
|
| 34 |
-
- Length Ratio: 0.
|
| 35 |
-
- Translation Length:
|
| 36 |
- Reference Length: 1221.0
|
| 37 |
-
- Precision: 0.
|
| 38 |
-
- Recall: 0.
|
| 39 |
-
- F1: 0.
|
| 40 |
- Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
|
| 41 |
|
| 42 |
## Model description
|
|
@@ -70,18 +70,18 @@ The following hyperparameters were used during training:
|
|
| 70 |
|
| 71 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
|
| 72 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
|
| 73 |
-
| No log | 1.0 | 7 |
|
| 74 |
-
| No log | 2.0 | 14 | 23.
|
| 75 |
-
| No log | 3.0 | 21 |
|
| 76 |
-
| No log | 4.0 | 28 | 19.
|
| 77 |
-
| No log | 5.0 | 35 |
|
| 78 |
-
| No log | 6.0 | 42 | 16.
|
| 79 |
-
| No log | 7.0 | 49 |
|
| 80 |
-
| No log | 8.0 | 56 |
|
| 81 |
-
| No log | 9.0 | 63 |
|
| 82 |
-
| No log | 10.0 | 70 |
|
| 83 |
-
| No log | 11.0 | 77 |
|
| 84 |
-
| No log | 12.0 | 84 |
|
| 85 |
|
| 86 |
|
| 87 |
### Framework versions
|
|
|
|
| 22 |
|
| 23 |
This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
|
| 24 |
It achieves the following results on the evaluation set:
|
| 25 |
+
- Loss: 5.9110
|
| 26 |
+
- Rouge1: 0.1621
|
| 27 |
+
- Rouge2: 0.032
|
| 28 |
+
- Rougel: 0.1254
|
| 29 |
+
- Rougelsum: 0.126
|
| 30 |
- Gen Len: 20.0
|
| 31 |
+
- Bleu: 0.0
|
| 32 |
+
- Precisions: 0.051
|
| 33 |
+
- Brevity Penalty: 0.5104
|
| 34 |
+
- Length Ratio: 0.5979
|
| 35 |
+
- Translation Length: 730.0
|
| 36 |
- Reference Length: 1221.0
|
| 37 |
+
- Precision: 0.84
|
| 38 |
+
- Recall: 0.8425
|
| 39 |
+
- F1: 0.8411
|
| 40 |
- Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
|
| 41 |
|
| 42 |
## Model description
|
|
|
|
| 70 |
|
| 71 |
| Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
|
| 72 |
|:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
|
| 73 |
+
| No log | 1.0 | 7 | 25.6156 | 0.2003 | 0.0566 | 0.1655 | 0.1653 | 20.0 | 0.0197 | 0.072 | 0.5371 | 0.6167 | 753.0 | 1221.0 | 0.8574 | 0.8505 | 0.8538 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 74 |
+
| No log | 2.0 | 14 | 23.0400 | 0.1953 | 0.0521 | 0.1616 | 0.1611 | 20.0 | 0.0186 | 0.0701 | 0.5302 | 0.6118 | 747.0 | 1221.0 | 0.8566 | 0.8497 | 0.8531 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 75 |
+
| No log | 3.0 | 21 | 20.9401 | 0.1905 | 0.0494 | 0.157 | 0.1574 | 20.0 | 0.0145 | 0.0629 | 0.5394 | 0.6183 | 755.0 | 1221.0 | 0.8581 | 0.8507 | 0.8543 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 76 |
+
| No log | 4.0 | 28 | 19.2980 | 0.1983 | 0.0521 | 0.1626 | 0.1635 | 20.0 | 0.0157 | 0.0682 | 0.5337 | 0.6143 | 750.0 | 1221.0 | 0.8595 | 0.8511 | 0.8552 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 77 |
+
| No log | 5.0 | 35 | 17.7768 | 0.2064 | 0.0605 | 0.1704 | 0.1719 | 20.0 | 0.0217 | 0.0758 | 0.5302 | 0.6118 | 747.0 | 1221.0 | 0.8612 | 0.8524 | 0.8567 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 78 |
+
| No log | 6.0 | 42 | 16.2705 | 0.2071 | 0.0636 | 0.1691 | 0.1699 | 20.0 | 0.0273 | 0.0818 | 0.5313 | 0.6126 | 748.0 | 1221.0 | 0.8599 | 0.851 | 0.8553 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 79 |
+
| No log | 7.0 | 49 | 14.6608 | 0.2018 | 0.0607 | 0.1676 | 0.1689 | 20.0 | 0.0263 | 0.0797 | 0.5244 | 0.6077 | 742.0 | 1221.0 | 0.8575 | 0.85 | 0.8536 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 80 |
+
| No log | 8.0 | 56 | 12.7872 | 0.1914 | 0.0533 | 0.1564 | 0.1566 | 20.0 | 0.0263 | 0.0771 | 0.5267 | 0.6093 | 744.0 | 1221.0 | 0.8528 | 0.8469 | 0.8498 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 81 |
+
| No log | 9.0 | 63 | 10.6116 | 0.2038 | 0.0562 | 0.1631 | 0.1637 | 20.0 | 0.0295 | 0.0836 | 0.5232 | 0.6069 | 741.0 | 1221.0 | 0.8542 | 0.8486 | 0.8513 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 82 |
+
| No log | 10.0 | 70 | 8.3062 | 0.1963 | 0.0497 | 0.1559 | 0.1558 | 20.0 | 0.0244 | 0.0747 | 0.5244 | 0.6077 | 742.0 | 1221.0 | 0.8504 | 0.8472 | 0.8487 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 83 |
+
| No log | 11.0 | 77 | 6.5339 | 0.1794 | 0.0401 | 0.1404 | 0.1411 | 20.0 | 0.0182 | 0.0649 | 0.5186 | 0.6036 | 737.0 | 1221.0 | 0.8448 | 0.8443 | 0.8444 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 84 |
+
| No log | 12.0 | 84 | 5.9110 | 0.1621 | 0.032 | 0.1254 | 0.126 | 20.0 | 0.0 | 0.051 | 0.5104 | 0.5979 | 730.0 | 1221.0 | 0.84 | 0.8425 | 0.8411 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
|
| 85 |
|
| 86 |
|
| 87 |
### Framework versions
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1187780840
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:748f5f9c089be571577d07d78ef4cf5f65da2e1bd0a088cd7116d576b2486cc3
|
| 3 |
size 1187780840
|
runs/Jul14_11-23-08_tardis/events.out.tfevents.1752484990.tardis.133679.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:676384abafcdcb3d65d82654256ce160c691bb8cdc82553c31a4a2525466abf3
|
| 3 |
+
size 19283
|
tokenizer.json
CHANGED
|
@@ -1,7 +1,21 @@
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
-
"truncation":
|
| 4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
"added_tokens": [
|
| 6 |
{
|
| 7 |
"id": 0,
|
|
|
|
| 1 |
{
|
| 2 |
"version": "1.0",
|
| 3 |
+
"truncation": {
|
| 4 |
+
"direction": "Right",
|
| 5 |
+
"max_length": 64,
|
| 6 |
+
"strategy": "LongestFirst",
|
| 7 |
+
"stride": 0
|
| 8 |
+
},
|
| 9 |
+
"padding": {
|
| 10 |
+
"strategy": {
|
| 11 |
+
"Fixed": 64
|
| 12 |
+
},
|
| 13 |
+
"direction": "Right",
|
| 14 |
+
"pad_to_multiple_of": null,
|
| 15 |
+
"pad_id": 0,
|
| 16 |
+
"pad_type_id": 0,
|
| 17 |
+
"pad_token": "<pad>"
|
| 18 |
+
},
|
| 19 |
"added_tokens": [
|
| 20 |
{
|
| 21 |
"id": 0,
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 5905
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:52b7fc4ba8d20c804d0f6fcd2fac8ff4f5b6178b46348f33459b12af5aec5ef9
|
| 3 |
size 5905
|