floflodebilbao commited on
Commit
94f2194
·
verified ·
1 Parent(s): a6194d3

End of training

Browse files
README.md CHANGED
@@ -22,21 +22,21 @@ should probably proofread and complete it, then remove this comment. -->
22
 
23
  This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
24
  It achieves the following results on the evaluation set:
25
- - Loss: 9.3636
26
- - Rouge1: 0.1964
27
- - Rouge2: 0.0473
28
- - Rougel: 0.1532
29
- - Rougelsum: 0.1538
30
  - Gen Len: 20.0
31
- - Bleu: 0.0207
32
- - Precisions: 0.0708
33
- - Brevity Penalty: 0.529
34
- - Length Ratio: 0.611
35
- - Translation Length: 746.0
36
  - Reference Length: 1221.0
37
- - Precision: 0.8524
38
- - Recall: 0.8477
39
- - F1: 0.85
40
  - Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
41
 
42
  ## Model description
@@ -70,18 +70,18 @@ The following hyperparameters were used during training:
70
 
71
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
72
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
73
- | No log | 1.0 | 7 | 26.0895 | 0.2073 | 0.0553 | 0.1695 | 0.1697 | 20.0 | 0.0198 | 0.0737 | 0.5394 | 0.6183 | 755.0 | 1221.0 | 0.8581 | 0.8517 | 0.8548 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
74
- | No log | 2.0 | 14 | 23.5786 | 0.1852 | 0.0458 | 0.1528 | 0.1531 | 20.0 | 0.0162 | 0.0646 | 0.5325 | 0.6134 | 749.0 | 1221.0 | 0.8547 | 0.8489 | 0.8517 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
75
- | No log | 3.0 | 21 | 21.4863 | 0.1833 | 0.0465 | 0.1532 | 0.1525 | 20.0 | 0.0157 | 0.0623 | 0.5267 | 0.6093 | 744.0 | 1221.0 | 0.8552 | 0.8486 | 0.8518 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
76
- | No log | 4.0 | 28 | 19.7741 | 0.1922 | 0.0521 | 0.1598 | 0.1595 | 20.0 | 0.0153 | 0.0655 | 0.5337 | 0.6143 | 750.0 | 1221.0 | 0.8575 | 0.8506 | 0.8539 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
77
- | No log | 5.0 | 35 | 18.2754 | 0.1896 | 0.0495 | 0.1581 | 0.1577 | 20.0 | 0.0153 | 0.0667 | 0.5267 | 0.6093 | 744.0 | 1221.0 | 0.8565 | 0.8497 | 0.853 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
78
- | No log | 6.0 | 42 | 16.9095 | 0.1914 | 0.0495 | 0.1582 | 0.1578 | 20.0 | 0.0153 | 0.0665 | 0.529 | 0.611 | 746.0 | 1221.0 | 0.857 | 0.8498 | 0.8533 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
79
- | No log | 7.0 | 49 | 15.5684 | 0.1928 | 0.0531 | 0.1586 | 0.158 | 20.0 | 0.019 | 0.0702 | 0.5255 | 0.6085 | 743.0 | 1221.0 | 0.8564 | 0.8495 | 0.8528 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
80
- | No log | 8.0 | 56 | 14.1588 | 0.1885 | 0.0497 | 0.1545 | 0.1544 | 20.0 | 0.0197 | 0.0688 | 0.5279 | 0.6102 | 745.0 | 1221.0 | 0.8542 | 0.8482 | 0.8511 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
81
- | No log | 9.0 | 63 | 12.6292 | 0.1768 | 0.0366 | 0.1401 | 0.1402 | 20.0 | 0.0113 | 0.0586 | 0.5209 | 0.6052 | 739.0 | 1221.0 | 0.8507 | 0.8458 | 0.8482 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
82
- | No log | 10.0 | 70 | 11.0702 | 0.1915 | 0.0442 | 0.1507 | 0.1504 | 20.0 | 0.0172 | 0.0665 | 0.5232 | 0.6069 | 741.0 | 1221.0 | 0.8532 | 0.8472 | 0.8501 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
83
- | No log | 11.0 | 77 | 9.8692 | 0.1943 | 0.0473 | 0.1519 | 0.1521 | 20.0 | 0.0207 | 0.0704 | 0.5337 | 0.6143 | 750.0 | 1221.0 | 0.853 | 0.848 | 0.8504 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
84
- | No log | 12.0 | 84 | 9.3636 | 0.1964 | 0.0473 | 0.1532 | 0.1538 | 20.0 | 0.0207 | 0.0708 | 0.529 | 0.611 | 746.0 | 1221.0 | 0.8524 | 0.8477 | 0.85 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
85
 
86
 
87
  ### Framework versions
 
22
 
23
  This model is a fine-tuned version of [google/long-t5-tglobal-base](https://huggingface.co/google/long-t5-tglobal-base) on an unknown dataset.
24
  It achieves the following results on the evaluation set:
25
+ - Loss: 5.9110
26
+ - Rouge1: 0.1621
27
+ - Rouge2: 0.032
28
+ - Rougel: 0.1254
29
+ - Rougelsum: 0.126
30
  - Gen Len: 20.0
31
+ - Bleu: 0.0
32
+ - Precisions: 0.051
33
+ - Brevity Penalty: 0.5104
34
+ - Length Ratio: 0.5979
35
+ - Translation Length: 730.0
36
  - Reference Length: 1221.0
37
+ - Precision: 0.84
38
+ - Recall: 0.8425
39
+ - F1: 0.8411
40
  - Hashcode: roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1)
41
 
42
  ## Model description
 
70
 
71
  | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Precision | Recall | F1 | Hashcode |
72
  |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:|:---------:|:-------:|:------:|:----------:|:---------------:|:------------:|:------------------:|:----------------:|:---------:|:------:|:------:|:---------------------------------------------------------:|
73
+ | No log | 1.0 | 7 | 25.6156 | 0.2003 | 0.0566 | 0.1655 | 0.1653 | 20.0 | 0.0197 | 0.072 | 0.5371 | 0.6167 | 753.0 | 1221.0 | 0.8574 | 0.8505 | 0.8538 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
74
+ | No log | 2.0 | 14 | 23.0400 | 0.1953 | 0.0521 | 0.1616 | 0.1611 | 20.0 | 0.0186 | 0.0701 | 0.5302 | 0.6118 | 747.0 | 1221.0 | 0.8566 | 0.8497 | 0.8531 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
75
+ | No log | 3.0 | 21 | 20.9401 | 0.1905 | 0.0494 | 0.157 | 0.1574 | 20.0 | 0.0145 | 0.0629 | 0.5394 | 0.6183 | 755.0 | 1221.0 | 0.8581 | 0.8507 | 0.8543 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
76
+ | No log | 4.0 | 28 | 19.2980 | 0.1983 | 0.0521 | 0.1626 | 0.1635 | 20.0 | 0.0157 | 0.0682 | 0.5337 | 0.6143 | 750.0 | 1221.0 | 0.8595 | 0.8511 | 0.8552 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
77
+ | No log | 5.0 | 35 | 17.7768 | 0.2064 | 0.0605 | 0.1704 | 0.1719 | 20.0 | 0.0217 | 0.0758 | 0.5302 | 0.6118 | 747.0 | 1221.0 | 0.8612 | 0.8524 | 0.8567 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
78
+ | No log | 6.0 | 42 | 16.2705 | 0.2071 | 0.0636 | 0.1691 | 0.1699 | 20.0 | 0.0273 | 0.0818 | 0.5313 | 0.6126 | 748.0 | 1221.0 | 0.8599 | 0.851 | 0.8553 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
79
+ | No log | 7.0 | 49 | 14.6608 | 0.2018 | 0.0607 | 0.1676 | 0.1689 | 20.0 | 0.0263 | 0.0797 | 0.5244 | 0.6077 | 742.0 | 1221.0 | 0.8575 | 0.85 | 0.8536 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
80
+ | No log | 8.0 | 56 | 12.7872 | 0.1914 | 0.0533 | 0.1564 | 0.1566 | 20.0 | 0.0263 | 0.0771 | 0.5267 | 0.6093 | 744.0 | 1221.0 | 0.8528 | 0.8469 | 0.8498 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
81
+ | No log | 9.0 | 63 | 10.6116 | 0.2038 | 0.0562 | 0.1631 | 0.1637 | 20.0 | 0.0295 | 0.0836 | 0.5232 | 0.6069 | 741.0 | 1221.0 | 0.8542 | 0.8486 | 0.8513 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
82
+ | No log | 10.0 | 70 | 8.3062 | 0.1963 | 0.0497 | 0.1559 | 0.1558 | 20.0 | 0.0244 | 0.0747 | 0.5244 | 0.6077 | 742.0 | 1221.0 | 0.8504 | 0.8472 | 0.8487 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
83
+ | No log | 11.0 | 77 | 6.5339 | 0.1794 | 0.0401 | 0.1404 | 0.1411 | 20.0 | 0.0182 | 0.0649 | 0.5186 | 0.6036 | 737.0 | 1221.0 | 0.8448 | 0.8443 | 0.8444 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
84
+ | No log | 12.0 | 84 | 5.9110 | 0.1621 | 0.032 | 0.1254 | 0.126 | 20.0 | 0.0 | 0.051 | 0.5104 | 0.5979 | 730.0 | 1221.0 | 0.84 | 0.8425 | 0.8411 | roberta-large_L17_no-idf_version=0.3.12(hug_trans=4.53.1) |
85
 
86
 
87
  ### Framework versions
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:895f8e3502d4e79d013360d6c31ef86aa8b7a6eaceb3e4a2c92fe5c9950b32b2
3
  size 1187780840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:748f5f9c089be571577d07d78ef4cf5f65da2e1bd0a088cd7116d576b2486cc3
3
  size 1187780840
runs/Jul14_11-23-08_tardis/events.out.tfevents.1752484990.tardis.133679.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:676384abafcdcb3d65d82654256ce160c691bb8cdc82553c31a4a2525466abf3
3
+ size 19283
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 64,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 64
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 0,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<pad>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac349db475e2e5d7ee45c8492b0d94086f85332b6aebe6ffc46b7433eca31484
3
  size 5905
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52b7fc4ba8d20c804d0f6fcd2fac8ff4f5b6178b46348f33459b12af5aec5ef9
3
  size 5905