CNR-ILC commited on
Commit
3c9a6a4
·
verified ·
1 Parent(s): 45b35d2

ILC-CNR/gs-greBERTa

Browse files
README.md CHANGED
@@ -1,21 +1,26 @@
1
  ---
2
  library_name: transformers
3
- base_model: Jacobo/aristoBERTo
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
- - name: model-checkpoints
8
  results: []
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
- # model-checkpoints
15
 
16
- This model is a fine-tuned version of [Jacobo/aristoBERTo](https://huggingface.co/Jacobo/aristoBERTo) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 2.0007
 
 
 
 
19
 
20
  ## Model description
21
 
@@ -36,7 +41,7 @@ More information needed
36
  The following hyperparameters were used during training:
37
  - learning_rate: 5e-05
38
  - train_batch_size: 16
39
- - eval_batch_size: 16
40
  - seed: 42
41
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
42
  - lr_scheduler_type: linear
@@ -45,23 +50,23 @@ The following hyperparameters were used during training:
45
 
46
  ### Training results
47
 
48
- | Training Loss | Epoch | Step | Validation Loss |
49
- |:-------------:|:-----:|:-----:|:---------------:|
50
- | 2.9726 | 1.0 | 5652 | 2.5978 |
51
- | 2.5639 | 2.0 | 11304 | 2.3816 |
52
- | 2.4014 | 3.0 | 16956 | 2.2645 |
53
- | 2.2988 | 4.0 | 22608 | 2.2166 |
54
- | 2.2089 | 5.0 | 28260 | 2.1206 |
55
- | 2.1371 | 6.0 | 33912 | 2.0921 |
56
- | 2.0808 | 7.0 | 39564 | 2.0355 |
57
- | 2.0361 | 8.0 | 45216 | 2.0193 |
58
- | 2.001 | 9.0 | 50868 | 1.9988 |
59
- | 1.977 | 10.0 | 56520 | 1.9922 |
60
 
61
 
62
  ### Framework versions
63
 
64
  - Transformers 4.51.3
65
  - Pytorch 2.7.0+cu126
66
- - Datasets 3.5.0
67
  - Tokenizers 0.21.1
 
1
  ---
2
  library_name: transformers
3
+ license: apache-2.0
4
+ base_model: bowphs/GreBerta
5
  tags:
6
  - generated_from_trainer
7
  model-index:
8
+ - name: gs-greBERTa
9
  results: []
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
+ # gs-greBERTa
16
 
17
+ This model is a fine-tuned version of [bowphs/GreBerta](https://huggingface.co/bowphs/GreBerta) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.5786
20
+ - Top1 Acc: 0.8589
21
+ - Top5 Acc: 0.9202
22
+ - Top10 Acc: 0.9448
23
+ - Top20 Acc: 0.9632
24
 
25
  ## Model description
26
 
 
41
  The following hyperparameters were used during training:
42
  - learning_rate: 5e-05
43
  - train_batch_size: 16
44
+ - eval_batch_size: 8
45
  - seed: 42
46
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: linear
 
50
 
51
  ### Training results
52
 
53
+ | Training Loss | Epoch | Step | Validation Loss | Top1 Acc | Top5 Acc | Top10 Acc | Top20 Acc |
54
+ |:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|:---------:|:---------:|
55
+ | 1.2561 | 1.0 | 5634 | 0.9441 | 0.7965 | 0.8837 | 0.9302 | 0.9593 |
56
+ | 0.969 | 2.0 | 11268 | 0.8007 | 0.8028 | 0.9296 | 0.9366 | 0.9648 |
57
+ | 0.8407 | 3.0 | 16902 | 0.7505 | 0.8092 | 0.9249 | 0.9480 | 0.9769 |
58
+ | 0.7791 | 4.0 | 22536 | 0.6900 | 0.825 | 0.9313 | 0.95 | 0.975 |
59
+ | 0.7264 | 5.0 | 28170 | 0.6541 | 0.8824 | 0.9471 | 0.9706 | 0.9765 |
60
+ | 0.6872 | 6.0 | 33804 | 0.6343 | 0.8344 | 0.9264 | 0.9571 | 0.9877 |
61
+ | 0.6553 | 7.0 | 39438 | 0.6069 | 0.8705 | 0.9568 | 0.9712 | 0.9784 |
62
+ | 0.6479 | 8.0 | 45072 | 0.5924 | 0.8905 | 0.9708 | 0.9854 | 0.9927 |
63
+ | 0.6181 | 9.0 | 50706 | 0.5827 | 0.8834 | 0.9571 | 0.9693 | 0.9816 |
64
+ | 0.6051 | 10.0 | 56340 | 0.5851 | 0.8922 | 0.9701 | 1.0 | 1.0 |
65
 
66
 
67
  ### Framework versions
68
 
69
  - Transformers 4.51.3
70
  - Pytorch 2.7.0+cu126
71
+ - Datasets 3.5.1
72
  - Tokenizers 0.21.1
all_results.json CHANGED
@@ -1,13 +1,17 @@
1
  {
2
  "epoch": 10.0,
3
- "eval_loss": 2.0006515979766846,
4
- "eval_runtime": 11.5932,
5
- "eval_samples_per_second": 839.288,
6
- "eval_steps_per_second": 52.531,
7
- "step": 56520,
8
- "total_flos": 3.105015735189504e+16,
9
- "train_loss": 2.067556116325038,
10
- "train_runtime": 7586.095,
11
- "train_samples_per_second": 124.388,
12
- "train_steps_per_second": 7.773
 
 
 
 
13
  }
 
1
  {
2
  "epoch": 10.0,
3
+ "eval_loss": 0.5786494612693787,
4
+ "eval_runtime": 736.6563,
5
+ "eval_samples_per_second": 3.851,
6
+ "eval_steps_per_second": 0.482,
7
+ "eval_top10_acc": 0.9447852760736196,
8
+ "eval_top1_acc": 0.8588957055214724,
9
+ "eval_top20_acc": 0.9631901840490797,
10
+ "eval_top5_acc": 0.9202453987730062,
11
+ "step": 56340,
12
+ "total_flos": 5.933070595915776e+16,
13
+ "train_loss": 0.7784920386062855,
14
+ "train_runtime": 14293.974,
15
+ "train_samples_per_second": 63.066,
16
+ "train_steps_per_second": 3.942
17
  }
config.json CHANGED
@@ -1,25 +1,26 @@
1
  {
2
  "architectures": [
3
- "BertForMaskedLM"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
 
6
  "classifier_dropout": null,
 
7
  "hidden_act": "gelu",
8
  "hidden_dropout_prob": 0.1,
9
  "hidden_size": 768,
10
  "initializer_range": 0.02,
11
  "intermediate_size": 3072,
12
  "layer_norm_eps": 1e-12,
13
- "max_position_embeddings": 512,
14
- "model_type": "bert",
15
  "num_attention_heads": 12,
16
  "num_hidden_layers": 12,
17
- "output_past": true,
18
- "pad_token_id": 0,
19
  "position_embedding_type": "absolute",
20
  "torch_dtype": "float32",
21
  "transformers_version": "4.51.3",
22
- "type_vocab_size": 2,
23
  "use_cache": true,
24
- "vocab_size": 35000
25
  }
 
1
  {
2
  "architectures": [
3
+ "RobertaForMaskedLM"
4
  ],
5
  "attention_probs_dropout_prob": 0.1,
6
+ "bos_token_id": 0,
7
  "classifier_dropout": null,
8
+ "eos_token_id": 2,
9
  "hidden_act": "gelu",
10
  "hidden_dropout_prob": 0.1,
11
  "hidden_size": 768,
12
  "initializer_range": 0.02,
13
  "intermediate_size": 3072,
14
  "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 514,
16
+ "model_type": "roberta",
17
  "num_attention_heads": 12,
18
  "num_hidden_layers": 12,
19
+ "pad_token_id": 1,
 
20
  "position_embedding_type": "absolute",
21
  "torch_dtype": "float32",
22
  "transformers_version": "4.51.3",
23
+ "type_vocab_size": 1,
24
  "use_cache": true,
25
+ "vocab_size": 52000
26
  }
eval_results.json CHANGED
@@ -1,7 +1,11 @@
1
  {
2
  "epoch": 10.0,
3
- "eval_loss": 2.0006515979766846,
4
- "eval_runtime": 11.5932,
5
- "eval_samples_per_second": 839.288,
6
- "eval_steps_per_second": 52.531
 
 
 
 
7
  }
 
1
  {
2
  "epoch": 10.0,
3
+ "eval_loss": 0.5786494612693787,
4
+ "eval_runtime": 736.6563,
5
+ "eval_samples_per_second": 3.851,
6
+ "eval_steps_per_second": 0.482,
7
+ "eval_top10_acc": 0.9447852760736196,
8
+ "eval_top1_acc": 0.8588957055214724,
9
+ "eval_top20_acc": 0.9631901840490797,
10
+ "eval_top5_acc": 0.9202453987730062
11
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2ac48d52adb26e2fddddb2ca2062b638e638322e6dcaa395cb7f33fd5e6f5338
3
- size 451855232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbc282e85cce75ef4bb3e2e8e5e0fe234ef492beebc040cd69cd3f3746b1457c
3
+ size 504150808
special_tokens_map.json CHANGED
@@ -1,7 +1,51 @@
1
  {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
 
1
  {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "<s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "<mask>",
25
+ "lstrip": true,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "<pad>",
32
+ "lstrip": false,
33
+ "normalized": true,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "</s>",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "<unk>",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
  }
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,58 +1,58 @@
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
- "content": "[PAD]",
5
  "lstrip": false,
6
- "normalized": false,
7
  "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
11
- "100": {
12
- "content": "[UNK]",
13
  "lstrip": false,
14
- "normalized": false,
15
  "rstrip": false,
16
  "single_word": false,
17
  "special": true
18
  },
19
- "101": {
20
- "content": "[CLS]",
21
  "lstrip": false,
22
- "normalized": false,
23
  "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
27
- "102": {
28
- "content": "[SEP]",
29
  "lstrip": false,
30
- "normalized": false,
31
  "rstrip": false,
32
  "single_word": false,
33
  "special": true
34
  },
35
- "103": {
36
- "content": "[MASK]",
37
- "lstrip": false,
38
- "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
42
  }
43
  },
44
- "clean_up_tokenization_spaces": true,
45
- "cls_token": "[CLS]",
46
- "do_basic_tokenize": true,
47
- "do_lower_case": true,
 
48
  "extra_special_tokens": {},
49
- "mask_token": "[MASK]",
50
  "model_max_length": 512,
51
- "never_split": null,
52
- "pad_token": "[PAD]",
53
- "sep_token": "[SEP]",
54
- "strip_accents": null,
55
- "tokenize_chinese_chars": true,
56
- "tokenizer_class": "BertTokenizer",
57
- "unk_token": "[UNK]"
58
  }
 
1
  {
2
+ "add_prefix_space": false,
3
  "added_tokens_decoder": {
4
  "0": {
5
+ "content": "<s>",
6
  "lstrip": false,
7
+ "normalized": true,
8
  "rstrip": false,
9
  "single_word": false,
10
  "special": true
11
  },
12
+ "1": {
13
+ "content": "<pad>",
14
  "lstrip": false,
15
+ "normalized": true,
16
  "rstrip": false,
17
  "single_word": false,
18
  "special": true
19
  },
20
+ "2": {
21
+ "content": "</s>",
22
  "lstrip": false,
23
+ "normalized": true,
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
  },
28
+ "3": {
29
+ "content": "<unk>",
30
  "lstrip": false,
31
+ "normalized": true,
32
  "rstrip": false,
33
  "single_word": false,
34
  "special": true
35
  },
36
+ "4": {
37
+ "content": "<mask>",
38
+ "lstrip": true,
39
+ "normalized": true,
40
  "rstrip": false,
41
  "single_word": false,
42
  "special": true
43
  }
44
  },
45
+ "bos_token": "<s>",
46
+ "clean_up_tokenization_spaces": false,
47
+ "cls_token": "<s>",
48
+ "eos_token": "</s>",
49
+ "errors": "replace",
50
  "extra_special_tokens": {},
51
+ "mask_token": "<mask>",
52
  "model_max_length": 512,
53
+ "pad_token": "<pad>",
54
+ "sep_token": "</s>",
55
+ "tokenizer_class": "RobertaTokenizer",
56
+ "trim_offsets": true,
57
+ "unk_token": "<unk>"
 
 
58
  }
train_results.json CHANGED
@@ -1,8 +1,9 @@
1
  {
2
  "epoch": 10.0,
3
- "eval_loss": 1.980637788772583,
4
- "eval_runtime": 11.5702,
5
- "eval_samples_per_second": 840.952,
6
- "eval_steps_per_second": 52.635,
7
- "step": 56520
 
8
  }
 
1
  {
2
  "epoch": 10.0,
3
+ "step": 56340,
4
+ "total_flos": 5.933070595915776e+16,
5
+ "train_loss": 0.7784920386062855,
6
+ "train_runtime": 14293.974,
7
+ "train_samples_per_second": 63.066,
8
+ "train_steps_per_second": 3.942
9
  }
trainer_state.json CHANGED
@@ -4,197 +4,225 @@
4
  "best_model_checkpoint": null,
5
  "epoch": 10.0,
6
  "eval_steps": 500,
7
- "global_step": 56520,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
- "grad_norm": 12.995927810668945,
15
- "learning_rate": 4.50070771408351e-05,
16
- "loss": 2.9726,
17
- "step": 5652
18
  },
19
  {
20
  "epoch": 1.0,
21
- "eval_loss": 2.5977725982666016,
22
- "eval_runtime": 19.8579,
23
- "eval_samples_per_second": 489.981,
24
- "eval_steps_per_second": 30.668,
25
- "step": 5652
 
 
 
 
26
  },
27
  {
28
  "epoch": 2.0,
29
- "grad_norm": 10.8749361038208,
30
- "learning_rate": 4.000796178343949e-05,
31
- "loss": 2.5639,
32
- "step": 11304
33
  },
34
  {
35
  "epoch": 2.0,
36
- "eval_loss": 2.381584405899048,
37
- "eval_runtime": 19.7379,
38
- "eval_samples_per_second": 492.959,
39
- "eval_steps_per_second": 30.854,
40
- "step": 11304
 
 
 
 
41
  },
42
  {
43
  "epoch": 3.0,
44
- "grad_norm": 11.009535789489746,
45
- "learning_rate": 3.500973106864827e-05,
46
- "loss": 2.4014,
47
- "step": 16956
48
  },
49
  {
50
  "epoch": 3.0,
51
- "eval_loss": 2.2645437717437744,
52
- "eval_runtime": 19.6521,
53
- "eval_samples_per_second": 495.112,
54
- "eval_steps_per_second": 30.989,
55
- "step": 16956
 
 
 
 
56
  },
57
  {
58
  "epoch": 4.0,
59
- "grad_norm": 11.495499610900879,
60
- "learning_rate": 3.0011500353857042e-05,
61
- "loss": 2.2988,
62
- "step": 22608
63
  },
64
  {
65
  "epoch": 4.0,
66
- "eval_loss": 2.2165863513946533,
67
- "eval_runtime": 19.8532,
68
- "eval_samples_per_second": 490.097,
69
- "eval_steps_per_second": 30.675,
70
- "step": 22608
 
 
 
 
71
  },
72
  {
73
  "epoch": 5.0,
74
- "grad_norm": 11.591689109802246,
75
- "learning_rate": 2.501238499646143e-05,
76
- "loss": 2.2089,
77
- "step": 28260
78
  },
79
  {
80
  "epoch": 5.0,
81
- "eval_loss": 2.120560884475708,
82
- "eval_runtime": 19.806,
83
- "eval_samples_per_second": 491.265,
84
- "eval_steps_per_second": 30.748,
85
- "step": 28260
 
 
 
 
86
  },
87
  {
88
  "epoch": 6.0,
89
- "grad_norm": 10.091324806213379,
90
- "learning_rate": 2.0015923566878984e-05,
91
- "loss": 2.1371,
92
- "step": 33912
93
  },
94
  {
95
  "epoch": 6.0,
96
- "eval_loss": 2.0921123027801514,
97
- "eval_runtime": 19.6342,
98
- "eval_samples_per_second": 495.564,
99
- "eval_steps_per_second": 31.017,
100
- "step": 33912
 
 
 
 
101
  },
102
  {
103
  "epoch": 7.0,
104
- "grad_norm": 9.590017318725586,
105
- "learning_rate": 1.5016808209483369e-05,
106
- "loss": 2.0808,
107
- "step": 39564
108
  },
109
  {
110
  "epoch": 7.0,
111
- "eval_loss": 2.035478115081787,
112
- "eval_runtime": 19.7938,
113
- "eval_samples_per_second": 491.568,
114
- "eval_steps_per_second": 30.767,
115
- "step": 39564
 
 
 
 
116
  },
117
  {
118
  "epoch": 8.0,
119
- "grad_norm": 11.4172945022583,
120
- "learning_rate": 1.0017692852087757e-05,
121
- "loss": 2.0361,
122
- "step": 45216
123
  },
124
  {
125
  "epoch": 8.0,
126
- "eval_loss": 2.019296169281006,
127
- "eval_runtime": 19.7587,
128
- "eval_samples_per_second": 492.442,
129
- "eval_steps_per_second": 30.822,
130
- "step": 45216
 
 
 
 
131
  },
132
  {
133
  "epoch": 9.0,
134
- "grad_norm": 13.36776351928711,
135
- "learning_rate": 5.019462137296532e-06,
136
- "loss": 2.001,
137
- "step": 50868
138
  },
139
  {
140
  "epoch": 9.0,
141
- "eval_loss": 1.9987555742263794,
142
- "eval_runtime": 19.4188,
143
- "eval_samples_per_second": 501.06,
144
- "eval_steps_per_second": 31.361,
145
- "step": 50868
 
 
 
 
146
  },
147
  {
148
  "epoch": 10.0,
149
- "grad_norm": 11.329434394836426,
150
- "learning_rate": 2.1231422505307857e-08,
151
- "loss": 1.977,
152
- "step": 56520
153
  },
154
  {
155
  "epoch": 10.0,
156
- "eval_loss": 1.992159366607666,
157
- "eval_runtime": 19.7519,
158
- "eval_samples_per_second": 492.612,
159
- "eval_steps_per_second": 30.833,
160
- "step": 56520
 
 
 
 
161
  },
162
  {
163
  "epoch": 10.0,
164
- "step": 56520,
165
- "total_flos": 2.975422706417664e+16,
166
- "train_loss": 2.2677627250309627,
167
- "train_runtime": 6971.4542,
168
- "train_samples_per_second": 129.738,
169
- "train_steps_per_second": 8.107
170
  },
171
  {
172
  "epoch": 10.0,
173
- "eval_loss": 1.9793492555618286,
174
- "eval_runtime": 19.7795,
175
- "eval_samples_per_second": 491.923,
176
- "eval_steps_per_second": 30.789,
177
- "step": 56520
178
- },
179
- {
180
- "epoch": 10.0,
181
- "eval_loss": 1.980637788772583,
182
- "eval_runtime": 11.5702,
183
- "eval_samples_per_second": 840.952,
184
- "eval_steps_per_second": 52.635,
185
- "step": 56520
186
- },
187
- {
188
- "epoch": 10.0,
189
- "eval_loss": 2.0006515979766846,
190
- "eval_runtime": 11.5932,
191
- "eval_samples_per_second": 839.288,
192
- "eval_steps_per_second": 52.531,
193
- "step": 56520
194
  }
195
  ],
196
  "logging_steps": 500,
197
- "max_steps": 56520,
198
  "num_input_tokens_seen": 0,
199
  "num_train_epochs": 10,
200
  "save_steps": 500,
@@ -210,7 +238,7 @@
210
  "attributes": {}
211
  }
212
  },
213
- "total_flos": 2.975422706417664e+16,
214
  "train_batch_size": 16,
215
  "trial_name": null,
216
  "trial_params": null
 
4
  "best_model_checkpoint": null,
5
  "epoch": 10.0,
6
  "eval_steps": 500,
7
+ "global_step": 56340,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 1.0,
14
+ "grad_norm": 5.6999335289001465,
15
+ "learning_rate": 4.5005324813631524e-05,
16
+ "loss": 1.2561,
17
+ "step": 5634
18
  },
19
  {
20
  "epoch": 1.0,
21
+ "eval_loss": 0.9440904259681702,
22
+ "eval_runtime": 727.0892,
23
+ "eval_samples_per_second": 3.902,
24
+ "eval_steps_per_second": 0.488,
25
+ "eval_top10_acc": 0.9302325581395349,
26
+ "eval_top1_acc": 0.7965116279069767,
27
+ "eval_top20_acc": 0.9593023255813954,
28
+ "eval_top5_acc": 0.8837209302325582,
29
+ "step": 5634
30
  },
31
  {
32
  "epoch": 2.0,
33
+ "grad_norm": 4.9532036781311035,
34
+ "learning_rate": 4.000887468938587e-05,
35
+ "loss": 0.969,
36
+ "step": 11268
37
  },
38
  {
39
  "epoch": 2.0,
40
+ "eval_loss": 0.8007155656814575,
41
+ "eval_runtime": 732.2211,
42
+ "eval_samples_per_second": 3.875,
43
+ "eval_steps_per_second": 0.485,
44
+ "eval_top10_acc": 0.9366197183098591,
45
+ "eval_top1_acc": 0.8028169014084507,
46
+ "eval_top20_acc": 0.9647887323943662,
47
+ "eval_top5_acc": 0.9295774647887324,
48
+ "step": 11268
49
  },
50
  {
51
  "epoch": 3.0,
52
+ "grad_norm": 4.767992973327637,
53
+ "learning_rate": 3.5011537096201635e-05,
54
+ "loss": 0.8407,
55
+ "step": 16902
56
  },
57
  {
58
  "epoch": 3.0,
59
+ "eval_loss": 0.750490665435791,
60
+ "eval_runtime": 746.6021,
61
+ "eval_samples_per_second": 3.8,
62
+ "eval_steps_per_second": 0.475,
63
+ "eval_top10_acc": 0.9479768786127167,
64
+ "eval_top1_acc": 0.8092485549132948,
65
+ "eval_top20_acc": 0.976878612716763,
66
+ "eval_top5_acc": 0.9248554913294798,
67
+ "step": 16902
68
  },
69
  {
70
  "epoch": 4.0,
71
+ "grad_norm": 4.196132659912109,
72
+ "learning_rate": 3.0011537096201635e-05,
73
+ "loss": 0.7791,
74
+ "step": 22536
75
  },
76
  {
77
  "epoch": 4.0,
78
+ "eval_loss": 0.6900169849395752,
79
+ "eval_runtime": 744.1442,
80
+ "eval_samples_per_second": 3.812,
81
+ "eval_steps_per_second": 0.477,
82
+ "eval_top10_acc": 0.95,
83
+ "eval_top1_acc": 0.825,
84
+ "eval_top20_acc": 0.975,
85
+ "eval_top5_acc": 0.93125,
86
+ "step": 22536
87
  },
88
  {
89
  "epoch": 5.0,
90
+ "grad_norm": 4.124821186065674,
91
+ "learning_rate": 2.5014199503017393e-05,
92
+ "loss": 0.7264,
93
+ "step": 28170
94
  },
95
  {
96
  "epoch": 5.0,
97
+ "eval_loss": 0.6541450619697571,
98
+ "eval_runtime": 732.9508,
99
+ "eval_samples_per_second": 3.871,
100
+ "eval_steps_per_second": 0.484,
101
+ "eval_top10_acc": 0.9705882352941176,
102
+ "eval_top1_acc": 0.8823529411764706,
103
+ "eval_top20_acc": 0.9764705882352941,
104
+ "eval_top5_acc": 0.9470588235294117,
105
+ "step": 28170
106
  },
107
  {
108
  "epoch": 6.0,
109
+ "grad_norm": 4.150660991668701,
110
+ "learning_rate": 2.001508697195598e-05,
111
+ "loss": 0.6872,
112
+ "step": 33804
113
  },
114
  {
115
  "epoch": 6.0,
116
+ "eval_loss": 0.6342827081680298,
117
+ "eval_runtime": 735.093,
118
+ "eval_samples_per_second": 3.859,
119
+ "eval_steps_per_second": 0.483,
120
+ "eval_top10_acc": 0.9570552147239264,
121
+ "eval_top1_acc": 0.8343558282208589,
122
+ "eval_top20_acc": 0.9877300613496932,
123
+ "eval_top5_acc": 0.9263803680981595,
124
+ "step": 33804
125
  },
126
  {
127
  "epoch": 7.0,
128
+ "grad_norm": 4.235877990722656,
129
+ "learning_rate": 1.501597444089457e-05,
130
+ "loss": 0.6553,
131
+ "step": 39438
132
  },
133
  {
134
  "epoch": 7.0,
135
+ "eval_loss": 0.6069093942642212,
136
+ "eval_runtime": 729.8603,
137
+ "eval_samples_per_second": 3.887,
138
+ "eval_steps_per_second": 0.486,
139
+ "eval_top10_acc": 0.9712230215827338,
140
+ "eval_top1_acc": 0.8705035971223022,
141
+ "eval_top20_acc": 0.9784172661870504,
142
+ "eval_top5_acc": 0.9568345323741008,
143
+ "step": 39438
144
  },
145
  {
146
  "epoch": 8.0,
147
+ "grad_norm": 4.277960777282715,
148
+ "learning_rate": 1.0021299254526093e-05,
149
+ "loss": 0.6479,
150
+ "step": 45072
151
  },
152
  {
153
  "epoch": 8.0,
154
+ "eval_loss": 0.5924458503723145,
155
+ "eval_runtime": 735.4415,
156
+ "eval_samples_per_second": 3.858,
157
+ "eval_steps_per_second": 0.483,
158
+ "eval_top10_acc": 0.9854014598540146,
159
+ "eval_top1_acc": 0.8905109489051095,
160
+ "eval_top20_acc": 0.9927007299270073,
161
+ "eval_top5_acc": 0.9708029197080292,
162
+ "step": 45072
163
  },
164
  {
165
  "epoch": 9.0,
166
+ "grad_norm": 3.431784152984619,
167
+ "learning_rate": 5.023074192403266e-06,
168
+ "loss": 0.6181,
169
+ "step": 50706
170
  },
171
  {
172
  "epoch": 9.0,
173
+ "eval_loss": 0.5826597213745117,
174
+ "eval_runtime": 740.5141,
175
+ "eval_samples_per_second": 3.831,
176
+ "eval_steps_per_second": 0.479,
177
+ "eval_top10_acc": 0.9693251533742331,
178
+ "eval_top1_acc": 0.8834355828220859,
179
+ "eval_top20_acc": 0.9815950920245399,
180
+ "eval_top5_acc": 0.9570552147239264,
181
+ "step": 50706
182
  },
183
  {
184
  "epoch": 10.0,
185
+ "grad_norm": 4.740328788757324,
186
+ "learning_rate": 2.5736599219027333e-08,
187
+ "loss": 0.6051,
188
+ "step": 56340
189
  },
190
  {
191
  "epoch": 10.0,
192
+ "eval_loss": 0.5851157307624817,
193
+ "eval_runtime": 739.7027,
194
+ "eval_samples_per_second": 3.835,
195
+ "eval_steps_per_second": 0.48,
196
+ "eval_top10_acc": 1.0,
197
+ "eval_top1_acc": 0.8922155688622755,
198
+ "eval_top20_acc": 1.0,
199
+ "eval_top5_acc": 0.9700598802395209,
200
+ "step": 56340
201
  },
202
  {
203
  "epoch": 10.0,
204
+ "step": 56340,
205
+ "total_flos": 5.933070595915776e+16,
206
+ "train_loss": 0.7784920386062855,
207
+ "train_runtime": 14293.974,
208
+ "train_samples_per_second": 63.066,
209
+ "train_steps_per_second": 3.942
210
  },
211
  {
212
  "epoch": 10.0,
213
+ "eval_loss": 0.5786494612693787,
214
+ "eval_runtime": 736.6563,
215
+ "eval_samples_per_second": 3.851,
216
+ "eval_steps_per_second": 0.482,
217
+ "eval_top10_acc": 0.9447852760736196,
218
+ "eval_top1_acc": 0.8588957055214724,
219
+ "eval_top20_acc": 0.9631901840490797,
220
+ "eval_top5_acc": 0.9202453987730062,
221
+ "step": 56340
 
 
 
 
 
 
 
 
 
 
 
 
222
  }
223
  ],
224
  "logging_steps": 500,
225
+ "max_steps": 56340,
226
  "num_input_tokens_seen": 0,
227
  "num_train_epochs": 10,
228
  "save_steps": 500,
 
238
  "attributes": {}
239
  }
240
  },
241
+ "total_flos": 5.933070595915776e+16,
242
  "train_batch_size": 16,
243
  "trial_name": null,
244
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b67c2a3013a187a18e50cba2537a82dc555169447cfcc62169c9520a9e77c19
3
- size 5713
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d86f31b8645edb5a8bc695e48ee26e49fb8c60618206647233a2cba8872a191c
3
+ size 5649