tzhao3 commited on
Commit
0fbe06d
·
1 Parent(s): b99cf91

Upload 11 files

Browse files
config.json CHANGED
@@ -19,7 +19,7 @@
19
  "position_embedding_type": "absolute",
20
  "problem_type": "single_label_classification",
21
  "torch_dtype": "float32",
22
- "transformers_version": "4.29.2",
23
  "type_vocab_size": 2,
24
  "use_cache": true,
25
  "vocab_size": 30522
 
19
  "position_embedding_type": "absolute",
20
  "problem_type": "single_label_classification",
21
  "torch_dtype": "float32",
22
+ "transformers_version": "4.26.1",
23
  "type_vocab_size": 2,
24
  "use_cache": true,
25
  "vocab_size": 30522
optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:084f7060a2d4e459e8f34a8c9d28522799810579ffff7777c4d91e6e62f8f071
3
  size 331073989
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9822d451c23c9d1a4219b5d731af3967b5694b24922818c968112db44e13a654
3
  size 331073989
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4b1f5e604bb2113932238477c13a8f1860a1c905693facf5d783e96f99a79aef
3
  size 165547637
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2767f946ca70f29f55cd0c13d8e46517e20c1b62ea5c70450eed00ea7f1d8d74
3
  size 165547637
rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8332db9dd41ad5c0859a96a265c824d1244e1ac28822bb3c87d560fccdc26cda
3
- size 14575
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:110f64100f09e85f8a8659dd6c765d0c05410e455733b95db82da65c9a985306
3
+ size 15597
scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54a776579ae12db457f9c3df76d9976cab904241a9c5dffa4cafa356e6aca712
3
  size 627
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:466f0b6297493eed05f2d1ce63809b3d054e71f229177cc77584e4462fd9dace
3
  size 627
tokenizer_config.json CHANGED
@@ -1,11 +1,12 @@
1
  {
2
- "clean_up_tokenization_spaces": true,
3
  "cls_token": "[CLS]",
4
  "do_lower_case": true,
5
  "mask_token": "[MASK]",
6
  "model_max_length": 512,
 
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
 
9
  "strip_accents": null,
10
  "tokenize_chinese_chars": true,
11
  "tokenizer_class": "BertTokenizer",
 
1
  {
 
2
  "cls_token": "[CLS]",
3
  "do_lower_case": true,
4
  "mask_token": "[MASK]",
5
  "model_max_length": 512,
6
+ "name_or_path": "bert-base-uncased",
7
  "pad_token": "[PAD]",
8
  "sep_token": "[SEP]",
9
+ "special_tokens_map_file": null,
10
  "strip_accents": null,
11
  "tokenize_chinese_chars": true,
12
  "tokenizer_class": "BertTokenizer",
trainer_state.json CHANGED
@@ -1,283 +1,115 @@
1
  {
2
- "best_metric": 0.9478841870824053,
3
- "best_model_checkpoint": "finetune-bert-Med-SST2/checkpoint-18945",
4
  "epoch": 5.0,
5
- "global_step": 18945,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
10
- {
11
- "epoch": 0.13,
12
- "learning_rate": 4.86803906043811e-05,
13
- "loss": 0.416,
14
- "step": 500
15
- },
16
- {
17
- "epoch": 0.26,
18
- "learning_rate": 4.7360781208762204e-05,
19
- "loss": 0.3286,
20
- "step": 1000
21
- },
22
- {
23
- "epoch": 0.4,
24
- "learning_rate": 4.604117181314331e-05,
25
- "loss": 0.29,
26
- "step": 1500
27
- },
28
  {
29
  "epoch": 0.53,
30
- "learning_rate": 4.472156241752442e-05,
31
- "loss": 0.2661,
32
- "step": 2000
33
- },
34
- {
35
- "epoch": 0.66,
36
- "learning_rate": 4.340195302190552e-05,
37
- "loss": 0.2539,
38
- "step": 2500
39
- },
40
- {
41
- "epoch": 0.79,
42
- "learning_rate": 4.208234362628662e-05,
43
- "loss": 0.2475,
44
- "step": 3000
45
- },
46
- {
47
- "epoch": 0.92,
48
- "learning_rate": 4.076273423066773e-05,
49
- "loss": 0.2378,
50
- "step": 3500
51
  },
52
  {
53
  "epoch": 1.0,
54
- "eval_accuracy": 0.9287305122494433,
55
- "eval_loss": 0.19024422764778137,
56
- "eval_runtime": 10.4888,
57
- "eval_samples_per_second": 642.112,
58
- "eval_steps_per_second": 40.138,
59
- "step": 3789
60
- },
61
- {
62
- "epoch": 1.06,
63
- "learning_rate": 3.944312483504883e-05,
64
- "loss": 0.2065,
65
- "step": 4000
66
- },
67
- {
68
- "epoch": 1.19,
69
- "learning_rate": 3.812351543942993e-05,
70
- "loss": 0.1747,
71
- "step": 4500
72
- },
73
- {
74
- "epoch": 1.32,
75
- "learning_rate": 3.680390604381103e-05,
76
- "loss": 0.1775,
77
- "step": 5000
78
  },
79
  {
80
- "epoch": 1.45,
81
- "learning_rate": 3.5484296648192136e-05,
82
- "loss": 0.1797,
83
- "step": 5500
84
  },
85
  {
86
  "epoch": 1.58,
87
- "learning_rate": 3.4164687252573244e-05,
88
- "loss": 0.1649,
89
- "step": 6000
90
- },
91
- {
92
- "epoch": 1.72,
93
- "learning_rate": 3.2845077856954345e-05,
94
- "loss": 0.1741,
95
- "step": 6500
96
- },
97
- {
98
- "epoch": 1.85,
99
- "learning_rate": 3.1525468461335445e-05,
100
- "loss": 0.1713,
101
- "step": 7000
102
- },
103
- {
104
- "epoch": 1.98,
105
- "learning_rate": 3.0205859065716553e-05,
106
- "loss": 0.1725,
107
- "step": 7500
108
  },
109
  {
110
  "epoch": 2.0,
111
- "eval_accuracy": 0.9449146250927988,
112
- "eval_loss": 0.1960470974445343,
113
- "eval_runtime": 10.3689,
114
- "eval_samples_per_second": 649.541,
115
- "eval_steps_per_second": 40.602,
116
- "step": 7578
117
  },
118
  {
119
  "epoch": 2.11,
120
- "learning_rate": 2.8886249670097653e-05,
121
- "loss": 0.1167,
122
- "step": 8000
123
- },
124
- {
125
- "epoch": 2.24,
126
- "learning_rate": 2.7566640274478754e-05,
127
- "loss": 0.1264,
128
- "step": 8500
129
- },
130
- {
131
- "epoch": 2.38,
132
- "learning_rate": 2.6247030878859858e-05,
133
- "loss": 0.1206,
134
- "step": 9000
135
- },
136
- {
137
- "epoch": 2.51,
138
- "learning_rate": 2.4927421483240962e-05,
139
- "loss": 0.1157,
140
- "step": 9500
141
  },
142
  {
143
  "epoch": 2.64,
144
- "learning_rate": 2.3607812087622063e-05,
145
- "loss": 0.1227,
146
- "step": 10000
147
- },
148
- {
149
- "epoch": 2.77,
150
- "learning_rate": 2.228820269200317e-05,
151
- "loss": 0.1207,
152
- "step": 10500
153
- },
154
- {
155
- "epoch": 2.9,
156
- "learning_rate": 2.096859329638427e-05,
157
- "loss": 0.1201,
158
- "step": 11000
159
  },
160
  {
161
  "epoch": 3.0,
162
- "eval_accuracy": 0.9469933184855234,
163
- "eval_loss": 0.23182542622089386,
164
- "eval_runtime": 10.3744,
165
- "eval_samples_per_second": 649.192,
166
- "eval_steps_per_second": 40.581,
167
- "step": 11367
168
- },
169
- {
170
- "epoch": 3.04,
171
- "learning_rate": 1.9648983900765375e-05,
172
- "loss": 0.1192,
173
- "step": 11500
174
- },
175
- {
176
- "epoch": 3.17,
177
- "learning_rate": 1.8329374505146475e-05,
178
- "loss": 0.0794,
179
- "step": 12000
180
- },
181
- {
182
- "epoch": 3.3,
183
- "learning_rate": 1.7009765109527583e-05,
184
- "loss": 0.0819,
185
- "step": 12500
186
- },
187
- {
188
- "epoch": 3.43,
189
- "learning_rate": 1.5690155713908683e-05,
190
- "loss": 0.0957,
191
- "step": 13000
192
  },
193
  {
194
- "epoch": 3.56,
195
- "learning_rate": 1.4370546318289787e-05,
196
- "loss": 0.0822,
197
- "step": 13500
198
  },
199
  {
200
  "epoch": 3.69,
201
- "learning_rate": 1.305093692267089e-05,
202
- "loss": 0.0909,
203
- "step": 14000
204
- },
205
- {
206
- "epoch": 3.83,
207
- "learning_rate": 1.1731327527051992e-05,
208
- "loss": 0.0895,
209
- "step": 14500
210
- },
211
- {
212
- "epoch": 3.96,
213
- "learning_rate": 1.0411718131433096e-05,
214
- "loss": 0.0811,
215
- "step": 15000
216
  },
217
  {
218
  "epoch": 4.0,
219
- "eval_accuracy": 0.947735708982925,
220
- "eval_loss": 0.2433549165725708,
221
- "eval_runtime": 10.3802,
222
- "eval_samples_per_second": 648.831,
223
- "eval_steps_per_second": 40.558,
224
- "step": 15156
225
- },
226
- {
227
- "epoch": 4.09,
228
- "learning_rate": 9.092108735814198e-06,
229
- "loss": 0.0591,
230
- "step": 15500
231
  },
232
  {
233
  "epoch": 4.22,
234
- "learning_rate": 7.772499340195302e-06,
235
- "loss": 0.0589,
236
- "step": 16000
237
- },
238
- {
239
- "epoch": 4.35,
240
- "learning_rate": 6.4528899445764055e-06,
241
- "loss": 0.0553,
242
- "step": 16500
243
- },
244
- {
245
- "epoch": 4.49,
246
- "learning_rate": 5.133280548957509e-06,
247
- "loss": 0.0576,
248
- "step": 17000
249
- },
250
- {
251
- "epoch": 4.62,
252
- "learning_rate": 3.813671153338612e-06,
253
- "loss": 0.061,
254
- "step": 17500
255
  },
256
  {
257
  "epoch": 4.75,
258
- "learning_rate": 2.494061757719715e-06,
259
- "loss": 0.0533,
260
- "step": 18000
261
- },
262
- {
263
- "epoch": 4.88,
264
- "learning_rate": 1.1744523621008182e-06,
265
- "loss": 0.0648,
266
- "step": 18500
267
  },
268
  {
269
  "epoch": 5.0,
270
- "eval_accuracy": 0.9478841870824053,
271
- "eval_loss": 0.27820324897766113,
272
- "eval_runtime": 10.4353,
273
- "eval_samples_per_second": 645.403,
274
- "eval_steps_per_second": 40.344,
275
- "step": 18945
276
  }
277
  ],
278
- "max_steps": 18945,
279
  "num_train_epochs": 5,
280
- "total_flos": 1630739369585616.0,
281
  "trial_name": null,
282
  "trial_params": null
283
  }
 
1
  {
2
+ "best_metric": 0.9498144023756496,
3
+ "best_model_checkpoint": "finetune-bert-Med-SST2\\checkpoint-3792",
4
  "epoch": 5.0,
5
+ "global_step": 4740,
6
  "is_hyper_param_search": false,
7
  "is_local_process_zero": true,
8
  "is_world_process_zero": true,
9
  "log_history": [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  {
11
  "epoch": 0.53,
12
+ "learning_rate": 4.4725738396624474e-05,
13
+ "loss": 0.3201,
14
+ "step": 500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  },
16
  {
17
  "epoch": 1.0,
18
+ "eval_accuracy": 0.9319970304380104,
19
+ "eval_loss": 0.18009190261363983,
20
+ "eval_runtime": 8.6367,
21
+ "eval_samples_per_second": 779.81,
22
+ "eval_steps_per_second": 12.273,
23
+ "step": 948
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  },
25
  {
26
+ "epoch": 1.05,
27
+ "learning_rate": 3.945147679324895e-05,
28
+ "loss": 0.2192,
29
+ "step": 1000
30
  },
31
  {
32
  "epoch": 1.58,
33
+ "learning_rate": 3.4177215189873416e-05,
34
+ "loss": 0.1468,
35
+ "step": 1500
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
36
  },
37
  {
38
  "epoch": 2.0,
39
+ "eval_accuracy": 0.9452115812917594,
40
+ "eval_loss": 0.16460973024368286,
41
+ "eval_runtime": 9.1827,
42
+ "eval_samples_per_second": 733.445,
43
+ "eval_steps_per_second": 11.543,
44
+ "step": 1896
45
  },
46
  {
47
  "epoch": 2.11,
48
+ "learning_rate": 2.8902953586497894e-05,
49
+ "loss": 0.1255,
50
+ "step": 2000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
51
  },
52
  {
53
  "epoch": 2.64,
54
+ "learning_rate": 2.3628691983122365e-05,
55
+ "loss": 0.094,
56
+ "step": 2500
 
 
 
 
 
 
 
 
 
 
 
 
57
  },
58
  {
59
  "epoch": 3.0,
60
+ "eval_accuracy": 0.9483296213808463,
61
+ "eval_loss": 0.17333245277404785,
62
+ "eval_runtime": 9.4505,
63
+ "eval_samples_per_second": 712.663,
64
+ "eval_steps_per_second": 11.216,
65
+ "step": 2844
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
66
  },
67
  {
68
+ "epoch": 3.16,
69
+ "learning_rate": 1.8354430379746836e-05,
70
+ "loss": 0.0816,
71
+ "step": 3000
72
  },
73
  {
74
  "epoch": 3.69,
75
+ "learning_rate": 1.308016877637131e-05,
76
+ "loss": 0.0645,
77
+ "step": 3500
 
 
 
 
 
 
 
 
 
 
 
 
78
  },
79
  {
80
  "epoch": 4.0,
81
+ "eval_accuracy": 0.9498144023756496,
82
+ "eval_loss": 0.19728632271289825,
83
+ "eval_runtime": 9.0479,
84
+ "eval_samples_per_second": 744.374,
85
+ "eval_steps_per_second": 11.715,
86
+ "step": 3792
 
 
 
 
 
 
87
  },
88
  {
89
  "epoch": 4.22,
90
+ "learning_rate": 7.805907172995782e-06,
91
+ "loss": 0.0538,
92
+ "step": 4000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
93
  },
94
  {
95
  "epoch": 4.75,
96
+ "learning_rate": 2.531645569620253e-06,
97
+ "loss": 0.0424,
98
+ "step": 4500
 
 
 
 
 
 
99
  },
100
  {
101
  "epoch": 5.0,
102
+ "eval_accuracy": 0.9484780994803267,
103
+ "eval_loss": 0.2303222268819809,
104
+ "eval_runtime": 9.3041,
105
+ "eval_samples_per_second": 723.873,
106
+ "eval_steps_per_second": 11.393,
107
+ "step": 4740
108
  }
109
  ],
110
+ "max_steps": 4740,
111
  "num_train_epochs": 5,
112
+ "total_flos": 2062420909600608.0,
113
  "trial_name": null,
114
  "trial_params": null
115
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d28093861817ce584c9417a32a2951733caa888fd46d56babe7bda6b3efd018f
3
- size 3899
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:541291882fa1c55858b0ae60dc84c3ac0b0b4904fc379c3ea1509409619514bf
3
+ size 3515