jialicheng commited on
Commit
580c816
·
verified ·
1 Parent(s): 0c1ad9b

Upload folder using huggingface_hub

Browse files
all_results.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ave/id": 0.7202648869574215,
3
+ "ave/ood": 0.6133564614050304,
4
+ "ave/stress": 76.8397970175716,
5
+ "ave/transfer": 0.8147013782542112,
6
+ "epoch": 3.0,
7
+ "eval_accuracy": 70.34313725490196,
8
+ "eval_combined_score": 0.8147013782542112,
9
+ "eval_f1": 0.8147013782542112,
10
+ "eval_loss": 0.7417141795158386,
11
+ "eval_runtime": 1.0302,
12
+ "eval_samples_per_second": 396.047,
13
+ "eval_steps_per_second": 1.941,
14
+ "gap/ood": 0.1069084255523911,
15
+ "gap/stress": -4.813308321829439,
16
+ "gap/transfer": -0.09443649129678966,
17
+ "ood": 0.6133564614050304,
18
+ "qqp": 0.7202648869574215,
19
+ "stress-antonym": 98.87050936384173,
20
+ "stress-length_mismatch": 70.7109634551495,
21
+ "stress-negation": 72.69133164151195,
22
+ "stress-spelling_error": 69.53682797339354,
23
+ "stress-word_overlap": 72.38935265396125,
24
+ "train_loss": 0.950482799304254,
25
+ "train_runtime": 3151.1724,
26
+ "train_samples_per_second": 346.391,
27
+ "train_steps_per_second": 10.825,
28
+ "transfer": 0.8147013782542112
29
+ }
ood_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_combined_score": 0.6133564614050304,
3
+ "eval_f1": 0.6133564614050304,
4
+ "eval_loss": 1.627290964126587,
5
+ "eval_runtime": 3.5206,
6
+ "eval_samples_per_second": 2272.336,
7
+ "eval_steps_per_second": 9.089
8
+ }
pred_ood.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1260f432e3e474085be4e6840c043d3091554ab9bf6b0c67fd64bd88487c3888
3
+ size 64128
pred_qqp.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f68d11e49fa2e6a6e096f4d96c5acaa5365a8e5186c0d6feb3c29e406e93ebc
3
+ size 323568
pred_stress-antonym.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55809ffe9e5c570100bed7eb2273476c085a73baefec21de92b52d79d51714bc
3
+ size 55640
pred_stress-length_mismatch.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9baa8046e626df3ae118b6e5a071b1cdacdd38ff7ad385f2f9fadaa698f992f0
3
+ size 323568
pred_stress-negation.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f458c476865214b1ca8de0cbd8f500a327920c2eb013810d770eee87f7bd989e
3
+ size 323568
pred_stress-spelling_error.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb344c68ea31d81dc8b67d49bf7f5921691b9bfc0c28d4fb99d2e425df16d98d
3
+ size 323560
pred_stress-word_overlap.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f5a44194663d846c26900c6bb4ee1a59329e9dd4ffeb717d303a76c25130937
3
+ size 323568
pred_transfer.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d88ad499efb00addb7a2ebc39c9d353f08af7148e0fb3583c9b46e363f43f69
3
+ size 3392
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f52ae1fd202f0ab7d22da42963d40a4316ad4cd21de7410d3445cfa65555fd7a
3
+ size 17907539
qqp_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_combined_score": 0.7202648869574215,
3
+ "eval_f1": 0.7202648869574215,
4
+ "eval_loss": 0.5221115350723267,
5
+ "eval_runtime": 13.1461,
6
+ "eval_samples_per_second": 3075.444,
7
+ "eval_steps_per_second": 12.019
8
+ }
rep_ood.npz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:877862d151cd253d979b4d74cc9db19182f0e4ce309fc07fbaadfd4510b1837e
3
+ size 2097169238
rep_ood.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eb07f1cd33343ad1dc065284bc46630d2e4468225872140cf80324d0c4e380d0
3
+ size 8208726
rep_qqp.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:25bbb44f6ab1c9a6bb4d0b20282c7cc4e289bba7f3371b10371120428770865a
3
+ size 41484538
rep_transfer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28955c7836a2a7b4867aad73d88ee1eee5414a6432f575401ee7a44a903c41a6
3
+ size 419832
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
stress-antonym_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 97.76624873901139,
4
+ "eval_combined_score": 98.31837905142656,
5
+ "eval_f1": 98.87050936384173,
6
+ "eval_loss": 0.15600410103797913,
7
+ "eval_runtime": 16.709,
8
+ "eval_samples_per_second": 415.286,
9
+ "eval_steps_per_second": 1.676
10
+ }
stress-length_mismatch_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 78.1944100915162,
4
+ "eval_combined_score": 74.45268677333286,
5
+ "eval_f1": 70.7109634551495,
6
+ "eval_loss": 0.44404110312461853,
7
+ "eval_runtime": 95.6539,
8
+ "eval_samples_per_second": 422.67,
9
+ "eval_steps_per_second": 1.652
10
+ }
stress-negation_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 74.71432104872619,
4
+ "eval_combined_score": 73.70282634511906,
5
+ "eval_f1": 72.69133164151195,
6
+ "eval_loss": 0.5010897517204285,
7
+ "eval_runtime": 96.6967,
8
+ "eval_samples_per_second": 418.111,
9
+ "eval_steps_per_second": 1.634
10
+ }
stress-spelling_error_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 75.30485542556086,
4
+ "eval_combined_score": 72.4208416994772,
5
+ "eval_f1": 69.53682797339354,
6
+ "eval_loss": 0.4926765561103821,
7
+ "eval_runtime": 95.7074,
8
+ "eval_samples_per_second": 422.423,
9
+ "eval_steps_per_second": 1.651
10
+ }
stress-word_overlap_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 73.98466485283205,
4
+ "eval_combined_score": 73.18700875339665,
5
+ "eval_f1": 72.38935265396125,
6
+ "eval_loss": 0.5098819732666016,
7
+ "eval_runtime": 96.8935,
8
+ "eval_samples_per_second": 417.262,
9
+ "eval_steps_per_second": 1.631
10
+ }
test_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ave/id": 0.7202648869574215,
3
+ "ave/ood": 0.6133564614050304,
4
+ "ave/transfer": 0.8147013782542112,
5
+ "gap/ood": 0.1069084255523911,
6
+ "gap/transfer": -0.09443649129678966,
7
+ "ood": 0.6133564614050304,
8
+ "qqp": 0.7202648869574215,
9
+ "transfer": 0.8147013782542112
10
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
+ "do_lower_case": true,
5
+ "mask_token": "[MASK]",
6
+ "model_max_length": 1000000000000000019884624838656,
7
+ "name_or_path": "google/bert_uncased_L-2_H-128_A-2",
8
+ "never_split": null,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "special_tokens_map_file": null,
12
+ "strip_accents": null,
13
+ "tokenize_chinese_chars": true,
14
+ "tokenizer_class": "BertTokenizer",
15
+ "unk_token": "[UNK]"
16
+ }
train_results.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "train_loss": 0.950482799304254,
4
+ "train_runtime": 3151.1724,
5
+ "train_samples_per_second": 346.391,
6
+ "train_steps_per_second": 10.825
7
+ }
trainer_state.json ADDED
@@ -0,0 +1,466 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 72.02648869574216,
3
+ "best_model_checkpoint": "outputs/bert-tiny/read/qqp_87/checkpoint-34113",
4
+ "epoch": 3.0,
5
+ "global_step": 34113,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.04,
12
+ "learning_rate": 1.9706856623574592e-05,
13
+ "loss": 1.2441,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 0.09,
18
+ "learning_rate": 1.9413713247149183e-05,
19
+ "loss": 1.1557,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 0.13,
24
+ "learning_rate": 1.9120569870723774e-05,
25
+ "loss": 1.1329,
26
+ "step": 1500
27
+ },
28
+ {
29
+ "epoch": 0.18,
30
+ "learning_rate": 1.882742649429836e-05,
31
+ "loss": 1.1005,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 0.22,
36
+ "learning_rate": 1.8534283117872952e-05,
37
+ "loss": 1.0837,
38
+ "step": 2500
39
+ },
40
+ {
41
+ "epoch": 0.26,
42
+ "learning_rate": 1.8241139741447543e-05,
43
+ "loss": 1.0757,
44
+ "step": 3000
45
+ },
46
+ {
47
+ "epoch": 0.31,
48
+ "learning_rate": 1.7947996365022134e-05,
49
+ "loss": 1.0474,
50
+ "step": 3500
51
+ },
52
+ {
53
+ "epoch": 0.35,
54
+ "learning_rate": 1.7654852988596725e-05,
55
+ "loss": 1.0291,
56
+ "step": 4000
57
+ },
58
+ {
59
+ "epoch": 0.4,
60
+ "learning_rate": 1.7361709612171312e-05,
61
+ "loss": 1.034,
62
+ "step": 4500
63
+ },
64
+ {
65
+ "epoch": 0.44,
66
+ "learning_rate": 1.7068566235745903e-05,
67
+ "loss": 1.03,
68
+ "step": 5000
69
+ },
70
+ {
71
+ "epoch": 0.48,
72
+ "learning_rate": 1.6775422859320497e-05,
73
+ "loss": 1.0145,
74
+ "step": 5500
75
+ },
76
+ {
77
+ "epoch": 0.53,
78
+ "learning_rate": 1.6482279482895085e-05,
79
+ "loss": 1.01,
80
+ "step": 6000
81
+ },
82
+ {
83
+ "epoch": 0.57,
84
+ "learning_rate": 1.6189136106469676e-05,
85
+ "loss": 1.0004,
86
+ "step": 6500
87
+ },
88
+ {
89
+ "epoch": 0.62,
90
+ "learning_rate": 1.5895992730044266e-05,
91
+ "loss": 1.0056,
92
+ "step": 7000
93
+ },
94
+ {
95
+ "epoch": 0.66,
96
+ "learning_rate": 1.5602849353618857e-05,
97
+ "loss": 0.9981,
98
+ "step": 7500
99
+ },
100
+ {
101
+ "epoch": 0.7,
102
+ "learning_rate": 1.5309705977193448e-05,
103
+ "loss": 0.9831,
104
+ "step": 8000
105
+ },
106
+ {
107
+ "epoch": 0.75,
108
+ "learning_rate": 1.5016562600768037e-05,
109
+ "loss": 0.9832,
110
+ "step": 8500
111
+ },
112
+ {
113
+ "epoch": 0.79,
114
+ "learning_rate": 1.4723419224342628e-05,
115
+ "loss": 0.9904,
116
+ "step": 9000
117
+ },
118
+ {
119
+ "epoch": 0.84,
120
+ "learning_rate": 1.4430275847917217e-05,
121
+ "loss": 0.9848,
122
+ "step": 9500
123
+ },
124
+ {
125
+ "epoch": 0.88,
126
+ "learning_rate": 1.4137132471491808e-05,
127
+ "loss": 0.9755,
128
+ "step": 10000
129
+ },
130
+ {
131
+ "epoch": 0.92,
132
+ "learning_rate": 1.3843989095066399e-05,
133
+ "loss": 0.9573,
134
+ "step": 10500
135
+ },
136
+ {
137
+ "epoch": 0.97,
138
+ "learning_rate": 1.3550845718640988e-05,
139
+ "loss": 0.9667,
140
+ "step": 11000
141
+ },
142
+ {
143
+ "epoch": 1.0,
144
+ "eval_accuracy": 67.91244125649271,
145
+ "eval_combined_score": 68.34303011193873,
146
+ "eval_f1": 68.77361896738476,
147
+ "eval_loss": 0.5894622802734375,
148
+ "eval_runtime": 90.4161,
149
+ "eval_samples_per_second": 447.155,
150
+ "eval_steps_per_second": 1.747,
151
+ "step": 11371
152
+ },
153
+ {
154
+ "epoch": 1.01,
155
+ "learning_rate": 1.3257702342215579e-05,
156
+ "loss": 0.9553,
157
+ "step": 11500
158
+ },
159
+ {
160
+ "epoch": 1.06,
161
+ "learning_rate": 1.2964558965790168e-05,
162
+ "loss": 0.9386,
163
+ "step": 12000
164
+ },
165
+ {
166
+ "epoch": 1.1,
167
+ "learning_rate": 1.2671415589364759e-05,
168
+ "loss": 0.9366,
169
+ "step": 12500
170
+ },
171
+ {
172
+ "epoch": 1.14,
173
+ "learning_rate": 1.2378272212939351e-05,
174
+ "loss": 0.94,
175
+ "step": 13000
176
+ },
177
+ {
178
+ "epoch": 1.19,
179
+ "learning_rate": 1.208512883651394e-05,
180
+ "loss": 0.9323,
181
+ "step": 13500
182
+ },
183
+ {
184
+ "epoch": 1.23,
185
+ "learning_rate": 1.1791985460088531e-05,
186
+ "loss": 0.9444,
187
+ "step": 14000
188
+ },
189
+ {
190
+ "epoch": 1.28,
191
+ "learning_rate": 1.149884208366312e-05,
192
+ "loss": 0.9268,
193
+ "step": 14500
194
+ },
195
+ {
196
+ "epoch": 1.32,
197
+ "learning_rate": 1.1205698707237711e-05,
198
+ "loss": 0.9294,
199
+ "step": 15000
200
+ },
201
+ {
202
+ "epoch": 1.36,
203
+ "learning_rate": 1.0912555330812302e-05,
204
+ "loss": 0.9281,
205
+ "step": 15500
206
+ },
207
+ {
208
+ "epoch": 1.41,
209
+ "learning_rate": 1.0619411954386891e-05,
210
+ "loss": 0.9398,
211
+ "step": 16000
212
+ },
213
+ {
214
+ "epoch": 1.45,
215
+ "learning_rate": 1.0326268577961482e-05,
216
+ "loss": 0.9178,
217
+ "step": 16500
218
+ },
219
+ {
220
+ "epoch": 1.5,
221
+ "learning_rate": 1.0033125201536071e-05,
222
+ "loss": 0.9188,
223
+ "step": 17000
224
+ },
225
+ {
226
+ "epoch": 1.54,
227
+ "learning_rate": 9.739981825110662e-06,
228
+ "loss": 0.9254,
229
+ "step": 17500
230
+ },
231
+ {
232
+ "epoch": 1.58,
233
+ "learning_rate": 9.446838448685253e-06,
234
+ "loss": 0.9184,
235
+ "step": 18000
236
+ },
237
+ {
238
+ "epoch": 1.63,
239
+ "learning_rate": 9.153695072259843e-06,
240
+ "loss": 0.9133,
241
+ "step": 18500
242
+ },
243
+ {
244
+ "epoch": 1.67,
245
+ "learning_rate": 8.860551695834433e-06,
246
+ "loss": 0.928,
247
+ "step": 19000
248
+ },
249
+ {
250
+ "epoch": 1.71,
251
+ "learning_rate": 8.567408319409023e-06,
252
+ "loss": 0.9055,
253
+ "step": 19500
254
+ },
255
+ {
256
+ "epoch": 1.76,
257
+ "learning_rate": 8.274264942983614e-06,
258
+ "loss": 0.9082,
259
+ "step": 20000
260
+ },
261
+ {
262
+ "epoch": 1.8,
263
+ "learning_rate": 7.981121566558205e-06,
264
+ "loss": 0.9268,
265
+ "step": 20500
266
+ },
267
+ {
268
+ "epoch": 1.85,
269
+ "learning_rate": 7.687978190132794e-06,
270
+ "loss": 0.9281,
271
+ "step": 21000
272
+ },
273
+ {
274
+ "epoch": 1.89,
275
+ "learning_rate": 7.394834813707384e-06,
276
+ "loss": 0.9294,
277
+ "step": 21500
278
+ },
279
+ {
280
+ "epoch": 1.93,
281
+ "learning_rate": 7.101691437281976e-06,
282
+ "loss": 0.9226,
283
+ "step": 22000
284
+ },
285
+ {
286
+ "epoch": 1.98,
287
+ "learning_rate": 6.808548060856566e-06,
288
+ "loss": 0.9154,
289
+ "step": 22500
290
+ },
291
+ {
292
+ "epoch": 2.0,
293
+ "eval_accuracy": 71.36532278011379,
294
+ "eval_combined_score": 71.09875067982031,
295
+ "eval_f1": 70.83217857952684,
296
+ "eval_loss": 0.5470175743103027,
297
+ "eval_runtime": 89.5769,
298
+ "eval_samples_per_second": 451.344,
299
+ "eval_steps_per_second": 1.764,
300
+ "step": 22742
301
+ },
302
+ {
303
+ "epoch": 2.02,
304
+ "learning_rate": 6.515404684431156e-06,
305
+ "loss": 0.9096,
306
+ "step": 23000
307
+ },
308
+ {
309
+ "epoch": 2.07,
310
+ "learning_rate": 6.222261308005746e-06,
311
+ "loss": 0.899,
312
+ "step": 23500
313
+ },
314
+ {
315
+ "epoch": 2.11,
316
+ "learning_rate": 5.929117931580336e-06,
317
+ "loss": 0.8881,
318
+ "step": 24000
319
+ },
320
+ {
321
+ "epoch": 2.15,
322
+ "learning_rate": 5.6359745551549274e-06,
323
+ "loss": 0.8952,
324
+ "step": 24500
325
+ },
326
+ {
327
+ "epoch": 2.2,
328
+ "learning_rate": 5.3428311787295174e-06,
329
+ "loss": 0.8906,
330
+ "step": 25000
331
+ },
332
+ {
333
+ "epoch": 2.24,
334
+ "learning_rate": 5.049687802304107e-06,
335
+ "loss": 0.9021,
336
+ "step": 25500
337
+ },
338
+ {
339
+ "epoch": 2.29,
340
+ "learning_rate": 4.756544425878697e-06,
341
+ "loss": 0.8924,
342
+ "step": 26000
343
+ },
344
+ {
345
+ "epoch": 2.33,
346
+ "learning_rate": 4.463401049453288e-06,
347
+ "loss": 0.8812,
348
+ "step": 26500
349
+ },
350
+ {
351
+ "epoch": 2.37,
352
+ "learning_rate": 4.170257673027878e-06,
353
+ "loss": 0.892,
354
+ "step": 27000
355
+ },
356
+ {
357
+ "epoch": 2.42,
358
+ "learning_rate": 3.877114296602469e-06,
359
+ "loss": 0.8974,
360
+ "step": 27500
361
+ },
362
+ {
363
+ "epoch": 2.46,
364
+ "learning_rate": 3.583970920177059e-06,
365
+ "loss": 0.893,
366
+ "step": 28000
367
+ },
368
+ {
369
+ "epoch": 2.51,
370
+ "learning_rate": 3.290827543751649e-06,
371
+ "loss": 0.8885,
372
+ "step": 28500
373
+ },
374
+ {
375
+ "epoch": 2.55,
376
+ "learning_rate": 2.9976841673262398e-06,
377
+ "loss": 0.8936,
378
+ "step": 29000
379
+ },
380
+ {
381
+ "epoch": 2.59,
382
+ "learning_rate": 2.7045407909008298e-06,
383
+ "loss": 0.8972,
384
+ "step": 29500
385
+ },
386
+ {
387
+ "epoch": 2.64,
388
+ "learning_rate": 2.41139741447542e-06,
389
+ "loss": 0.8764,
390
+ "step": 30000
391
+ },
392
+ {
393
+ "epoch": 2.68,
394
+ "learning_rate": 2.1182540380500106e-06,
395
+ "loss": 0.8957,
396
+ "step": 30500
397
+ },
398
+ {
399
+ "epoch": 2.73,
400
+ "learning_rate": 1.825110661624601e-06,
401
+ "loss": 0.8902,
402
+ "step": 31000
403
+ },
404
+ {
405
+ "epoch": 2.77,
406
+ "learning_rate": 1.531967285199191e-06,
407
+ "loss": 0.8869,
408
+ "step": 31500
409
+ },
410
+ {
411
+ "epoch": 2.81,
412
+ "learning_rate": 1.2388239087737813e-06,
413
+ "loss": 0.8809,
414
+ "step": 32000
415
+ },
416
+ {
417
+ "epoch": 2.86,
418
+ "learning_rate": 9.456805323483717e-07,
419
+ "loss": 0.8876,
420
+ "step": 32500
421
+ },
422
+ {
423
+ "epoch": 2.9,
424
+ "learning_rate": 6.525371559229619e-07,
425
+ "loss": 0.9025,
426
+ "step": 33000
427
+ },
428
+ {
429
+ "epoch": 2.95,
430
+ "learning_rate": 3.5939377949755233e-07,
431
+ "loss": 0.872,
432
+ "step": 33500
433
+ },
434
+ {
435
+ "epoch": 2.99,
436
+ "learning_rate": 6.62504030721426e-08,
437
+ "loss": 0.8964,
438
+ "step": 34000
439
+ },
440
+ {
441
+ "epoch": 3.0,
442
+ "eval_accuracy": 73.25253524610437,
443
+ "eval_combined_score": 72.63951197092327,
444
+ "eval_f1": 72.02648869574216,
445
+ "eval_loss": 0.5221115350723267,
446
+ "eval_runtime": 89.4824,
447
+ "eval_samples_per_second": 451.821,
448
+ "eval_steps_per_second": 1.766,
449
+ "step": 34113
450
+ },
451
+ {
452
+ "epoch": 3.0,
453
+ "step": 34113,
454
+ "total_flos": 0.0,
455
+ "train_loss": 0.950482799304254,
456
+ "train_runtime": 3151.1724,
457
+ "train_samples_per_second": 346.391,
458
+ "train_steps_per_second": 10.825
459
+ }
460
+ ],
461
+ "max_steps": 34113,
462
+ "num_train_epochs": 3,
463
+ "total_flos": 0.0,
464
+ "trial_name": null,
465
+ "trial_params": null
466
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:19b26c5be5cf5528350e46bfe6be2c85bd79b47a1f52c8e8b8b01cd8bbd3c8c8
3
+ size 4091
transfer_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_combined_score": 0.8147013782542112,
3
+ "eval_f1": 0.8147013782542112,
4
+ "eval_loss": 0.7417141795158386,
5
+ "eval_runtime": 1.0302,
6
+ "eval_samples_per_second": 396.047,
7
+ "eval_steps_per_second": 1.941
8
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff