jialicheng commited on
Commit
7c8f80f
·
verified ·
1 Parent(s): 58c3d56

Upload folder using huggingface_hub

Browse files
all_results.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ave/id": 0.814885685064828,
3
+ "ave/ood": 0.6040905770635501,
4
+ "ave/stress": 77.33218437992298,
5
+ "ave/transfer": 0.7922077922077924,
6
+ "epoch": 3.0,
7
+ "eval_accuracy": 68.62745098039215,
8
+ "eval_combined_score": 0.7922077922077924,
9
+ "eval_f1": 0.7922077922077924,
10
+ "eval_loss": 0.8416510224342346,
11
+ "eval_runtime": 2.1923,
12
+ "eval_samples_per_second": 186.11,
13
+ "eval_steps_per_second": 0.912,
14
+ "gap/ood": 0.21079510800127788,
15
+ "gap/stress": 4.156384126559814,
16
+ "gap/transfer": 0.022677892857035586,
17
+ "ood": 0.6040905770635501,
18
+ "qqp": 0.814885685064828,
19
+ "stress-antonym": 70.67374895163545,
20
+ "stress-length_mismatch": 77.4223759095166,
21
+ "stress-negation": 81.71691107055196,
22
+ "stress-spelling_error": 75.08700102354146,
23
+ "stress-word_overlap": 81.76088494436941,
24
+ "train_loss": 0.6687755357152263,
25
+ "train_runtime": 7270.4925,
26
+ "train_samples_per_second": 150.133,
27
+ "train_steps_per_second": 4.692,
28
+ "transfer": 0.7922077922077924
29
+ }
ood_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_combined_score": 0.6040905770635501,
3
+ "eval_f1": 0.6040905770635501,
4
+ "eval_loss": 1.6317058801651,
5
+ "eval_runtime": 23.59,
6
+ "eval_samples_per_second": 339.127,
7
+ "eval_steps_per_second": 1.357
8
+ }
pred_ood.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:147fde8e4388463262853c50aaee36e29a83340f798b320a7ab57a013a622621
3
+ size 64128
pred_qqp.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26f816519dcc408966dc4eb6335fde7974a7ee7a4a270bb4b58a1b3f821c4ddb
3
+ size 323568
pred_stress-antonym.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:62fe94b8a7ff316f15958517d13505a9b89e2f328c28b21a3bdb7711ff37866a
3
+ size 55640
pred_stress-length_mismatch.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3690046dcf8126eb96ee8a80ca6bf64bca26df68d771875d3d5ad51d7f260a5
3
+ size 323568
pred_stress-negation.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4da9eab8378a0abf4c762d269d10b1f9664821de70bd99263ced080dc5b862a8
3
+ size 323568
pred_stress-spelling_error.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:478a94e2e102173ad0150d0a300da94ffda7081b6b6048cb9ac650b1aca667ad
3
+ size 323560
pred_stress-word_overlap.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c0962808319090258cfd4de1923ad8f02d759cdee12b36c3cd230f7f09a155e
3
+ size 323568
pred_transfer.npz.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc0b7c6dff30c24b4ba20e9a55a97f1252e5ac82c2a1023ad4ca097a46585948
3
+ size 3392
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35442494bfd0ff646f239ed1d2e14f7d4131ef4b867deb09542351de2a22fc7a
3
+ size 183448325
qqp_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_combined_score": 0.814885685064828,
3
+ "eval_f1": 0.814885685064828,
4
+ "eval_loss": 0.3177751898765564,
5
+ "eval_runtime": 112.2847,
6
+ "eval_samples_per_second": 360.067,
7
+ "eval_steps_per_second": 1.407
8
+ }
rep_ood.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddb89c1bfa9bf260691246d03b2b479d4094d2e506fe3758efb36189401e309b
3
+ size 131141490
rep_qqp.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65f1bce10e7e1bbaee84c4eb49ebf16509c603acd9c0ac144fb6f324a92b648a
3
+ size 662749338
rep_transfer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c99473290b0c47facac82de2ba2ed4c7c6449048f9727315c26563b06205a9c
3
+ size 6690122
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
stress-antonym_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 54.647643752702116,
4
+ "eval_combined_score": 62.660696352168785,
5
+ "eval_f1": 70.67374895163545,
6
+ "eval_loss": 1.0299919843673706,
7
+ "eval_runtime": 18.1327,
8
+ "eval_samples_per_second": 382.68,
9
+ "eval_steps_per_second": 1.544
10
+ }
stress-length_mismatch_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 84.49666089537472,
4
+ "eval_combined_score": 80.95951840244567,
5
+ "eval_f1": 77.4223759095166,
6
+ "eval_loss": 0.35398587584495544,
7
+ "eval_runtime": 106.1714,
8
+ "eval_samples_per_second": 380.799,
9
+ "eval_steps_per_second": 1.488
10
+ }
stress-negation_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 85.62948305713579,
4
+ "eval_combined_score": 83.67319706384387,
5
+ "eval_f1": 81.71691107055196,
6
+ "eval_loss": 0.35075807571411133,
7
+ "eval_runtime": 105.8746,
8
+ "eval_samples_per_second": 381.867,
9
+ "eval_steps_per_second": 1.492
10
+ }
stress-spelling_error_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 81.93870736352619,
4
+ "eval_combined_score": 78.51285419353383,
5
+ "eval_f1": 75.08700102354146,
6
+ "eval_loss": 0.4257154166698456,
7
+ "eval_runtime": 106.3656,
8
+ "eval_samples_per_second": 380.095,
9
+ "eval_steps_per_second": 1.485
10
+ }
stress-word_overlap_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "eval_accuracy": 85.97081375216423,
4
+ "eval_combined_score": 83.86584934826682,
5
+ "eval_f1": 81.76088494436941,
6
+ "eval_loss": 0.33746978640556335,
7
+ "eval_runtime": 106.1412,
8
+ "eval_samples_per_second": 380.908,
9
+ "eval_steps_per_second": 1.489
10
+ }
test_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "ave/id": 0.814885685064828,
3
+ "ave/ood": 0.6040905770635501,
4
+ "ave/transfer": 0.7922077922077924,
5
+ "gap/ood": 0.21079510800127788,
6
+ "gap/transfer": 0.022677892857035586,
7
+ "ood": 0.6040905770635501,
8
+ "qqp": 0.814885685064828,
9
+ "transfer": 0.7922077922077924
10
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "do_basic_tokenize": true,
4
+ "do_lower_case": true,
5
+ "mask_token": "[MASK]",
6
+ "model_max_length": 1000000000000000019884624838656,
7
+ "name_or_path": "google/bert_uncased_L-8_H-512_A-8",
8
+ "never_split": null,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "special_tokens_map_file": null,
12
+ "strip_accents": null,
13
+ "tokenize_chinese_chars": true,
14
+ "tokenizer_class": "BertTokenizer",
15
+ "unk_token": "[UNK]"
16
+ }
train_results.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "train_loss": 0.6687755357152263,
4
+ "train_runtime": 7270.4925,
5
+ "train_samples_per_second": 150.133,
6
+ "train_steps_per_second": 4.692
7
+ }
trainer_state.json ADDED
@@ -0,0 +1,466 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 81.48856850648279,
3
+ "best_model_checkpoint": "outputs/bert-medium/read/qqp_21/checkpoint-34113",
4
+ "epoch": 3.0,
5
+ "global_step": 34113,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.04,
12
+ "learning_rate": 1.9706856623574592e-05,
13
+ "loss": 1.1052,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 0.09,
18
+ "learning_rate": 1.9413713247149183e-05,
19
+ "loss": 0.9827,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 0.13,
24
+ "learning_rate": 1.9120569870723774e-05,
25
+ "loss": 0.9411,
26
+ "step": 1500
27
+ },
28
+ {
29
+ "epoch": 0.18,
30
+ "learning_rate": 1.882742649429836e-05,
31
+ "loss": 0.9184,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 0.22,
36
+ "learning_rate": 1.8534283117872952e-05,
37
+ "loss": 0.8824,
38
+ "step": 2500
39
+ },
40
+ {
41
+ "epoch": 0.26,
42
+ "learning_rate": 1.8241139741447543e-05,
43
+ "loss": 0.8506,
44
+ "step": 3000
45
+ },
46
+ {
47
+ "epoch": 0.31,
48
+ "learning_rate": 1.7947996365022134e-05,
49
+ "loss": 0.8293,
50
+ "step": 3500
51
+ },
52
+ {
53
+ "epoch": 0.35,
54
+ "learning_rate": 1.7654852988596725e-05,
55
+ "loss": 0.8302,
56
+ "step": 4000
57
+ },
58
+ {
59
+ "epoch": 0.4,
60
+ "learning_rate": 1.7361709612171312e-05,
61
+ "loss": 0.8115,
62
+ "step": 4500
63
+ },
64
+ {
65
+ "epoch": 0.44,
66
+ "learning_rate": 1.7068566235745903e-05,
67
+ "loss": 0.7882,
68
+ "step": 5000
69
+ },
70
+ {
71
+ "epoch": 0.48,
72
+ "learning_rate": 1.6775422859320497e-05,
73
+ "loss": 0.802,
74
+ "step": 5500
75
+ },
76
+ {
77
+ "epoch": 0.53,
78
+ "learning_rate": 1.6482279482895085e-05,
79
+ "loss": 0.7852,
80
+ "step": 6000
81
+ },
82
+ {
83
+ "epoch": 0.57,
84
+ "learning_rate": 1.6189136106469676e-05,
85
+ "loss": 0.7845,
86
+ "step": 6500
87
+ },
88
+ {
89
+ "epoch": 0.62,
90
+ "learning_rate": 1.5895992730044266e-05,
91
+ "loss": 0.7755,
92
+ "step": 7000
93
+ },
94
+ {
95
+ "epoch": 0.66,
96
+ "learning_rate": 1.5602849353618857e-05,
97
+ "loss": 0.7744,
98
+ "step": 7500
99
+ },
100
+ {
101
+ "epoch": 0.7,
102
+ "learning_rate": 1.5309705977193448e-05,
103
+ "loss": 0.7629,
104
+ "step": 8000
105
+ },
106
+ {
107
+ "epoch": 0.75,
108
+ "learning_rate": 1.5016562600768037e-05,
109
+ "loss": 0.7497,
110
+ "step": 8500
111
+ },
112
+ {
113
+ "epoch": 0.79,
114
+ "learning_rate": 1.4723419224342628e-05,
115
+ "loss": 0.7299,
116
+ "step": 9000
117
+ },
118
+ {
119
+ "epoch": 0.84,
120
+ "learning_rate": 1.4430275847917217e-05,
121
+ "loss": 0.7426,
122
+ "step": 9500
123
+ },
124
+ {
125
+ "epoch": 0.88,
126
+ "learning_rate": 1.4137132471491808e-05,
127
+ "loss": 0.7292,
128
+ "step": 10000
129
+ },
130
+ {
131
+ "epoch": 0.92,
132
+ "learning_rate": 1.3843989095066399e-05,
133
+ "loss": 0.7333,
134
+ "step": 10500
135
+ },
136
+ {
137
+ "epoch": 0.97,
138
+ "learning_rate": 1.3550845718640988e-05,
139
+ "loss": 0.7246,
140
+ "step": 11000
141
+ },
142
+ {
143
+ "epoch": 1.0,
144
+ "eval_accuracy": 83.31189710610933,
145
+ "eval_combined_score": 81.1331624399507,
146
+ "eval_f1": 78.95442777379208,
147
+ "eval_loss": 0.36778926849365234,
148
+ "eval_runtime": 106.4889,
149
+ "eval_samples_per_second": 379.664,
150
+ "eval_steps_per_second": 1.484,
151
+ "step": 11371
152
+ },
153
+ {
154
+ "epoch": 1.01,
155
+ "learning_rate": 1.3257702342215579e-05,
156
+ "loss": 0.6926,
157
+ "step": 11500
158
+ },
159
+ {
160
+ "epoch": 1.06,
161
+ "learning_rate": 1.2964558965790168e-05,
162
+ "loss": 0.656,
163
+ "step": 12000
164
+ },
165
+ {
166
+ "epoch": 1.1,
167
+ "learning_rate": 1.2671415589364759e-05,
168
+ "loss": 0.6554,
169
+ "step": 12500
170
+ },
171
+ {
172
+ "epoch": 1.14,
173
+ "learning_rate": 1.2378272212939351e-05,
174
+ "loss": 0.6457,
175
+ "step": 13000
176
+ },
177
+ {
178
+ "epoch": 1.19,
179
+ "learning_rate": 1.208512883651394e-05,
180
+ "loss": 0.6531,
181
+ "step": 13500
182
+ },
183
+ {
184
+ "epoch": 1.23,
185
+ "learning_rate": 1.1791985460088531e-05,
186
+ "loss": 0.6553,
187
+ "step": 14000
188
+ },
189
+ {
190
+ "epoch": 1.28,
191
+ "learning_rate": 1.149884208366312e-05,
192
+ "loss": 0.6447,
193
+ "step": 14500
194
+ },
195
+ {
196
+ "epoch": 1.32,
197
+ "learning_rate": 1.1205698707237711e-05,
198
+ "loss": 0.623,
199
+ "step": 15000
200
+ },
201
+ {
202
+ "epoch": 1.36,
203
+ "learning_rate": 1.0912555330812302e-05,
204
+ "loss": 0.64,
205
+ "step": 15500
206
+ },
207
+ {
208
+ "epoch": 1.41,
209
+ "learning_rate": 1.0619411954386891e-05,
210
+ "loss": 0.6371,
211
+ "step": 16000
212
+ },
213
+ {
214
+ "epoch": 1.45,
215
+ "learning_rate": 1.0326268577961482e-05,
216
+ "loss": 0.6519,
217
+ "step": 16500
218
+ },
219
+ {
220
+ "epoch": 1.5,
221
+ "learning_rate": 1.0033125201536071e-05,
222
+ "loss": 0.6368,
223
+ "step": 17000
224
+ },
225
+ {
226
+ "epoch": 1.54,
227
+ "learning_rate": 9.739981825110662e-06,
228
+ "loss": 0.6436,
229
+ "step": 17500
230
+ },
231
+ {
232
+ "epoch": 1.58,
233
+ "learning_rate": 9.446838448685253e-06,
234
+ "loss": 0.6382,
235
+ "step": 18000
236
+ },
237
+ {
238
+ "epoch": 1.63,
239
+ "learning_rate": 9.153695072259843e-06,
240
+ "loss": 0.625,
241
+ "step": 18500
242
+ },
243
+ {
244
+ "epoch": 1.67,
245
+ "learning_rate": 8.860551695834433e-06,
246
+ "loss": 0.638,
247
+ "step": 19000
248
+ },
249
+ {
250
+ "epoch": 1.71,
251
+ "learning_rate": 8.567408319409023e-06,
252
+ "loss": 0.6326,
253
+ "step": 19500
254
+ },
255
+ {
256
+ "epoch": 1.76,
257
+ "learning_rate": 8.274264942983614e-06,
258
+ "loss": 0.6194,
259
+ "step": 20000
260
+ },
261
+ {
262
+ "epoch": 1.8,
263
+ "learning_rate": 7.981121566558205e-06,
264
+ "loss": 0.624,
265
+ "step": 20500
266
+ },
267
+ {
268
+ "epoch": 1.85,
269
+ "learning_rate": 7.687978190132794e-06,
270
+ "loss": 0.635,
271
+ "step": 21000
272
+ },
273
+ {
274
+ "epoch": 1.89,
275
+ "learning_rate": 7.394834813707384e-06,
276
+ "loss": 0.629,
277
+ "step": 21500
278
+ },
279
+ {
280
+ "epoch": 1.93,
281
+ "learning_rate": 7.101691437281976e-06,
282
+ "loss": 0.6191,
283
+ "step": 22000
284
+ },
285
+ {
286
+ "epoch": 1.98,
287
+ "learning_rate": 6.808548060856566e-06,
288
+ "loss": 0.6142,
289
+ "step": 22500
290
+ },
291
+ {
292
+ "epoch": 2.0,
293
+ "eval_accuracy": 85.40934949295078,
294
+ "eval_combined_score": 83.03703568914804,
295
+ "eval_f1": 80.6647218853453,
296
+ "eval_loss": 0.32606932520866394,
297
+ "eval_runtime": 105.2719,
298
+ "eval_samples_per_second": 384.053,
299
+ "eval_steps_per_second": 1.501,
300
+ "step": 22742
301
+ },
302
+ {
303
+ "epoch": 2.02,
304
+ "learning_rate": 6.515404684431156e-06,
305
+ "loss": 0.591,
306
+ "step": 23000
307
+ },
308
+ {
309
+ "epoch": 2.07,
310
+ "learning_rate": 6.222261308005746e-06,
311
+ "loss": 0.5543,
312
+ "step": 23500
313
+ },
314
+ {
315
+ "epoch": 2.11,
316
+ "learning_rate": 5.929117931580336e-06,
317
+ "loss": 0.5572,
318
+ "step": 24000
319
+ },
320
+ {
321
+ "epoch": 2.15,
322
+ "learning_rate": 5.6359745551549274e-06,
323
+ "loss": 0.5657,
324
+ "step": 24500
325
+ },
326
+ {
327
+ "epoch": 2.2,
328
+ "learning_rate": 5.3428311787295174e-06,
329
+ "loss": 0.5572,
330
+ "step": 25000
331
+ },
332
+ {
333
+ "epoch": 2.24,
334
+ "learning_rate": 5.049687802304107e-06,
335
+ "loss": 0.5606,
336
+ "step": 25500
337
+ },
338
+ {
339
+ "epoch": 2.29,
340
+ "learning_rate": 4.756544425878697e-06,
341
+ "loss": 0.5592,
342
+ "step": 26000
343
+ },
344
+ {
345
+ "epoch": 2.33,
346
+ "learning_rate": 4.463401049453288e-06,
347
+ "loss": 0.564,
348
+ "step": 26500
349
+ },
350
+ {
351
+ "epoch": 2.37,
352
+ "learning_rate": 4.170257673027878e-06,
353
+ "loss": 0.5672,
354
+ "step": 27000
355
+ },
356
+ {
357
+ "epoch": 2.42,
358
+ "learning_rate": 3.877114296602469e-06,
359
+ "loss": 0.5639,
360
+ "step": 27500
361
+ },
362
+ {
363
+ "epoch": 2.46,
364
+ "learning_rate": 3.583970920177059e-06,
365
+ "loss": 0.5605,
366
+ "step": 28000
367
+ },
368
+ {
369
+ "epoch": 2.51,
370
+ "learning_rate": 3.290827543751649e-06,
371
+ "loss": 0.5448,
372
+ "step": 28500
373
+ },
374
+ {
375
+ "epoch": 2.55,
376
+ "learning_rate": 2.9976841673262398e-06,
377
+ "loss": 0.5588,
378
+ "step": 29000
379
+ },
380
+ {
381
+ "epoch": 2.59,
382
+ "learning_rate": 2.7045407909008298e-06,
383
+ "loss": 0.5416,
384
+ "step": 29500
385
+ },
386
+ {
387
+ "epoch": 2.64,
388
+ "learning_rate": 2.41139741447542e-06,
389
+ "loss": 0.5532,
390
+ "step": 30000
391
+ },
392
+ {
393
+ "epoch": 2.68,
394
+ "learning_rate": 2.1182540380500106e-06,
395
+ "loss": 0.5475,
396
+ "step": 30500
397
+ },
398
+ {
399
+ "epoch": 2.73,
400
+ "learning_rate": 1.825110661624601e-06,
401
+ "loss": 0.5412,
402
+ "step": 31000
403
+ },
404
+ {
405
+ "epoch": 2.77,
406
+ "learning_rate": 1.531967285199191e-06,
407
+ "loss": 0.554,
408
+ "step": 31500
409
+ },
410
+ {
411
+ "epoch": 2.81,
412
+ "learning_rate": 1.2388239087737813e-06,
413
+ "loss": 0.5532,
414
+ "step": 32000
415
+ },
416
+ {
417
+ "epoch": 2.86,
418
+ "learning_rate": 9.456805323483717e-07,
419
+ "loss": 0.5533,
420
+ "step": 32500
421
+ },
422
+ {
423
+ "epoch": 2.9,
424
+ "learning_rate": 6.525371559229619e-07,
425
+ "loss": 0.5351,
426
+ "step": 33000
427
+ },
428
+ {
429
+ "epoch": 2.95,
430
+ "learning_rate": 3.5939377949755233e-07,
431
+ "loss": 0.5447,
432
+ "step": 33500
433
+ },
434
+ {
435
+ "epoch": 2.99,
436
+ "learning_rate": 6.62504030721426e-08,
437
+ "loss": 0.5349,
438
+ "step": 34000
439
+ },
440
+ {
441
+ "epoch": 3.0,
442
+ "eval_accuracy": 86.12169181301013,
443
+ "eval_combined_score": 83.80513015974645,
444
+ "eval_f1": 81.48856850648279,
445
+ "eval_loss": 0.3177751898765564,
446
+ "eval_runtime": 105.2798,
447
+ "eval_samples_per_second": 384.024,
448
+ "eval_steps_per_second": 1.501,
449
+ "step": 34113
450
+ },
451
+ {
452
+ "epoch": 3.0,
453
+ "step": 34113,
454
+ "total_flos": 0.0,
455
+ "train_loss": 0.6687755357152263,
456
+ "train_runtime": 7270.4925,
457
+ "train_samples_per_second": 150.133,
458
+ "train_steps_per_second": 4.692
459
+ }
460
+ ],
461
+ "max_steps": 34113,
462
+ "num_train_epochs": 3,
463
+ "total_flos": 0.0,
464
+ "trial_name": null,
465
+ "trial_params": null
466
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:571423933137e161e1bf46056d30e0a33302368e398c441b6375c02b2784c456
3
+ size 4091
transfer_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eval_combined_score": 0.7922077922077924,
3
+ "eval_f1": 0.7922077922077924,
4
+ "eval_loss": 0.8416510224342346,
5
+ "eval_runtime": 2.1923,
6
+ "eval_samples_per_second": 186.11,
7
+ "eval_steps_per_second": 0.912
8
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff