andrew-healey commited on
Commit
688fff0
·
verified ·
1 Parent(s): 7cf62a7

Upload folder using huggingface_hub

Browse files
attention_kindselective_n_heads4_seed1340/args.json CHANGED
@@ -1 +1 @@
1
- {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_9/attention_kindselective_n_heads4_seed1340", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 4, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 8750, "warmup_steps": 500, "group": "wider_is_better_9", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1340, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 120, "total_batch_size": 61440, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 3e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "3e-5_61440_4_1340", "n_embd": 256}
 
1
+ {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_9/attention_kindselective_n_heads4_seed1340", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 4, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 8750, "warmup_steps": 500, "group": "wider_is_better_9", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1340, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 120, "total_batch_size": 61440, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 5.5e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "5.5e-5_61440_4_1340", "n_embd": 256}
attention_kindselective_n_heads4_seed1340/log2.txt CHANGED
@@ -1,533 +1,267 @@
1
  max_steps: 8750
2
  0 val loss 11.2709
3
- 0 val perplexity 78503.8281
4
- 0 val loss 11.2709
5
- 0 val perplexity 78503.8281
6
- 0 train 11.277061 (lr=7.0000e-08) (hash(x)=164406924)
7
- 0 train 11.277061 (lr=6.0000e-08) (hash(x)=164406924)
8
- 100 val loss 10.1502
9
- 100 val perplexity 25596.3652
10
- 100 train 10.113564 (lr=7.0700e-06) (hash(x)=144903932)
11
- 100 val loss 10.2327
12
- 100 val perplexity 27796.1465
13
- 100 train 10.198875 (lr=6.0600e-06) (hash(x)=144903932)
14
- 200 val loss 9.7291
15
- 200 val perplexity 16799.9590
16
- 200 train 9.783854 (lr=1.4070e-05) (hash(x)=167734596)
17
- 200 val loss 9.8080
18
- 200 val perplexity 18177.9453
19
- 200 train 9.861107 (lr=1.2060e-05) (hash(x)=167734596)
20
- 300 val loss 8.8589
21
- 300 val perplexity 7036.8594
22
- 300 train 8.904337 (lr=2.1070e-05) (hash(x)=149619098)
23
- 300 val loss 8.9841
24
- 300 val perplexity 7975.3730
25
- 300 train 9.031675 (lr=1.8060e-05) (hash(x)=149619098)
26
- 400 val loss 7.9793
27
- 400 val perplexity 2919.9771
28
- 400 train 7.923836 (lr=2.8070e-05) (hash(x)=153710890)
29
- 400 val loss 8.0370
30
- 400 val perplexity 3093.2683
31
- 400 train 7.987201 (lr=2.4060e-05) (hash(x)=153710890)
32
- 500 val loss 7.6324
33
- 500 val perplexity 2063.9663
34
- 500 train 7.716890 (lr=3.5000e-05) (hash(x)=145450636)
35
- 500 val loss 7.6279
36
- 500 val perplexity 2054.7944
37
- 500 train 7.708719 (lr=3.0000e-05) (hash(x)=145450636)
38
- 600 val loss 7.4815
39
- 600 val perplexity 1774.9750
40
- 600 train 7.341413 (lr=3.4989e-05) (hash(x)=145249251)
41
- 600 val loss 7.4696
42
- 600 val perplexity 1753.8359
43
- 600 train 7.332487 (lr=2.9990e-05) (hash(x)=145249251)
44
- 700 val loss 7.3877
45
- 700 val perplexity 1615.9683
46
- 700 train 7.501635 (lr=3.4954e-05) (hash(x)=150475545)
47
- 700 val loss 7.3795
48
- 700 val perplexity 1602.7659
49
- 700 train 7.492246 (lr=2.9961e-05) (hash(x)=150475545)
50
- 800 val loss 7.3168
51
- 800 val perplexity 1505.4194
52
- 800 train 7.283746 (lr=3.4897e-05) (hash(x)=144483776)
53
- 800 val loss 7.3040
54
- 800 val perplexity 1486.2216
55
- 800 train 7.270497 (lr=2.9912e-05) (hash(x)=144483776)
56
- 900 val loss 7.2365
57
- 900 val perplexity 1389.2202
58
- 900 train 7.342116 (lr=3.4818e-05) (hash(x)=157916369)
59
- 900 val loss 7.2214
60
- 900 val perplexity 1368.4169
61
- 900 train 7.322839 (lr=2.9844e-05) (hash(x)=157916369)
62
- 1000 val loss 7.1684
63
- 1000 val perplexity 1297.7793
64
- 1000 train 7.195076 (lr=3.4715e-05) (hash(x)=154856891)
65
- 1000 val loss 7.1551
66
- 1000 val perplexity 1280.6429
67
- 1000 train 7.181716 (lr=2.9756e-05) (hash(x)=154856891)
68
- 1100 val loss 7.1025
69
- 1100 val perplexity 1215.0176
70
- 1100 train 6.832980 (lr=3.4591e-05) (hash(x)=136191502)
71
- 1100 val loss 7.0845
72
- 1100 val perplexity 1193.2688
73
- 1100 train 6.817375 (lr=2.9649e-05) (hash(x)=136191502)
74
- 1200 val loss 7.0249
75
- 1200 val perplexity 1124.2922
76
- 1200 train 6.977524 (lr=3.4444e-05) (hash(x)=148021541)
77
- 1200 val loss 7.0108
78
- 1200 val perplexity 1108.5863
79
- 1200 train 6.962654 (lr=2.9523e-05) (hash(x)=148021541)
80
- 1300 val loss 6.9649
81
- 1300 val perplexity 1058.8282
82
- 1300 train 6.899820 (lr=3.4275e-05) (hash(x)=146005217)
83
- 1300 val loss 6.9486
84
- 1300 val perplexity 1041.7388
85
- 1300 train 6.883662 (lr=2.9378e-05) (hash(x)=146005217)
86
- 1400 val loss 6.9124
87
- 1400 val perplexity 1004.6208
88
- 1400 train 6.817109 (lr=3.4084e-05) (hash(x)=146019502)
89
- 1400 val loss 6.8887
90
- 1400 val perplexity 981.1172
91
- 1400 train 6.794360 (lr=2.9215e-05) (hash(x)=146019502)
92
- 1500 val loss 6.8527
93
- 1500 val perplexity 946.4700
94
- 1500 train 6.795928 (lr=3.3872e-05) (hash(x)=150127281)
95
- 1500 val loss 6.8294
96
- 1500 val perplexity 924.6553
97
- 1500 train 6.769631 (lr=2.9033e-05) (hash(x)=150127281)
98
- 1600 val loss 6.7911
99
- 1600 val perplexity 889.9340
100
- 1600 train 6.759404 (lr=3.3638e-05) (hash(x)=154120875)
101
- 1600 val loss 6.7634
102
- 1600 val perplexity 865.6116
103
- 1600 train 6.731661 (lr=2.8833e-05) (hash(x)=154120875)
104
- 1700 val loss 6.7391
105
- 1700 val perplexity 844.7706
106
- 1700 train 6.959402 (lr=3.3384e-05) (hash(x)=155797680)
107
- 1700 val loss 6.7154
108
- 1700 val perplexity 825.0399
109
- 1700 train 6.934373 (lr=2.8615e-05) (hash(x)=155797680)
110
- 1800 val loss 6.6668
111
- 1800 val perplexity 785.8712
112
- 1800 train 6.665632 (lr=3.3109e-05) (hash(x)=156809396)
113
- 1800 val loss 6.6459
114
- 1800 val perplexity 769.6384
115
- 1800 train 6.640855 (lr=2.8379e-05) (hash(x)=156809396)
116
- 1900 val loss 6.6110
117
- 1900 val perplexity 743.2146
118
- 1900 train 6.502274 (lr=3.2814e-05) (hash(x)=144640294)
119
- 1900 val loss 6.5938
120
- 1900 val perplexity 730.5507
121
- 1900 train 6.482723 (lr=2.8127e-05) (hash(x)=144640294)
122
- 2000 val loss 6.5645
123
- 2000 val perplexity 709.4272
124
- 2000 train 6.437298 (lr=3.2500e-05) (hash(x)=162831106)
125
- 2000 val loss 6.5512
126
- 2000 val perplexity 700.1097
127
- 2000 train 6.425098 (lr=2.7857e-05) (hash(x)=162831106)
128
- 2100 val loss 6.5289
129
- 2100 val perplexity 684.6605
130
- 2100 train 6.518277 (lr=3.2166e-05) (hash(x)=158239484)
131
- 2100 val loss 6.5258
132
- 2100 val perplexity 682.5104
133
- 2100 train 6.522050 (lr=2.7571e-05) (hash(x)=158239484)
134
- 2200 val loss 6.4757
135
- 2200 val perplexity 649.1948
136
- 2200 train 6.351269 (lr=3.1813e-05) (hash(x)=140504180)
137
- 2200 val loss 6.4764
138
- 2200 val perplexity 649.6376
139
- 2200 train 6.350248 (lr=2.7269e-05) (hash(x)=140504180)
140
- 2300 val loss 6.4330
141
- 2300 val perplexity 622.0456
142
- 2300 train 6.415581 (lr=3.1443e-05) (hash(x)=142234024)
143
- 2300 val loss 6.4349
144
- 2300 val perplexity 623.2421
145
- 2300 train 6.418139 (lr=2.6951e-05) (hash(x)=142234024)
146
- 2400 val loss 6.4024
147
- 2400 val perplexity 603.2999
148
- 2400 train 6.397040 (lr=3.1054e-05) (hash(x)=143091562)
149
- 2400 val loss 6.4081
150
- 2400 val perplexity 606.7314
151
- 2400 train 6.401471 (lr=2.6618e-05) (hash(x)=143091562)
152
- 2500 val loss 6.3863
153
- 2500 val perplexity 593.6433
154
- 2500 train 6.290105 (lr=3.0649e-05) (hash(x)=149857456)
155
- 2500 val loss 6.3969
156
- 2500 val perplexity 599.9606
157
- 2500 train 6.296983 (lr=2.6270e-05) (hash(x)=149857456)
158
- 2600 val loss 6.3419
159
- 2600 val perplexity 567.8901
160
- 2600 train 6.379487 (lr=3.0227e-05) (hash(x)=146191551)
161
- 2600 val loss 6.3522
162
- 2600 val perplexity 573.7252
163
- 2600 train 6.395041 (lr=2.5909e-05) (hash(x)=146191551)
164
- 2700 val loss 6.3120
165
- 2700 val perplexity 551.1568
166
- 2700 train 6.235927 (lr=2.9789e-05) (hash(x)=145375752)
167
- 2700 val loss 6.3210
168
- 2700 val perplexity 556.1021
169
- 2700 train 6.239472 (lr=2.5533e-05) (hash(x)=145375752)
170
- 2800 val loss 6.2871
171
- 2800 val perplexity 537.5940
172
- 2800 train 6.166823 (lr=2.9336e-05) (hash(x)=151568014)
173
- 2800 val loss 6.3012
174
- 2800 val perplexity 545.2515
175
- 2800 train 6.188046 (lr=2.5145e-05) (hash(x)=151568014)
176
- 2900 val loss 6.2707
177
- 2900 val perplexity 528.8716
178
- 2900 train 6.104226 (lr=2.8868e-05) (hash(x)=149366597)
179
- 2900 val loss 6.2892
180
- 2900 val perplexity 538.7473
181
- 2900 train 6.116791 (lr=2.4744e-05) (hash(x)=149366597)
182
- 3000 val loss 6.2343
183
- 3000 val perplexity 509.9225
184
- 3000 train 6.170280 (lr=2.8386e-05) (hash(x)=150464442)
185
- 3000 val loss 6.2545
186
- 3000 val perplexity 520.3714
187
- 3000 train 6.192822 (lr=2.4331e-05) (hash(x)=150464442)
188
- 3100 val loss 6.2172
189
- 3100 val perplexity 501.3207
190
- 3100 train 6.346448 (lr=2.7891e-05) (hash(x)=182449036)
191
- 3100 val loss 6.2397
192
- 3100 val perplexity 512.7269
193
- 3100 train 6.366556 (lr=2.3906e-05) (hash(x)=182449036)
194
- 3200 val loss 6.1963
195
- 3200 val perplexity 490.9159
196
- 3200 train 6.149408 (lr=2.7383e-05) (hash(x)=140141286)
197
- 3200 val loss 6.2185
198
- 3200 val perplexity 501.9713
199
- 3200 train 6.174580 (lr=2.3471e-05) (hash(x)=140141286)
200
- 3300 val loss 6.1753
201
- 3300 val perplexity 480.7094
202
- 3300 train 6.136339 (lr=2.6864e-05) (hash(x)=148099414)
203
- 3300 val loss 6.1991
204
- 3300 val perplexity 492.3220
205
- 3300 train 6.160706 (lr=2.3026e-05) (hash(x)=148099414)
206
- 3400 val loss 6.1553
207
- 3400 val perplexity 471.2260
208
- 3400 train 6.015404 (lr=2.6333e-05) (hash(x)=142633951)
209
- 3400 val loss 6.1802
210
- 3400 val perplexity 483.0799
211
- 3400 train 6.039740 (lr=2.2572e-05) (hash(x)=142633951)
212
- 3500 val loss 6.1481
213
- 3500 val perplexity 467.8076
214
- 3500 train 6.004545 (lr=2.5793e-05) (hash(x)=148368965)
215
- 3500 val loss 6.1706
216
- 3500 val perplexity 478.4758
217
- 3500 train 6.029538 (lr=2.2108e-05) (hash(x)=148368965)
218
- 3600 val loss 6.1221
219
- 3600 val perplexity 455.8235
220
- 3600 train 6.150208 (lr=2.5243e-05) (hash(x)=152372067)
221
- 3600 val loss 6.1490
222
- 3600 val perplexity 468.2530
223
- 3600 train 6.173963 (lr=2.1637e-05) (hash(x)=152372067)
224
- 3700 val loss 6.1051
225
- 3700 val perplexity 448.1516
226
- 3700 train 6.197345 (lr=2.4684e-05) (hash(x)=168885609)
227
- 3700 val loss 6.1346
228
- 3700 val perplexity 461.5741
229
- 3700 train 6.226795 (lr=2.1158e-05) (hash(x)=168885609)
230
- 3800 val loss 6.0966
231
- 3800 val perplexity 444.3637
232
- 3800 train 5.907861 (lr=2.4117e-05) (hash(x)=141633734)
233
- 3800 val loss 6.1256
234
- 3800 val perplexity 457.4419
235
- 3800 train 5.937433 (lr=2.0672e-05) (hash(x)=141633734)
236
- 3900 val loss 6.0723
237
- 3900 val perplexity 433.6902
238
- 3900 train 5.966411 (lr=2.3543e-05) (hash(x)=153141007)
239
- 3900 val loss 6.1031
240
- 3900 val perplexity 447.2220
241
- 3900 train 5.999508 (lr=2.0180e-05) (hash(x)=153141007)
242
- 4000 val loss 6.0612
243
- 4000 val perplexity 428.9083
244
- 4000 train 5.994430 (lr=2.2963e-05) (hash(x)=160577202)
245
- 4000 val loss 6.0919
246
- 4000 val perplexity 442.2613
247
- 4000 train 6.026505 (lr=1.9683e-05) (hash(x)=160577202)
248
- 4100 val loss 6.0522
249
- 4100 val perplexity 425.0566
250
- 4100 train 6.113316 (lr=2.2378e-05) (hash(x)=153858169)
251
- 4100 val loss 6.0831
252
- 4100 val perplexity 438.4046
253
- 4100 train 6.139560 (lr=1.9181e-05) (hash(x)=153858169)
254
- 4200 val loss 6.0324
255
- 4200 val perplexity 416.7092
256
- 4200 train 6.035808 (lr=2.1788e-05) (hash(x)=155889149)
257
- 4200 val loss 6.0647
258
- 4200 val perplexity 430.4065
259
- 4200 train 6.064448 (lr=1.8675e-05) (hash(x)=155889149)
260
- 4300 val loss 6.0199
261
- 4300 val perplexity 411.5509
262
- 4300 train 5.834055 (lr=2.1194e-05) (hash(x)=152294662)
263
- 4300 val loss 6.0560
264
- 4300 val perplexity 426.6565
265
- 4300 train 5.866488 (lr=1.8166e-05) (hash(x)=152294662)
266
- 4400 val loss 6.0108
267
- 4400 val perplexity 407.8034
268
- 4400 train 5.899076 (lr=2.0598e-05) (hash(x)=141804386)
269
- 4400 val loss 6.0450
270
- 4400 val perplexity 421.9777
271
- 4400 train 5.932065 (lr=1.7655e-05) (hash(x)=141804386)
272
- 4500 val loss 5.9941
273
- 4500 val perplexity 401.0631
274
- 4500 train 5.893041 (lr=1.9999e-05) (hash(x)=151095242)
275
- 4500 val loss 6.0274
276
- 4500 val perplexity 414.6237
277
- 4500 train 5.928043 (lr=1.7142e-05) (hash(x)=151095242)
278
- 4600 val loss 5.9986
279
- 4600 val perplexity 402.8594
280
- 4600 train 5.935762 (lr=1.9400e-05) (hash(x)=156414699)
281
- 4600 val loss 6.0307
282
- 4600 val perplexity 416.0135
283
- 4600 train 5.966719 (lr=1.6629e-05) (hash(x)=156414699)
284
- 4700 val loss 5.9728
285
- 4700 val perplexity 392.6121
286
- 4700 train 5.997845 (lr=1.8800e-05) (hash(x)=161556686)
287
- 4700 val loss 6.0094
288
- 4700 val perplexity 407.2528
289
- 4700 train 6.033873 (lr=1.6114e-05) (hash(x)=161556686)
290
- 4800 val loss 5.9603
291
- 4800 val perplexity 387.7329
292
- 4800 train 5.870970 (lr=1.8201e-05) (hash(x)=149000293)
293
- 4800 val loss 5.9985
294
- 4800 val perplexity 402.8394
295
- 4800 train 5.906920 (lr=1.5601e-05) (hash(x)=149000293)
296
- 4900 val loss 5.9508
297
- 4900 val perplexity 384.0769
298
- 4900 train 5.752374 (lr=1.7604e-05) (hash(x)=154349989)
299
- 4900 val loss 5.9899
300
- 4900 val perplexity 399.3921
301
- 4900 train 5.792023 (lr=1.5089e-05) (hash(x)=154349989)
302
- 5000 val loss 5.9401
303
- 5000 val perplexity 379.9893
304
- 5000 train 5.769830 (lr=1.7009e-05) (hash(x)=131475967)
305
- 5000 val loss 5.9785
306
- 5000 val perplexity 394.8312
307
- 5000 train 5.818519 (lr=1.4579e-05) (hash(x)=131475967)
308
- 5100 val loss 5.9320
309
- 5100 val perplexity 376.9128
310
- 5100 train 5.830674 (lr=1.6417e-05) (hash(x)=149717902)
311
- 5100 val loss 5.9694
312
- 5100 val perplexity 391.2876
313
- 5100 train 5.866533 (lr=1.4071e-05) (hash(x)=149717902)
314
- 5200 val loss 5.9241
315
- 5200 val perplexity 373.9456
316
- 5200 train 5.780983 (lr=1.5829e-05) (hash(x)=151407999)
317
- 5200 val loss 5.9621
318
- 5200 val perplexity 388.4129
319
- 5200 train 5.815117 (lr=1.3568e-05) (hash(x)=151407999)
320
- 5300 val loss 5.9147
321
- 5300 val perplexity 370.4412
322
- 5300 train 6.021460 (lr=1.5246e-05) (hash(x)=168602728)
323
- 5300 val loss 5.9521
324
- 5300 val perplexity 384.5754
325
- 5300 train 6.062849 (lr=1.3068e-05) (hash(x)=168602728)
326
- 5400 val loss 5.9028
327
- 5400 val perplexity 366.0675
328
- 5400 train 5.860891 (lr=1.4669e-05) (hash(x)=158344511)
329
- 5400 val loss 5.9456
330
- 5400 val perplexity 382.0638
331
- 5400 train 5.902995 (lr=1.2573e-05) (hash(x)=158344511)
332
- 5500 val loss 5.8960
333
- 5500 val perplexity 363.5726
334
- 5500 train 5.857149 (lr=1.4099e-05) (hash(x)=148350057)
335
- 5500 val loss 5.9365
336
- 5500 val perplexity 378.6254
337
- 5500 train 5.899691 (lr=1.2085e-05) (hash(x)=148350057)
338
- 5600 val loss 5.8877
339
- 5600 val perplexity 360.5695
340
- 5600 train 5.956222 (lr=1.3536e-05) (hash(x)=153847323)
341
- 5600 val loss 5.9293
342
- 5600 val perplexity 375.8728
343
- 5600 train 5.996018 (lr=1.1602e-05) (hash(x)=153847323)
344
- 5700 val loss 5.8798
345
- 5700 val perplexity 357.7410
346
- 5700 train 5.829589 (lr=1.2981e-05) (hash(x)=156607405)
347
- 5700 val loss 5.9220
348
- 5700 val perplexity 373.1628
349
- 5700 train 5.870955 (lr=1.1127e-05) (hash(x)=156607405)
350
- 5800 val loss 5.8736
351
- 5800 val perplexity 355.5310
352
- 5800 train 5.641397 (lr=1.2436e-05) (hash(x)=145115031)
353
- 5800 val loss 5.9151
354
- 5800 val perplexity 370.6071
355
- 5800 train 5.676248 (lr=1.0659e-05) (hash(x)=145115031)
356
- 5900 val loss 5.8659
357
- 5900 val perplexity 352.7947
358
- 5900 train 5.737990 (lr=1.1900e-05) (hash(x)=141584622)
359
- 5900 val loss 5.9075
360
- 5900 val perplexity 367.7683
361
- 5900 train 5.776842 (lr=1.0200e-05) (hash(x)=141584622)
362
- 6000 val loss 5.8580
363
- 6000 val perplexity 350.0075
364
- 6000 train 5.787103 (lr=1.1375e-05) (hash(x)=146613857)
365
- 6000 val loss 5.9002
366
- 6000 val perplexity 365.0927
367
- 6000 train 5.830695 (lr=9.7500e-06) (hash(x)=146613857)
368
- 6100 val loss 5.8535
369
- 6100 val perplexity 348.4407
370
- 6100 train 5.786377 (lr=1.0861e-05) (hash(x)=144621768)
371
- 6100 val loss 5.8963
372
- 6100 val perplexity 363.7028
373
- 6100 train 5.814258 (lr=9.3098e-06) (hash(x)=144621768)
374
- 6200 val loss 5.8504
375
- 6200 val perplexity 347.3798
376
- 6200 train 5.775459 (lr=1.0360e-05) (hash(x)=146521760)
377
- 6200 val loss 5.8932
378
- 6200 val perplexity 362.5698
379
- 6200 train 5.820045 (lr=8.8800e-06) (hash(x)=146521760)
380
- 6300 val loss 5.8390
381
- 6300 val perplexity 343.4239
382
- 6300 train 5.671170 (lr=9.8715e-06) (hash(x)=161378136)
383
- 6300 val loss 5.8837
384
- 6300 val perplexity 359.1181
385
- 6300 train 5.718021 (lr=8.4613e-06) (hash(x)=161378136)
386
- 6400 val loss 5.8362
387
- 6400 val perplexity 342.4716
388
- 6400 train 5.602135 (lr=9.3966e-06) (hash(x)=141624235)
389
- 6400 val loss 5.8797
390
- 6400 val perplexity 357.7079
391
- 6400 train 5.647285 (lr=8.0542e-06) (hash(x)=141624235)
392
- 6500 val loss 5.8298
393
- 6500 val perplexity 340.3030
394
- 6500 train 5.743033 (lr=8.9359e-06) (hash(x)=151197095)
395
- 6500 val loss 5.8733
396
- 6500 val perplexity 355.4235
397
- 6500 train 5.777500 (lr=7.6594e-06) (hash(x)=151197095)
398
- 6600 val loss 5.8275
399
- 6600 val perplexity 339.4921
400
- 6600 train 5.840244 (lr=8.4903e-06) (hash(x)=153269571)
401
- 6600 val loss 5.8719
402
- 6600 val perplexity 354.9266
403
- 6600 train 5.881926 (lr=7.2774e-06) (hash(x)=153269571)
404
- 6700 val loss 5.8217
405
- 6700 val perplexity 337.5518
406
- 6700 train 5.783973 (lr=8.0602e-06) (hash(x)=146111181)
407
- 6700 val loss 5.8653
408
- 6700 val perplexity 352.5977
409
- 6700 train 5.830712 (lr=6.9087e-06) (hash(x)=146111181)
410
- 6800 val loss 5.8137
411
- 6800 val perplexity 334.8429
412
- 6800 train 5.840888 (lr=7.6463e-06) (hash(x)=147269760)
413
- 6800 val loss 5.8585
414
- 6800 val perplexity 350.1996
415
- 6800 train 5.885352 (lr=6.5540e-06) (hash(x)=147269760)
416
- 6900 val loss 5.8129
417
- 6900 val perplexity 334.5923
418
- 6900 train 5.878181 (lr=7.2493e-06) (hash(x)=152912762)
419
- 6900 val loss 5.8558
420
- 6900 val perplexity 349.2571
421
- 6900 train 5.918777 (lr=6.2137e-06) (hash(x)=152912762)
422
- 7000 val loss 5.8089
423
- 7000 val perplexity 333.2421
424
- 7000 train 5.886887 (lr=6.8697e-06) (hash(x)=165412343)
425
- 7000 val loss 5.8537
426
- 7000 val perplexity 348.5204
427
- 7000 train 5.931127 (lr=5.8883e-06) (hash(x)=165412343)
428
- 7100 val loss 5.8002
429
- 7100 val perplexity 330.3812
430
- 7100 train 5.863431 (lr=6.5080e-06) (hash(x)=162866028)
431
- 7100 val loss 5.8450
432
- 7100 val perplexity 345.4860
433
- 7100 train 5.909845 (lr=5.5783e-06) (hash(x)=162866028)
434
- 7200 val loss 5.7969
435
- 7200 val perplexity 329.2704
436
- 7200 train 5.771354 (lr=6.1648e-06) (hash(x)=142998115)
437
- 7200 val loss 5.8425
438
- 7200 val perplexity 344.6447
439
- 7200 train 5.816433 (lr=5.2841e-06) (hash(x)=142998115)
440
- 7300 val loss 5.7948
441
- 7300 val perplexity 328.5770
442
- 7300 train 5.564707 (lr=5.8405e-06) (hash(x)=145486999)
443
- 7300 val loss 5.8414
444
- 7300 val perplexity 344.2689
445
- 7300 train 5.614068 (lr=5.0062e-06) (hash(x)=145486999)
446
- 7400 val loss 5.7891
447
- 7400 val perplexity 326.7220
448
- 7400 train 5.758606 (lr=5.5357e-06) (hash(x)=155325873)
449
- 7400 val loss 5.8346
450
- 7400 val perplexity 341.9388
451
- 7400 train 5.804873 (lr=4.7449e-06) (hash(x)=155325873)
452
- 7500 val loss 5.7857
453
- 7500 val perplexity 325.5937
454
- 7500 train 5.746189 (lr=5.2508e-06) (hash(x)=145131256)
455
- 7500 val loss 5.8316
456
- 7500 val perplexity 340.9057
457
- 7500 train 5.791263 (lr=4.5007e-06) (hash(x)=145131256)
458
- 7600 val loss 5.7893
459
- 7600 val perplexity 326.7786
460
- 7600 train 5.710508 (lr=4.9862e-06) (hash(x)=144008365)
461
- 7600 val loss 5.8346
462
- 7600 val perplexity 341.9225
463
- 7600 train 5.753230 (lr=4.2739e-06) (hash(x)=144008365)
464
- 7700 val loss 5.7797
465
- 7700 val perplexity 323.6634
466
- 7700 train 5.858647 (lr=4.7423e-06) (hash(x)=148848532)
467
- 7700 val loss 5.8260
468
- 7700 val perplexity 338.9968
469
- 7700 train 5.898289 (lr=4.0648e-06) (hash(x)=148848532)
470
- 7800 val loss 5.7777
471
- 7800 val perplexity 323.0294
472
- 7800 train 5.758301 (lr=4.5194e-06) (hash(x)=150391642)
473
- 7800 val loss 5.8235
474
- 7800 val perplexity 338.1660
475
- 7800 train 5.806548 (lr=3.8738e-06) (hash(x)=150391642)
476
- 7900 val loss 5.7780
477
- 7900 val perplexity 323.1080
478
- 7900 train 5.534154 (lr=4.3179e-06) (hash(x)=152191414)
479
- 7900 val loss 5.8236
480
- 7900 val perplexity 338.1740
481
- 7900 train 5.581667 (lr=3.7010e-06) (hash(x)=152191414)
482
- 8000 val loss 5.7726
483
- 8000 val perplexity 321.3700
484
- 8000 train 5.851005 (lr=4.1380e-06) (hash(x)=159755587)
485
- 8000 val loss 5.8195
486
- 8000 val perplexity 336.8142
487
- 8000 train 5.901532 (lr=3.5468e-06) (hash(x)=159755587)
488
- 8100 val loss 5.7703
489
- 8100 val perplexity 320.6418
490
- 8100 train 5.791206 (lr=3.9800e-06) (hash(x)=156664468)
491
- 8100 val loss 5.8174
492
- 8100 val perplexity 336.0982
493
- 8100 train 5.838668 (lr=3.4114e-06) (hash(x)=156664468)
494
- 8200 val loss 5.7692
495
- 8200 val perplexity 320.2684
496
- 8200 val loss 5.8163
497
- 8200 val perplexity 335.7272
498
- 8200 train 5.611233 (lr=3.8442e-06) (hash(x)=139457379)
499
- 8200 train 5.657933 (lr=3.2950e-06) (hash(x)=139457379)
500
- 8300 val loss 5.7722
501
- 8300 val perplexity 321.2451
502
- 8300 val loss 5.8194
503
- 8300 val perplexity 336.7803
504
- 8300 train 5.541981 (lr=3.7307e-06) (hash(x)=145478564)
505
- 8300 train 5.593337 (lr=3.1977e-06) (hash(x)=145478564)
506
- 8400 val loss 5.7649
507
- 8400 val perplexity 318.9180
508
- 8400 val loss 5.8112
509
- 8400 val perplexity 334.0220
510
- 8400 train 5.704823 (lr=3.6397e-06) (hash(x)=154982769)
511
- 8400 train 5.757417 (lr=3.1197e-06) (hash(x)=154982769)
512
- 8500 val loss 5.7627
513
- 8500 val perplexity 318.2200
514
- 8500 val loss 5.8100
515
- 8500 val perplexity 333.6224
516
- 8500 train 5.712815 (lr=3.5713e-06) (hash(x)=145798118)
517
- 8500 train 5.755826 (lr=3.0611e-06) (hash(x)=145798118)
518
- 8600 val loss 5.7622
519
- 8600 val perplexity 318.0369
520
- 8600 val loss 5.8088
521
- 8600 val perplexity 333.2224
522
- 8600 train 5.668952 (lr=3.5257e-06) (hash(x)=143231551)
523
- 8600 train 5.717763 (lr=3.0220e-06) (hash(x)=143231551)
524
- 8700 val loss 5.7658
525
- 8700 val perplexity 319.1926
526
- 8700 val loss 5.8134
527
- 8700 val perplexity 334.7692
528
- 8700 train 5.667092 (lr=3.5029e-06) (hash(x)=154780112)
529
- 8700 train 5.715662 (lr=3.0024e-06) (hash(x)=154780112)
530
- 8749 val loss 5.7583
531
- 8749 val perplexity 316.8212
532
- 8749 val loss 5.8060
533
- 8749 val perplexity 332.2806
 
1
  max_steps: 8750
2
  0 val loss 11.2709
3
+ 0 val perplexity 78502.8516
4
+ 0 train 11.277068 (lr=1.1000e-07) (hash(x)=164406924)
5
+ 100 val loss 10.0085
6
+ 100 val perplexity 22213.9395
7
+ 100 train 9.970803 (lr=1.1110e-05) (hash(x)=144903932)
8
+ 200 val loss 9.2914
9
+ 200 val perplexity 10844.2412
10
+ 200 train 9.357403 (lr=2.2110e-05) (hash(x)=167734596)
11
+ 300 val loss 8.0630
12
+ 300 val perplexity 3174.7661
13
+ 300 train 8.075590 (lr=3.3110e-05) (hash(x)=149619098)
14
+ 400 val loss 7.5692
15
+ 400 val perplexity 1937.4963
16
+ 400 train 7.450818 (lr=4.4110e-05) (hash(x)=153710890)
17
+ 500 val loss 7.4143
18
+ 500 val perplexity 1659.5359
19
+ 500 train 7.504872 (lr=5.5000e-05) (hash(x)=145450636)
20
+ 600 val loss 7.2988
21
+ 600 val perplexity 1478.4788
22
+ 600 train 7.166701 (lr=5.4982e-05) (hash(x)=145249251)
23
+ 700 val loss 7.1882
24
+ 700 val perplexity 1323.7411
25
+ 700 train 7.318045 (lr=5.4928e-05) (hash(x)=150475545)
26
+ 800 val loss 7.0767
27
+ 800 val perplexity 1184.1125
28
+ 800 train 7.045276 (lr=5.4839e-05) (hash(x)=144483776)
29
+ 900 val loss 6.9781
30
+ 900 val perplexity 1072.9233
31
+ 900 train 7.080307 (lr=5.4713e-05) (hash(x)=157916369)
32
+ 1000 val loss 6.9082
33
+ 1000 val perplexity 1000.4556
34
+ 1000 train 6.925345 (lr=5.4553e-05) (hash(x)=154856891)
35
+ 1100 val loss 6.8256
36
+ 1100 val perplexity 921.1058
37
+ 1100 train 6.522852 (lr=5.4357e-05) (hash(x)=136191502)
38
+ 1200 val loss 6.7440
39
+ 1200 val perplexity 848.9789
40
+ 1200 train 6.700300 (lr=5.4126e-05) (hash(x)=148021541)
41
+ 1300 val loss 6.6680
42
+ 1300 val perplexity 786.7943
43
+ 1300 train 6.599460 (lr=5.3860e-05) (hash(x)=146005217)
44
+ 1400 val loss 6.6030
45
+ 1400 val perplexity 737.3328
46
+ 1400 train 6.514089 (lr=5.3561e-05) (hash(x)=146019502)
47
+ 1500 val loss 6.5557
48
+ 1500 val perplexity 703.2271
49
+ 1500 train 6.492791 (lr=5.3227e-05) (hash(x)=150127281)
50
+ 1600 val loss 6.5065
51
+ 1600 val perplexity 669.4506
52
+ 1600 train 6.472455 (lr=5.2860e-05) (hash(x)=154120875)
53
+ 1700 val loss 6.4733
54
+ 1700 val perplexity 647.6284
55
+ 1700 train 6.692543 (lr=5.2461e-05) (hash(x)=155797680)
56
+ 1800 val loss 6.4172
57
+ 1800 val perplexity 612.2867
58
+ 1800 train 6.403757 (lr=5.2029e-05) (hash(x)=156809396)
59
+ 1900 val loss 6.3877
60
+ 1900 val perplexity 594.5070
61
+ 1900 train 6.274245 (lr=5.1565e-05) (hash(x)=144640294)
62
+ 2000 val loss 6.3473
63
+ 2000 val perplexity 570.9570
64
+ 2000 train 6.208043 (lr=5.1071e-05) (hash(x)=162831106)
65
+ 2100 val loss 6.3462
66
+ 2100 val perplexity 570.3412
67
+ 2100 train 6.336015 (lr=5.0547e-05) (hash(x)=158239484)
68
+ 2200 val loss 6.2871
69
+ 2200 val perplexity 537.5853
70
+ 2200 train 6.161128 (lr=4.9993e-05) (hash(x)=140504180)
71
+ 2300 val loss 6.2516
72
+ 2300 val perplexity 518.8192
73
+ 2300 train 6.238686 (lr=4.9410e-05) (hash(x)=142234024)
74
+ 2400 val loss 6.2270
75
+ 2400 val perplexity 506.2221
76
+ 2400 train 6.225477 (lr=4.8800e-05) (hash(x)=143091562)
77
+ 2500 val loss 6.2183
78
+ 2500 val perplexity 501.8651
79
+ 2500 train 6.118163 (lr=4.8162e-05) (hash(x)=149857456)
80
+ 2600 val loss 6.1767
81
+ 2600 val perplexity 481.3983
82
+ 2600 train 6.211945 (lr=4.7499e-05) (hash(x)=146191551)
83
+ 2700 val loss 6.1543
84
+ 2700 val perplexity 470.7182
85
+ 2700 train 6.071248 (lr=4.6811e-05) (hash(x)=145375752)
86
+ 2800 val loss 6.1304
87
+ 2800 val perplexity 459.5979
88
+ 2800 train 6.014522 (lr=4.6099e-05) (hash(x)=151568014)
89
+ 2900 val loss 6.1218
90
+ 2900 val perplexity 455.7011
91
+ 2900 train 5.945108 (lr=4.5364e-05) (hash(x)=149366597)
92
+ 3000 val loss 6.0883
93
+ 3000 val perplexity 440.6820
94
+ 3000 train 6.013472 (lr=4.4606e-05) (hash(x)=150464442)
95
+ 3100 val loss 6.0724
96
+ 3100 val perplexity 433.7043
97
+ 3100 train 6.195862 (lr=4.3828e-05) (hash(x)=182449036)
98
+ 3200 val loss 6.0531
99
+ 3200 val perplexity 425.4502
100
+ 3200 train 6.008684 (lr=4.3031e-05) (hash(x)=140141286)
101
+ 3300 val loss 6.0351
102
+ 3300 val perplexity 417.8260
103
+ 3300 train 6.008709 (lr=4.2215e-05) (hash(x)=148099414)
104
+ 3400 val loss 6.0145
105
+ 3400 val perplexity 409.3233
106
+ 3400 train 5.879248 (lr=4.1381e-05) (hash(x)=142633951)
107
+ 3500 val loss 6.0038
108
+ 3500 val perplexity 404.9803
109
+ 3500 train 5.866429 (lr=4.0532e-05) (hash(x)=148368965)
110
+ 3600 val loss 5.9817
111
+ 3600 val perplexity 396.1001
112
+ 3600 train 6.005436 (lr=3.9667e-05) (hash(x)=152372067)
113
+ 3700 val loss 5.9637
114
+ 3700 val perplexity 389.0613
115
+ 3700 train 6.042872 (lr=3.8789e-05) (hash(x)=168885609)
116
+ 3800 val loss 5.9535
117
+ 3800 val perplexity 385.1143
118
+ 3800 train 5.766492 (lr=3.7898e-05) (hash(x)=141633734)
119
+ 3900 val loss 5.9283
120
+ 3900 val perplexity 375.5027
121
+ 3900 train 5.815371 (lr=3.6996e-05) (hash(x)=153141007)
122
+ 4000 val loss 5.9161
123
+ 4000 val perplexity 370.9533
124
+ 4000 train 5.847968 (lr=3.6085e-05) (hash(x)=160577202)
125
+ 4100 val loss 5.9108
126
+ 4100 val perplexity 368.9956
127
+ 4100 train 5.976177 (lr=3.5165e-05) (hash(x)=153858169)
128
+ 4200 val loss 5.8859
129
+ 4200 val perplexity 359.9236
130
+ 4200 train 5.900618 (lr=3.4238e-05) (hash(x)=155889149)
131
+ 4300 val loss 5.8756
132
+ 4300 val perplexity 356.2298
133
+ 4300 train 5.686732 (lr=3.3305e-05) (hash(x)=152294662)
134
+ 4400 val loss 5.8614
135
+ 4400 val perplexity 351.2089
136
+ 4400 train 5.759041 (lr=3.2368e-05) (hash(x)=141804386)
137
+ 4500 val loss 5.8419
138
+ 4500 val perplexity 344.4301
139
+ 4500 train 5.738522 (lr=3.1428e-05) (hash(x)=151095242)
140
+ 4600 val loss 5.8479
141
+ 4600 val perplexity 346.4965
142
+ 4600 train 5.792981 (lr=3.0486e-05) (hash(x)=156414699)
143
+ 4700 val loss 5.8213
144
+ 4700 val perplexity 337.4196
145
+ 4700 train 5.835704 (lr=2.9543e-05) (hash(x)=161556686)
146
+ 4800 val loss 5.8061
147
+ 4800 val perplexity 332.3093
148
+ 4800 train 5.715901 (lr=2.8602e-05) (hash(x)=149000293)
149
+ 4900 val loss 5.7968
150
+ 4900 val perplexity 329.2426
151
+ 4900 train 5.584097 (lr=2.7663e-05) (hash(x)=154349989)
152
+ 5000 val loss 5.7843
153
+ 5000 val perplexity 325.1438
154
+ 5000 train 5.614993 (lr=2.6728e-05) (hash(x)=131475967)
155
+ 5100 val loss 5.7728
156
+ 5100 val perplexity 321.4481
157
+ 5100 train 5.674039 (lr=2.5798e-05) (hash(x)=149717902)
158
+ 5200 val loss 5.7695
159
+ 5200 val perplexity 320.3890
160
+ 5200 train 5.634620 (lr=2.4874e-05) (hash(x)=151407999)
161
+ 5300 val loss 5.7545
162
+ 5300 val perplexity 315.6197
163
+ 5300 train 5.843950 (lr=2.3958e-05) (hash(x)=168602728)
164
+ 5400 val loss 5.7444
165
+ 5400 val perplexity 312.4250
166
+ 5400 train 5.700167 (lr=2.3051e-05) (hash(x)=158344511)
167
+ 5500 val loss 5.7369
168
+ 5500 val perplexity 310.0940
169
+ 5500 train 5.696602 (lr=2.2155e-05) (hash(x)=148350057)
170
+ 5600 val loss 5.7255
171
+ 5600 val perplexity 306.5844
172
+ 5600 train 5.792315 (lr=2.1271e-05) (hash(x)=153847323)
173
+ 5700 val loss 5.7168
174
+ 5700 val perplexity 303.9281
175
+ 5700 train 5.660481 (lr=2.0399e-05) (hash(x)=156607405)
176
+ 5800 val loss 5.7113
177
+ 5800 val perplexity 302.2715
178
+ 5800 train 5.476274 (lr=1.9542e-05) (hash(x)=145115031)
179
+ 5900 val loss 5.7032
180
+ 5900 val perplexity 299.8224
181
+ 5900 train 5.585124 (lr=1.8700e-05) (hash(x)=141584622)
182
+ 6000 val loss 5.6926
183
+ 6000 val perplexity 296.6544
184
+ 6000 train 5.623785 (lr=1.7875e-05) (hash(x)=146613857)
185
+ 6100 val loss 5.6872
186
+ 6100 val perplexity 295.0699
187
+ 6100 train 5.629707 (lr=1.7068e-05) (hash(x)=144621768)
188
+ 6200 val loss 5.6809
189
+ 6200 val perplexity 293.2248
190
+ 6200 train 5.612484 (lr=1.6280e-05) (hash(x)=146521760)
191
+ 6300 val loss 5.6744
192
+ 6300 val perplexity 291.3021
193
+ 6300 train 5.505092 (lr=1.5512e-05) (hash(x)=161378136)
194
+ 6400 val loss 5.6678
195
+ 6400 val perplexity 289.3883
196
+ 6400 train 5.434224 (lr=1.4766e-05) (hash(x)=141624235)
197
+ 6500 val loss 5.6614
198
+ 6500 val perplexity 287.5368
199
+ 6500 train 5.581383 (lr=1.4042e-05) (hash(x)=151197095)
200
+ 6600 val loss 5.6591
201
+ 6600 val perplexity 286.9044
202
+ 6600 train 5.677541 (lr=1.3342e-05) (hash(x)=153269571)
203
+ 6700 val loss 5.6526
204
+ 6700 val perplexity 285.0374
205
+ 6700 train 5.618601 (lr=1.2666e-05) (hash(x)=146111181)
206
+ 6800 val loss 5.6434
207
+ 6800 val perplexity 282.4169
208
+ 6800 train 5.672344 (lr=1.2016e-05) (hash(x)=147269760)
209
+ 6900 val loss 5.6390
210
+ 6900 val perplexity 281.1924
211
+ 6900 train 5.721029 (lr=1.1392e-05) (hash(x)=152912762)
212
+ 7000 val loss 5.6363
213
+ 7000 val perplexity 280.4270
214
+ 7000 train 5.724804 (lr=1.0795e-05) (hash(x)=165412343)
215
+ 7100 val loss 5.6272
216
+ 7100 val perplexity 277.8830
217
+ 7100 train 5.702348 (lr=1.0227e-05) (hash(x)=162866028)
218
+ 7200 val loss 5.6242
219
+ 7200 val perplexity 277.0432
220
+ 7200 train 5.594518 (lr=9.6875e-06) (hash(x)=142998115)
221
+ 7300 val loss 5.6216
222
+ 7300 val perplexity 276.3317
223
+ 7300 train 5.392981 (lr=9.1780e-06) (hash(x)=145486999)
224
+ 7400 val loss 5.6144
225
+ 7400 val perplexity 274.3558
226
+ 7400 train 5.591548 (lr=8.6990e-06) (hash(x)=155325873)
227
+ 7500 val loss 5.6108
228
+ 7500 val perplexity 273.3524
229
+ 7500 train 5.574790 (lr=8.2513e-06) (hash(x)=145131256)
230
+ 7600 val loss 5.6154
231
+ 7600 val perplexity 274.6144
232
+ 7600 train 5.539998 (lr=7.8355e-06) (hash(x)=144008365)
233
+ 7700 val loss 5.6057
234
+ 7700 val perplexity 271.9604
235
+ 7700 train 5.708950 (lr=7.4522e-06) (hash(x)=148848532)
236
+ 7800 val loss 5.6015
237
+ 7800 val perplexity 270.8214
238
+ 7800 train 5.587547 (lr=7.1019e-06) (hash(x)=150391642)
239
+ 7900 val loss 5.6012
240
+ 7900 val perplexity 270.7429
241
+ 7900 train 5.359458 (lr=6.7852e-06) (hash(x)=152191414)
242
+ 8000 val loss 5.5965
243
+ 8000 val perplexity 269.4868
244
+ 8000 train 5.674829 (lr=6.5025e-06) (hash(x)=159755587)
245
+ 8100 val loss 5.5929
246
+ 8100 val perplexity 268.5056
247
+ 8100 train 5.618392 (lr=6.2543e-06) (hash(x)=156664468)
248
+ 8200 val loss 5.5916
249
+ 8200 val perplexity 268.1588
250
+ 8200 train 5.440843 (lr=6.0408e-06) (hash(x)=139457379)
251
+ 8300 val loss 5.5967
252
+ 8300 val perplexity 269.5344
253
+ 8300 train 5.369544 (lr=5.8625e-06) (hash(x)=145478564)
254
+ 8400 val loss 5.5865
255
+ 8400 val perplexity 266.8061
256
+ 8400 train 5.525924 (lr=5.7195e-06) (hash(x)=154982769)
257
+ 8500 val loss 5.5846
258
+ 8500 val perplexity 266.2822
259
+ 8500 train 5.552999 (lr=5.6121e-06) (hash(x)=145798118)
260
+ 8600 val loss 5.5844
261
+ 8600 val perplexity 266.2398
262
+ 8600 train 5.494684 (lr=5.5404e-06) (hash(x)=143231551)
263
+ 8700 val loss 5.5872
264
+ 8700 val perplexity 266.9780
265
+ 8700 train 5.486710 (lr=5.5045e-06) (hash(x)=154780112)
266
+ 8749 val loss 5.5793
267
+ 8749 val perplexity 264.8789
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
attention_kindselective_n_heads4_seed1340/model_08749.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e0deb68cd19fbb766e0c31a28d0c7f0f91e0468f8c0970be55918f414ab5a2e1
3
  size 92843394
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2b41aec4b5aa02ea394f501290423bcd9f8ce3f46e72eca126939a6364eda72
3
  size 92843394
attention_kindselective_n_heads4_seed1340/optimizer_08749.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b680f6fd82e56ab5225beac288979a0f6e12d7c7d4ec8b7753e22389e7775a9d
3
  size 179406214
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77641ea9c6c521b32f7a6a6ed7c39d1f0621a3ef114f834ef2d5b923b8687f0b
3
  size 179406214