andrew-healey commited on
Commit
29a86a3
·
verified ·
1 Parent(s): 295810e

Upload folder using huggingface_hub

Browse files
attention_kindselective_n_heads4_seed1342/args.json CHANGED
@@ -1 +1 @@
1
- {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_9/attention_kindselective_n_heads4_seed1342", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 4, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 8750, "warmup_steps": 500, "group": "wider_is_better_9", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1342, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 120, "total_batch_size": 61440, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 3.5e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "3.5e-5_61440_4_1342", "n_embd": 256}
 
1
+ {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_9/attention_kindselective_n_heads4_seed1342", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 4, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 8750, "warmup_steps": 500, "group": "wider_is_better_9", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1342, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 120, "total_batch_size": 61440, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 4.5e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "4.5e-5_61440_4_1342", "n_embd": 256}
attention_kindselective_n_heads4_seed1342/log2.txt CHANGED
@@ -1,534 +1,267 @@
1
  max_steps: 8750
2
- max_steps: 8750
3
- 0 val loss 11.2898
4
- 0 val perplexity 80004.5469
5
  0 val loss 11.2898
6
  0 val perplexity 80004.5469
7
- 0 train 11.289051 (lr=7.0000e-08) (hash(x)=150332693)
8
- 0 train 11.289051 (lr=6.0000e-08) (hash(x)=150332693)
9
- 100 val loss 10.1794
10
- 100 val perplexity 26354.9375
11
- 100 val loss 10.1043
12
- 100 val perplexity 24447.8750
13
- 100 train 10.142110 (lr=6.0600e-06) (hash(x)=149277319)
14
- 100 train 10.061918 (lr=7.0700e-06) (hash(x)=149277319)
15
- 200 val loss 9.6815
16
- 200 val perplexity 16018.4980
17
- 200 train 9.752102 (lr=1.2060e-05) (hash(x)=155934075)
18
- 200 val loss 9.6158
19
- 200 val perplexity 15000.1738
20
- 200 train 9.686650 (lr=1.4070e-05) (hash(x)=155934075)
21
- 300 val loss 8.7807
22
- 300 val perplexity 6507.2988
23
- 300 train 8.821930 (lr=1.8060e-05) (hash(x)=165402628)
24
- 300 val loss 8.6249
25
- 300 val perplexity 5568.3394
26
- 300 train 8.665680 (lr=2.1070e-05) (hash(x)=165402628)
27
- 400 val loss 7.9230
28
- 400 val perplexity 2759.9607
29
- 400 train 7.778969 (lr=2.4060e-05) (hash(x)=139168939)
30
- 400 val loss 7.8122
31
- 400 val perplexity 2470.4434
32
- 400 train 7.653944 (lr=2.8070e-05) (hash(x)=139168939)
33
- 500 val loss 7.5502
34
- 500 val perplexity 1901.0621
35
- 500 train 7.481020 (lr=3.0000e-05) (hash(x)=147134338)
36
- 500 val loss 7.5233
37
- 500 val perplexity 1850.6179
38
- 500 train 7.451856 (lr=3.5000e-05) (hash(x)=147134338)
39
- 600 val loss 7.4285
40
- 600 val perplexity 1683.2417
41
- 600 train 7.348928 (lr=2.9990e-05) (hash(x)=159086459)
42
- 600 val loss 7.4080
43
- 600 val perplexity 1649.0697
44
- 600 train 7.328346 (lr=3.4989e-05) (hash(x)=159086459)
45
- 700 val loss 7.3568
46
- 700 val perplexity 1566.8014
47
- 700 train 7.395999 (lr=2.9961e-05) (hash(x)=152399099)
48
- 700 val loss 7.3374
49
- 700 val perplexity 1536.7510
50
- 700 train 7.378353 (lr=3.4954e-05) (hash(x)=152399099)
51
- 800 val loss 7.2636
52
- 800 val perplexity 1427.3462
53
- 800 train 7.262696 (lr=2.9912e-05) (hash(x)=156548843)
54
- 800 val loss 7.2412
55
- 800 val perplexity 1395.8164
56
- 800 train 7.243083 (lr=3.4897e-05) (hash(x)=156548843)
57
- 900 val loss 7.1888
58
- 900 val perplexity 1324.5342
59
- 900 train 7.158834 (lr=2.9844e-05) (hash(x)=157889911)
60
- 900 val loss 7.1467
61
- 900 val perplexity 1269.9143
62
- 900 train 7.119726 (lr=3.4818e-05) (hash(x)=157889911)
63
- 1000 val loss 7.1303
64
- 1000 val perplexity 1249.2928
65
- 1000 train 6.941356 (lr=2.9756e-05) (hash(x)=142943707)
66
- 1000 val loss 7.0711
67
- 1000 val perplexity 1177.5012
68
- 1000 train 6.876932 (lr=3.4715e-05) (hash(x)=142943707)
69
- 1100 val loss 7.0742
70
- 1100 val perplexity 1181.1373
71
- 1100 train 7.080656 (lr=2.9649e-05) (hash(x)=148938689)
72
- 1100 val loss 7.0131
73
- 1100 val perplexity 1111.0625
74
- 1100 train 7.017052 (lr=3.4591e-05) (hash(x)=148938689)
75
- 1200 val loss 7.0067
76
- 1200 val perplexity 1104.0490
77
- 1200 train 7.136171 (lr=2.9523e-05) (hash(x)=161962903)
78
- 1200 val loss 6.9357
79
- 1200 val perplexity 1028.3834
80
- 1200 train 7.066735 (lr=3.4444e-05) (hash(x)=161962903)
81
- 1300 val loss 6.9482
82
- 1300 val perplexity 1041.2402
83
- 1300 train 6.917204 (lr=2.9378e-05) (hash(x)=145979235)
84
- 1300 val loss 6.8651
85
- 1300 val perplexity 958.2105
86
- 1300 train 6.840671 (lr=3.4275e-05) (hash(x)=145979235)
87
- 1400 val loss 6.9017
88
- 1400 val perplexity 993.9407
89
- 1400 train 6.849684 (lr=2.9215e-05) (hash(x)=159157507)
90
- 1400 val loss 6.8026
91
- 1400 val perplexity 900.1846
92
- 1400 train 6.751725 (lr=3.4084e-05) (hash(x)=159157507)
93
- 1500 val loss 6.8547
94
- 1500 val perplexity 948.3015
95
- 1500 train 6.894870 (lr=2.9033e-05) (hash(x)=145943130)
96
- 1500 val loss 6.7512
97
- 1500 val perplexity 855.1268
98
- 1500 train 6.790061 (lr=3.3872e-05) (hash(x)=145943130)
99
- 1600 val loss 6.7962
100
- 1600 val perplexity 894.4285
101
- 1600 train 6.587233 (lr=2.8833e-05) (hash(x)=143249173)
102
- 1600 val loss 6.6718
103
- 1600 val perplexity 789.8278
104
- 1600 train 6.468535 (lr=3.3638e-05) (hash(x)=143249173)
105
- 1700 val loss 6.7377
106
- 1700 val perplexity 843.5904
107
- 1700 train 6.812438 (lr=2.8615e-05) (hash(x)=173907143)
108
- 1700 val loss 6.6123
109
- 1700 val perplexity 744.1997
110
- 1700 train 6.683813 (lr=3.3384e-05) (hash(x)=173907143)
111
- 1800 val loss 6.6922
112
- 1800 val perplexity 806.0727
113
- 1800 train 6.619445 (lr=2.8379e-05) (hash(x)=138363755)
114
- 1800 val loss 6.5679
115
- 1800 val perplexity 711.8691
116
- 1800 train 6.490389 (lr=3.3109e-05) (hash(x)=138363755)
117
- 1900 val loss 6.6449
118
- 1900 val perplexity 768.8578
119
- 1900 train 6.460667 (lr=2.8127e-05) (hash(x)=146341390)
120
- 1900 val loss 6.5231
121
- 1900 val perplexity 680.7050
122
- 1900 train 6.339457 (lr=3.2814e-05) (hash(x)=146341390)
123
- 2000 val loss 6.5923
124
- 2000 val perplexity 729.4580
125
- 2000 train 6.609406 (lr=2.7857e-05) (hash(x)=154559671)
126
- 2000 val loss 6.4736
127
- 2000 val perplexity 647.8088
128
- 2000 train 6.496451 (lr=3.2500e-05) (hash(x)=154559671)
129
- 2100 val loss 6.5569
130
- 2100 val perplexity 704.1078
131
- 2100 train 6.621999 (lr=2.7571e-05) (hash(x)=155607137)
132
- 2100 val loss 6.4424
133
- 2100 val perplexity 627.8850
134
- 2100 train 6.498925 (lr=3.2166e-05) (hash(x)=155607137)
135
- 2200 val loss 6.5175
136
- 2200 val perplexity 676.9085
137
- 2200 train 6.371143 (lr=2.7269e-05) (hash(x)=153010221)
138
- 2200 val loss 6.4053
139
- 2200 val perplexity 605.0538
140
- 2200 train 6.259152 (lr=3.1813e-05) (hash(x)=153010221)
141
- 2300 val loss 6.4806
142
- 2300 val perplexity 652.3843
143
- 2300 train 6.504387 (lr=2.6951e-05) (hash(x)=160097777)
144
- 2300 val loss 6.3697
145
- 2300 val perplexity 583.9057
146
- 2300 train 6.387885 (lr=3.1443e-05) (hash(x)=160097777)
147
- 2400 val loss 6.4437
148
- 2400 val perplexity 628.7560
149
- 2400 train 6.401982 (lr=2.6618e-05) (hash(x)=135590094)
150
- 2400 val loss 6.3429
151
- 2400 val perplexity 568.4287
152
- 2400 train 6.308095 (lr=3.1054e-05) (hash(x)=135590094)
153
- 2500 val loss 6.4175
154
- 2500 val perplexity 612.4648
155
- 2500 train 6.538583 (lr=2.6270e-05) (hash(x)=147042929)
156
- 2500 val loss 6.3173
157
- 2500 val perplexity 554.0828
158
- 2500 train 6.432256 (lr=3.0649e-05) (hash(x)=147042929)
159
- 2600 val loss 6.3833
160
- 2600 val perplexity 591.8883
161
- 2600 train 6.348204 (lr=2.5909e-05) (hash(x)=163793881)
162
- 2600 val loss 6.2817
163
- 2600 val perplexity 534.7117
164
- 2600 train 6.245691 (lr=3.0227e-05) (hash(x)=163793881)
165
- 2700 val loss 6.3584
166
- 2700 val perplexity 577.3007
167
- 2700 train 6.389218 (lr=2.5533e-05) (hash(x)=153598698)
168
- 2700 val loss 6.2576
169
- 2700 val perplexity 521.9579
170
- 2700 train 6.286983 (lr=2.9789e-05) (hash(x)=153598698)
171
- 2800 val loss 6.3386
172
- 2800 val perplexity 565.9789
173
- 2800 train 6.140002 (lr=2.5145e-05) (hash(x)=133439803)
174
- 2800 val loss 6.2427
175
- 2800 val perplexity 514.2557
176
- 2800 train 6.040616 (lr=2.9336e-05) (hash(x)=133439803)
177
- 2900 val loss 6.3106
178
- 2900 val perplexity 550.3931
179
- 2900 train 6.254204 (lr=2.4744e-05) (hash(x)=148239158)
180
- 2900 val loss 6.2123
181
- 2900 val perplexity 498.8553
182
- 2900 train 6.154753 (lr=2.8868e-05) (hash(x)=148239158)
183
- 3000 val loss 6.2870
184
- 3000 val perplexity 537.5233
185
- 3000 train 6.281301 (lr=2.4331e-05) (hash(x)=150991971)
186
- 3000 val loss 6.1935
187
- 3000 val perplexity 489.5722
188
- 3000 train 6.193192 (lr=2.8386e-05) (hash(x)=150991971)
189
- 3100 val loss 6.2690
190
- 3100 val perplexity 527.9424
191
- 3100 train 6.275957 (lr=2.3906e-05) (hash(x)=157791832)
192
- 3100 val loss 6.1763
193
- 3100 val perplexity 481.2045
194
- 3100 train 6.180146 (lr=2.7891e-05) (hash(x)=157791832)
195
- 3200 val loss 6.2427
196
- 3200 val perplexity 514.2594
197
- 3200 train 6.317775 (lr=2.3471e-05) (hash(x)=163746305)
198
- 3200 val loss 6.1504
199
- 3200 val perplexity 468.9072
200
- 3200 train 6.224569 (lr=2.7383e-05) (hash(x)=163746305)
201
- 3300 val loss 6.2315
202
- 3300 val perplexity 508.5287
203
- 3300 train 6.160985 (lr=2.3026e-05) (hash(x)=155120269)
204
- 3300 val loss 6.1413
205
- 3300 val perplexity 464.6666
206
- 3300 train 6.069816 (lr=2.6864e-05) (hash(x)=155120269)
207
- 3400 val loss 6.2016
208
- 3400 val perplexity 493.5572
209
- 3400 train 6.141113 (lr=2.2572e-05) (hash(x)=152748451)
210
- 3400 val loss 6.1122
211
- 3400 val perplexity 451.3326
212
- 3400 train 6.053649 (lr=2.6333e-05) (hash(x)=152748451)
213
- 3500 val loss 6.1898
214
- 3500 val perplexity 487.7594
215
- 3500 train 6.065310 (lr=2.2108e-05) (hash(x)=147395004)
216
- 3500 val loss 6.0981
217
- 3500 val perplexity 445.0312
218
- 3500 train 5.977302 (lr=2.5793e-05) (hash(x)=147395004)
219
- 3600 val loss 6.1764
220
- 3600 val perplexity 481.2518
221
- 3600 train 6.201883 (lr=2.1637e-05) (hash(x)=157554250)
222
- 3600 val loss 6.0854
223
- 3600 val perplexity 439.3748
224
- 3600 train 6.107175 (lr=2.5243e-05) (hash(x)=157554250)
225
- 3700 val loss 6.1583
226
- 3700 val perplexity 472.6462
227
- 3700 train 6.011199 (lr=2.1158e-05) (hash(x)=147953321)
228
- 3700 val loss 6.0681
229
- 3700 val perplexity 431.8563
230
- 3700 train 5.921242 (lr=2.4684e-05) (hash(x)=147953321)
231
- 3800 val loss 6.1382
232
- 3800 val perplexity 463.2388
233
- 3800 train 5.997552 (lr=2.0672e-05) (hash(x)=146898609)
234
- 3800 val loss 6.0475
235
- 3800 val perplexity 423.0692
236
- 3800 train 5.903200 (lr=2.4117e-05) (hash(x)=146898609)
237
- 3900 val loss 6.1327
238
- 3900 val perplexity 460.6937
239
- 3900 train 6.182793 (lr=2.0180e-05) (hash(x)=150115282)
240
- 3900 val loss 6.0465
241
- 3900 val perplexity 422.6220
242
- 3900 train 6.085420 (lr=2.3543e-05) (hash(x)=150115282)
243
- 4000 val loss 6.1066
244
- 4000 val perplexity 448.7970
245
- 4000 train 5.963685 (lr=1.9683e-05) (hash(x)=142448619)
246
- 4000 val loss 6.0158
247
- 4000 val perplexity 409.8366
248
- 4000 train 5.875909 (lr=2.2963e-05) (hash(x)=142448619)
249
- 4100 val loss 6.1067
250
- 4100 val perplexity 448.8636
251
- 4100 train 5.829423 (lr=1.9181e-05) (hash(x)=141244123)
252
- 4100 val loss 6.0114
253
- 4100 val perplexity 408.0352
254
- 4100 train 5.743440 (lr=2.2378e-05) (hash(x)=141244123)
255
- 4200 val loss 6.0938
256
- 4200 val perplexity 443.1073
257
- 4200 train 6.123459 (lr=1.8675e-05) (hash(x)=165913661)
258
- 4200 val loss 6.0059
259
- 4200 val perplexity 405.8220
260
- 4200 train 6.037457 (lr=2.1788e-05) (hash(x)=165913661)
261
- 4300 val loss 6.0734
262
- 4300 val perplexity 434.1442
263
- 4300 train 6.109362 (lr=1.8166e-05) (hash(x)=155744823)
264
- 4300 val loss 5.9803
265
- 4300 val perplexity 395.5565
266
- 4300 train 6.013528 (lr=2.1194e-05) (hash(x)=155744823)
267
- 4400 val loss 6.0636
268
- 4400 val perplexity 429.9022
269
- 4400 train 6.056406 (lr=1.7655e-05) (hash(x)=160461243)
270
- 4400 val loss 5.9719
271
- 4400 val perplexity 392.2354
272
- 4400 train 5.965618 (lr=2.0598e-05) (hash(x)=160461243)
273
- 4500 val loss 6.0538
274
- 4500 val perplexity 425.7418
275
- 4500 train 5.980674 (lr=1.7142e-05) (hash(x)=156073129)
276
- 4500 val loss 5.9604
277
- 4500 val perplexity 387.7795
278
- 4500 train 5.879493 (lr=1.9999e-05) (hash(x)=156073129)
279
- 4600 val loss 6.0368
280
- 4600 val perplexity 418.5724
281
- 4600 train 6.086256 (lr=1.6629e-05) (hash(x)=152294132)
282
- 4600 val loss 5.9469
283
- 4600 val perplexity 382.5828
284
- 4600 train 5.993891 (lr=1.9400e-05) (hash(x)=152294132)
285
- 4700 val loss 6.0269
286
- 4700 val perplexity 414.4126
287
- 4700 train 5.966726 (lr=1.6114e-05) (hash(x)=140904569)
288
- 4700 val loss 5.9346
289
- 4700 val perplexity 377.8774
290
- 4700 train 5.868339 (lr=1.8800e-05) (hash(x)=140904569)
291
- 4800 val loss 6.0212
292
- 4800 val perplexity 412.0571
293
- 4800 train 5.943006 (lr=1.5601e-05) (hash(x)=156191635)
294
- 4800 val loss 5.9285
295
- 4800 val perplexity 375.5807
296
- 4800 train 5.838706 (lr=1.8201e-05) (hash(x)=156191635)
297
- 4900 val loss 6.0084
298
- 4900 val perplexity 406.8464
299
- 4900 train 6.088559 (lr=1.5089e-05) (hash(x)=145063976)
300
- 4900 val loss 5.9164
301
- 4900 val perplexity 371.0731
302
- 4900 train 6.010023 (lr=1.7604e-05) (hash(x)=145063976)
303
- 5000 val loss 6.0026
304
- 5000 val perplexity 404.4739
305
- 5000 train 5.947785 (lr=1.4579e-05) (hash(x)=160110619)
306
- 5000 val loss 5.9095
307
- 5000 val perplexity 368.5036
308
- 5000 train 5.861709 (lr=1.7009e-05) (hash(x)=160110619)
309
- 5100 val loss 5.9873
310
- 5100 val perplexity 398.3270
311
- 5100 train 6.010914 (lr=1.4071e-05) (hash(x)=156270070)
312
- 5100 val loss 5.8941
313
- 5100 val perplexity 362.8853
314
- 5100 train 5.915792 (lr=1.6417e-05) (hash(x)=156270070)
315
- 5200 val loss 5.9787
316
- 5200 val perplexity 394.9133
317
- 5200 train 5.849009 (lr=1.3568e-05) (hash(x)=138954242)
318
- 5200 val loss 5.8882
319
- 5200 val perplexity 360.7615
320
- 5200 train 5.765492 (lr=1.5829e-05) (hash(x)=138954242)
321
- 5300 val loss 5.9750
322
- 5300 val perplexity 393.4499
323
- 5300 train 5.825052 (lr=1.3068e-05) (hash(x)=146472367)
324
- 5300 val loss 5.8781
325
- 5300 val perplexity 357.1307
326
- 5300 train 5.724410 (lr=1.5246e-05) (hash(x)=146472367)
327
- 5400 val loss 5.9621
328
- 5400 val perplexity 388.4326
329
- 5400 train 5.902071 (lr=1.2573e-05) (hash(x)=146753405)
330
- 5400 val loss 5.8667
331
- 5400 val perplexity 353.0806
332
- 5400 train 5.797910 (lr=1.4669e-05) (hash(x)=146753405)
333
- 5500 val loss 5.9555
334
- 5500 val perplexity 385.8689
335
- 5500 train 5.900256 (lr=1.2085e-05) (hash(x)=147757398)
336
- 5500 val loss 5.8596
337
- 5500 val perplexity 350.5704
338
- 5500 train 5.799724 (lr=1.4099e-05) (hash(x)=147757398)
339
- 5600 val loss 5.9502
340
- 5600 val perplexity 383.8358
341
- 5600 train 5.831348 (lr=1.1602e-05) (hash(x)=151291757)
342
- 5600 val loss 5.8539
343
- 5600 val perplexity 348.5839
344
- 5600 train 5.729513 (lr=1.3536e-05) (hash(x)=151291757)
345
- 5700 val loss 5.9380
346
- 5700 val perplexity 379.1620
347
- 5700 train 5.980654 (lr=1.1127e-05) (hash(x)=156274392)
348
- 5700 val loss 5.8421
349
- 5700 val perplexity 344.4909
350
- 5700 train 5.884508 (lr=1.2981e-05) (hash(x)=156274392)
351
- 5800 val loss 5.9322
352
- 5800 val perplexity 376.9763
353
- 5800 train 5.895813 (lr=1.0659e-05) (hash(x)=152612266)
354
- 5800 val loss 5.8349
355
- 5800 val perplexity 342.0416
356
- 5800 train 5.799439 (lr=1.2436e-05) (hash(x)=152612266)
357
- 5900 val loss 5.9274
358
- 5900 val perplexity 375.1920
359
- 5900 train 6.081093 (lr=1.0200e-05) (hash(x)=148973201)
360
- 5900 val loss 5.8301
361
- 5900 val perplexity 340.3840
362
- 5900 train 5.991538 (lr=1.1900e-05) (hash(x)=148973201)
363
- 6000 val loss 5.9201
364
- 6000 val perplexity 372.4622
365
- 6000 train 5.883375 (lr=9.7500e-06) (hash(x)=158311490)
366
- 6000 val loss 5.8212
367
- 6000 val perplexity 337.3903
368
- 6000 train 5.787454 (lr=1.1375e-05) (hash(x)=158311490)
369
- 6100 val loss 5.9126
370
- 6100 val perplexity 369.6491
371
- 6100 train 5.831893 (lr=9.3098e-06) (hash(x)=147541319)
372
- 6100 val loss 5.8151
373
- 6100 val perplexity 335.3258
374
- 6100 train 5.741688 (lr=1.0861e-05) (hash(x)=147541319)
375
- 6200 val loss 5.9105
376
- 6200 val perplexity 368.8990
377
- 6200 train 5.971362 (lr=8.8800e-06) (hash(x)=154753980)
378
- 6200 val loss 5.8099
379
- 6200 val perplexity 333.5760
380
- 6200 train 5.873833 (lr=1.0360e-05) (hash(x)=154753980)
381
- 6300 val loss 5.9034
382
- 6300 val perplexity 366.2755
383
- 6300 train 5.801030 (lr=8.4613e-06) (hash(x)=140798998)
384
- 6300 val loss 5.8038
385
- 6300 val perplexity 331.5592
386
- 6300 train 5.703178 (lr=9.8715e-06) (hash(x)=140798998)
387
- 6400 val loss 5.8987
388
- 6400 val perplexity 364.5718
389
- 6400 train 5.783006 (lr=8.0542e-06) (hash(x)=149299516)
390
- 6400 val loss 5.8005
391
- 6400 val perplexity 330.4565
392
- 6400 train 5.687269 (lr=9.3966e-06) (hash(x)=149299516)
393
- 6500 val loss 5.8946
394
- 6500 val perplexity 363.0667
395
- 6500 train 5.893285 (lr=7.6594e-06) (hash(x)=148495806)
396
- 6500 val loss 5.7949
397
- 6500 val perplexity 328.6337
398
- 6500 train 5.787768 (lr=8.9359e-06) (hash(x)=148495806)
399
- 6600 val loss 5.8868
400
- 6600 val perplexity 360.2643
401
- 6600 train 5.860414 (lr=7.2774e-06) (hash(x)=153664387)
402
- 6600 val loss 5.7871
403
- 6600 val perplexity 326.0732
404
- 6600 train 5.761164 (lr=8.4903e-06) (hash(x)=153664387)
405
- 6700 val loss 5.8848
406
- 6700 val perplexity 359.5459
407
- 6700 train 5.866759 (lr=6.9087e-06) (hash(x)=153333491)
408
- 6700 val loss 5.7843
409
- 6700 val perplexity 325.1652
410
- 6700 train 5.766811 (lr=8.0602e-06) (hash(x)=153333491)
411
- 6800 val loss 5.8856
412
- 6800 val perplexity 359.8354
413
- 6800 train 5.644751 (lr=6.5540e-06) (hash(x)=150149382)
414
- 6800 val loss 5.7835
415
- 6800 val perplexity 324.8955
416
- 6800 train 5.536487 (lr=7.6463e-06) (hash(x)=150149382)
417
- 6900 val loss 5.8772
418
- 6900 val perplexity 356.8184
419
- 6900 train 5.949944 (lr=6.2137e-06) (hash(x)=152379767)
420
- 6900 val loss 5.7778
421
- 6900 val perplexity 323.0596
422
- 6900 train 5.856633 (lr=7.2493e-06) (hash(x)=152379767)
423
- 7000 val loss 5.8711
424
- 7000 val perplexity 354.6424
425
- 7000 train 5.916762 (lr=5.8883e-06) (hash(x)=150432537)
426
- 7000 val loss 5.7705
427
- 7000 val perplexity 320.6822
428
- 7000 train 5.818938 (lr=6.8697e-06) (hash(x)=150432537)
429
- 7100 val loss 5.8700
430
- 7100 val perplexity 354.2435
431
- 7100 train 5.747322 (lr=5.5783e-06) (hash(x)=148277836)
432
- 7100 val loss 5.7707
433
- 7100 val perplexity 320.7480
434
- 7100 train 5.649139 (lr=6.5080e-06) (hash(x)=148277836)
435
- 7200 val loss 5.8696
436
- 7200 val perplexity 354.0941
437
- 7200 train 5.876884 (lr=5.2841e-06) (hash(x)=168591659)
438
- 7200 val loss 5.7699
439
- 7200 val perplexity 320.5116
440
- 7200 train 5.778949 (lr=6.1648e-06) (hash(x)=168591659)
441
- 7300 val loss 5.8627
442
- 7300 val perplexity 351.6757
443
- 7300 train 5.839412 (lr=5.0062e-06) (hash(x)=152764851)
444
- 7300 val loss 5.7618
445
- 7300 val perplexity 317.9327
446
- 7300 train 5.739857 (lr=5.8405e-06) (hash(x)=152764851)
447
- 7400 val loss 5.8589
448
- 7400 val perplexity 350.3225
449
- 7400 train 5.809220 (lr=4.7449e-06) (hash(x)=142920131)
450
- 7400 val loss 5.7583
451
- 7400 val perplexity 316.8126
452
- 7400 train 5.714902 (lr=5.5357e-06) (hash(x)=142920131)
453
- 7500 val loss 5.8567
454
- 7500 val perplexity 349.5636
455
- 7500 train 5.858554 (lr=4.5007e-06) (hash(x)=146954649)
456
- 7500 val loss 5.7558
457
- 7500 val perplexity 316.0316
458
- 7500 train 5.757757 (lr=5.2508e-06) (hash(x)=146954649)
459
- 7600 val loss 5.8574
460
- 7600 val perplexity 349.8061
461
- 7600 train 5.648992 (lr=4.2739e-06) (hash(x)=144529566)
462
- 7600 val loss 5.7568
463
- 7600 val perplexity 316.3331
464
- 7600 train 5.546754 (lr=4.9862e-06) (hash(x)=144529566)
465
- 7700 val loss 5.8515
466
- 7700 val perplexity 347.7727
467
- 7700 train 5.806117 (lr=4.0648e-06) (hash(x)=142669945)
468
- 7700 val loss 5.7503
469
- 7700 val perplexity 314.2869
470
- 7700 train 5.707147 (lr=4.7423e-06) (hash(x)=142669945)
471
- 7800 val loss 5.8481
472
- 7800 val perplexity 346.5742
473
- 7800 train 5.919096 (lr=3.8738e-06) (hash(x)=161563764)
474
- 7800 val loss 5.7467
475
- 7800 val perplexity 313.1556
476
- 7800 train 5.816212 (lr=4.5194e-06) (hash(x)=161563764)
477
- 7900 val loss 5.8457
478
- 7900 val perplexity 345.7390
479
- 7900 train 5.643712 (lr=3.7010e-06) (hash(x)=141516532)
480
- 7900 val loss 5.7451
481
- 7900 val perplexity 312.6412
482
- 7900 train 5.545619 (lr=4.3179e-06) (hash(x)=141516532)
483
- 8000 val loss 5.8473
484
- 8000 val perplexity 346.2927
485
- 8000 train 5.716923 (lr=3.5468e-06) (hash(x)=152418457)
486
- 8000 val loss 5.7454
487
- 8000 val perplexity 312.7387
488
- 8000 train 5.618537 (lr=4.1380e-06) (hash(x)=152418457)
489
- 8100 val loss 5.8420
490
- 8100 val perplexity 344.4742
491
- 8100 train 5.831784 (lr=3.4114e-06) (hash(x)=161109470)
492
- 8100 val loss 5.7407
493
- 8100 val perplexity 311.2804
494
- 8100 train 5.731179 (lr=3.9800e-06) (hash(x)=161109470)
495
- 8200 val loss 5.8385
496
- 8200 val perplexity 343.2647
497
- 8200 train 5.958980 (lr=3.2950e-06) (hash(x)=154408822)
498
- 8200 val loss 5.7371
499
- 8200 val perplexity 310.1499
500
- 8200 train 5.859767 (lr=3.8442e-06) (hash(x)=154408822)
501
- 8300 val loss 5.8363
502
- 8300 val perplexity 342.5162
503
- 8300 train 5.717995 (lr=3.1977e-06) (hash(x)=146431691)
504
- 8300 val loss 5.7353
505
- 8300 val perplexity 309.5942
506
- 8300 train 5.619819 (lr=3.7307e-06) (hash(x)=146431691)
507
- 8400 val loss 5.8365
508
- 8400 val perplexity 342.5848
509
- 8400 train 5.609816 (lr=3.1197e-06) (hash(x)=150182527)
510
- 8400 val loss 5.7352
511
- 8400 val perplexity 309.5880
512
- 8400 train 5.512958 (lr=3.6397e-06) (hash(x)=150182527)
513
- 8500 val loss 5.8311
514
- 8500 val perplexity 340.7312
515
- 8500 train 5.901333 (lr=3.0611e-06) (hash(x)=154828709)
516
- 8500 val loss 5.7297
517
- 8500 val perplexity 307.8887
518
- 8500 train 5.799944 (lr=3.5713e-06) (hash(x)=154828709)
519
- 8600 val loss 5.8306
520
- 8600 val perplexity 340.5702
521
- 8600 train 5.737008 (lr=3.0220e-06) (hash(x)=154995428)
522
- 8600 val loss 5.7294
523
- 8600 val perplexity 307.7847
524
- 8600 train 5.641396 (lr=3.5257e-06) (hash(x)=154995428)
525
- 8700 val loss 5.8302
526
- 8700 val perplexity 340.4246
527
- 8700 train 5.724449 (lr=3.0024e-06) (hash(x)=142111043)
528
- 8700 val loss 5.7292
529
- 8700 val perplexity 307.7150
530
- 8700 train 5.626750 (lr=3.5029e-06) (hash(x)=142111043)
531
- 8749 val loss 5.8284
532
- 8749 val perplexity 339.8080
533
- 8749 val loss 5.7264
534
- 8749 val perplexity 306.8476
 
1
  max_steps: 8750
 
 
 
2
  0 val loss 11.2898
3
  0 val perplexity 80004.5469
4
+ 0 train 11.289043 (lr=9.0000e-08) (hash(x)=150332693)
5
+ 100 val loss 10.0243
6
+ 100 val perplexity 22568.6875
7
+ 100 train 9.978326 (lr=9.0900e-06) (hash(x)=149277319)
8
+ 200 val loss 9.3785
9
+ 200 val perplexity 11831.2090
10
+ 200 train 9.445846 (lr=1.8090e-05) (hash(x)=155934075)
11
+ 300 val loss 8.2194
12
+ 300 val perplexity 3712.2971
13
+ 300 train 8.254610 (lr=2.7090e-05) (hash(x)=165402628)
14
+ 400 val loss 7.6163
15
+ 400 val perplexity 2031.0946
16
+ 400 train 7.421521 (lr=3.6090e-05) (hash(x)=139168939)
17
+ 500 val loss 7.4349
18
+ 500 val perplexity 1694.0548
19
+ 500 train 7.357471 (lr=4.5000e-05) (hash(x)=147134338)
20
+ 600 val loss 7.3305
21
+ 600 val perplexity 1526.1565
22
+ 600 train 7.243377 (lr=4.4985e-05) (hash(x)=159086459)
23
+ 700 val loss 7.2503
24
+ 700 val perplexity 1408.4863
25
+ 700 train 7.288620 (lr=4.4941e-05) (hash(x)=152399099)
26
+ 800 val loss 7.1363
27
+ 800 val perplexity 1256.7866
28
+ 800 train 7.140655 (lr=4.4868e-05) (hash(x)=156548843)
29
+ 900 val loss 7.0484
30
+ 900 val perplexity 1151.0243
31
+ 900 train 7.021039 (lr=4.4766e-05) (hash(x)=157889911)
32
+ 1000 val loss 6.9766
33
+ 1000 val perplexity 1071.3145
34
+ 1000 train 6.774483 (lr=4.4634e-05) (hash(x)=142943707)
35
+ 1100 val loss 6.9135
36
+ 1100 val perplexity 1005.7903
37
+ 1100 train 6.912020 (lr=4.4474e-05) (hash(x)=148938689)
38
+ 1200 val loss 6.8143
39
+ 1200 val perplexity 910.7801
40
+ 1200 train 6.954645 (lr=4.4285e-05) (hash(x)=161962903)
41
+ 1300 val loss 6.7411
42
+ 1300 val perplexity 846.4569
43
+ 1300 train 6.723616 (lr=4.4068e-05) (hash(x)=145979235)
44
+ 1400 val loss 6.6754
45
+ 1400 val perplexity 792.6756
46
+ 1400 train 6.625307 (lr=4.3822e-05) (hash(x)=159157507)
47
+ 1500 val loss 6.6153
48
+ 1500 val perplexity 746.3971
49
+ 1500 train 6.649820 (lr=4.3549e-05) (hash(x)=145943130)
50
+ 1600 val loss 6.5401
51
+ 1600 val perplexity 692.3466
52
+ 1600 train 6.341069 (lr=4.3249e-05) (hash(x)=143249173)
53
+ 1700 val loss 6.4872
54
+ 1700 val perplexity 656.6964
55
+ 1700 train 6.551693 (lr=4.2922e-05) (hash(x)=173907143)
56
+ 1800 val loss 6.4411
57
+ 1800 val perplexity 627.1124
58
+ 1800 train 6.368075 (lr=4.2569e-05) (hash(x)=138363755)
59
+ 1900 val loss 6.4086
60
+ 1900 val perplexity 607.0637
61
+ 1900 train 6.216325 (lr=4.2190e-05) (hash(x)=146341390)
62
+ 2000 val loss 6.3554
63
+ 2000 val perplexity 575.6113
64
+ 2000 train 6.384245 (lr=4.1785e-05) (hash(x)=154559671)
65
+ 2100 val loss 6.3281
66
+ 2100 val perplexity 560.0664
67
+ 2100 train 6.378413 (lr=4.1356e-05) (hash(x)=155607137)
68
+ 2200 val loss 6.2939
69
+ 2200 val perplexity 541.2359
70
+ 2200 train 6.150013 (lr=4.0903e-05) (hash(x)=153010221)
71
+ 2300 val loss 6.2609
72
+ 2300 val perplexity 523.7111
73
+ 2300 train 6.278090 (lr=4.0426e-05) (hash(x)=160097777)
74
+ 2400 val loss 6.2372
75
+ 2400 val perplexity 511.4395
76
+ 2400 train 6.214860 (lr=3.9927e-05) (hash(x)=135590094)
77
+ 2500 val loss 6.2120
78
+ 2500 val perplexity 498.6850
79
+ 2500 train 6.329730 (lr=3.9406e-05) (hash(x)=147042929)
80
+ 2600 val loss 6.1775
81
+ 2600 val perplexity 481.7953
82
+ 2600 train 6.138986 (lr=3.8863e-05) (hash(x)=163793881)
83
+ 2700 val loss 6.1626
84
+ 2700 val perplexity 474.6808
85
+ 2700 train 6.195170 (lr=3.8300e-05) (hash(x)=153598698)
86
+ 2800 val loss 6.1463
87
+ 2800 val perplexity 466.9887
88
+ 2800 train 5.944228 (lr=3.7717e-05) (hash(x)=133439803)
89
+ 2900 val loss 6.1207
90
+ 2900 val perplexity 455.1814
91
+ 2900 train 6.059956 (lr=3.7116e-05) (hash(x)=148239158)
92
+ 3000 val loss 6.0949
93
+ 3000 val perplexity 443.6075
94
+ 3000 train 6.099676 (lr=3.6496e-05) (hash(x)=150991971)
95
+ 3100 val loss 6.0827
96
+ 3100 val perplexity 438.2101
97
+ 3100 train 6.087112 (lr=3.5860e-05) (hash(x)=157791832)
98
+ 3200 val loss 6.0558
99
+ 3200 val perplexity 426.5774
100
+ 3200 train 6.124968 (lr=3.5207e-05) (hash(x)=163746305)
101
+ 3300 val loss 6.0481
102
+ 3300 val perplexity 423.3175
103
+ 3300 train 5.978697 (lr=3.4539e-05) (hash(x)=155120269)
104
+ 3400 val loss 6.0166
105
+ 3400 val perplexity 410.1862
106
+ 3400 train 5.958898 (lr=3.3857e-05) (hash(x)=152748451)
107
+ 3500 val loss 6.0043
108
+ 3500 val perplexity 405.1582
109
+ 3500 train 5.885794 (lr=3.3162e-05) (hash(x)=147395004)
110
+ 3600 val loss 5.9914
111
+ 3600 val perplexity 399.9552
112
+ 3600 train 6.009531 (lr=3.2455e-05) (hash(x)=157554250)
113
+ 3700 val loss 5.9740
114
+ 3700 val perplexity 393.0799
115
+ 3700 train 5.826887 (lr=3.1736e-05) (hash(x)=147953321)
116
+ 3800 val loss 5.9551
117
+ 3800 val perplexity 385.7162
118
+ 3800 train 5.805262 (lr=3.1008e-05) (hash(x)=146898609)
119
+ 3900 val loss 5.9564
120
+ 3900 val perplexity 386.2080
121
+ 3900 train 5.996837 (lr=3.0270e-05) (hash(x)=150115282)
122
+ 4000 val loss 5.9218
123
+ 4000 val perplexity 373.0663
124
+ 4000 train 5.790571 (lr=2.9524e-05) (hash(x)=142448619)
125
+ 4100 val loss 5.9223
126
+ 4100 val perplexity 373.2632
127
+ 4100 train 5.654961 (lr=2.8771e-05) (hash(x)=141244123)
128
+ 4200 val loss 5.9140
129
+ 4200 val perplexity 370.1866
130
+ 4200 train 5.927835 (lr=2.8013e-05) (hash(x)=165913661)
131
+ 4300 val loss 5.8861
132
+ 4300 val perplexity 360.0132
133
+ 4300 train 5.917614 (lr=2.7250e-05) (hash(x)=155744823)
134
+ 4400 val loss 5.8769
135
+ 4400 val perplexity 356.6874
136
+ 4400 train 5.872905 (lr=2.6483e-05) (hash(x)=160461243)
137
+ 4500 val loss 5.8657
138
+ 4500 val perplexity 352.7454
139
+ 4500 train 5.776689 (lr=2.5714e-05) (hash(x)=156073129)
140
+ 4600 val loss 5.8505
141
+ 4600 val perplexity 347.4088
142
+ 4600 train 5.897415 (lr=2.4943e-05) (hash(x)=152294132)
143
+ 4700 val loss 5.8393
144
+ 4700 val perplexity 343.5292
145
+ 4700 train 5.771152 (lr=2.4172e-05) (hash(x)=140904569)
146
+ 4800 val loss 5.8338
147
+ 4800 val perplexity 341.6528
148
+ 4800 train 5.739367 (lr=2.3402e-05) (hash(x)=156191635)
149
+ 4900 val loss 5.8181
150
+ 4900 val perplexity 336.3423
151
+ 4900 train 5.924228 (lr=2.2633e-05) (hash(x)=145063976)
152
+ 5000 val loss 5.8120
153
+ 5000 val perplexity 334.2776
154
+ 5000 train 5.754147 (lr=2.1868e-05) (hash(x)=160110619)
155
+ 5100 val loss 5.7954
156
+ 5100 val perplexity 328.7745
157
+ 5100 train 5.821685 (lr=2.1107e-05) (hash(x)=156270070)
158
+ 5200 val loss 5.7867
159
+ 5200 val perplexity 325.9519
160
+ 5200 train 5.674892 (lr=2.0351e-05) (hash(x)=138954242)
161
+ 5300 val loss 5.7796
162
+ 5300 val perplexity 323.6199
163
+ 5300 train 5.633004 (lr=1.9602e-05) (hash(x)=146472367)
164
+ 5400 val loss 5.7699
165
+ 5400 val perplexity 320.5030
166
+ 5400 train 5.702404 (lr=1.8860e-05) (hash(x)=146753405)
167
+ 5500 val loss 5.7575
168
+ 5500 val perplexity 316.5528
169
+ 5500 train 5.706319 (lr=1.8127e-05) (hash(x)=147757398)
170
+ 5600 val loss 5.7557
171
+ 5600 val perplexity 315.9973
172
+ 5600 train 5.624521 (lr=1.7403e-05) (hash(x)=151291757)
173
+ 5700 val loss 5.7424
174
+ 5700 val perplexity 311.8269
175
+ 5700 train 5.793109 (lr=1.6690e-05) (hash(x)=156274392)
176
+ 5800 val loss 5.7336
177
+ 5800 val perplexity 309.0707
178
+ 5800 train 5.702478 (lr=1.5989e-05) (hash(x)=152612266)
179
+ 5900 val loss 5.7301
180
+ 5900 val perplexity 307.9881
181
+ 5900 train 5.893963 (lr=1.5300e-05) (hash(x)=148973201)
182
+ 6000 val loss 5.7217
183
+ 6000 val perplexity 305.4335
184
+ 6000 train 5.687572 (lr=1.4625e-05) (hash(x)=158311490)
185
+ 6100 val loss 5.7135
186
+ 6100 val perplexity 302.9321
187
+ 6100 train 5.645096 (lr=1.3965e-05) (hash(x)=147541319)
188
+ 6200 val loss 5.7129
189
+ 6200 val perplexity 302.7402
190
+ 6200 train 5.777384 (lr=1.3320e-05) (hash(x)=154753980)
191
+ 6300 val loss 5.7037
192
+ 6300 val perplexity 299.9899
193
+ 6300 train 5.601553 (lr=1.2692e-05) (hash(x)=140798998)
194
+ 6400 val loss 5.6993
195
+ 6400 val perplexity 298.6710
196
+ 6400 train 5.588508 (lr=1.2081e-05) (hash(x)=149299516)
197
+ 6500 val loss 5.6953
198
+ 6500 val perplexity 297.4683
199
+ 6500 train 5.682362 (lr=1.1489e-05) (hash(x)=148495806)
200
+ 6600 val loss 5.6878
201
+ 6600 val perplexity 295.2325
202
+ 6600 train 5.665105 (lr=1.0916e-05) (hash(x)=153664387)
203
+ 6700 val loss 5.6849
204
+ 6700 val perplexity 294.3740
205
+ 6700 train 5.661620 (lr=1.0363e-05) (hash(x)=153333491)
206
+ 6800 val loss 5.6854
207
+ 6800 val perplexity 294.5452
208
+ 6800 train 5.438633 (lr=9.8310e-06) (hash(x)=150149382)
209
+ 6900 val loss 5.6780
210
+ 6900 val perplexity 292.3573
211
+ 6900 train 5.763326 (lr=9.3205e-06) (hash(x)=152379767)
212
+ 7000 val loss 5.6708
213
+ 7000 val perplexity 290.2598
214
+ 7000 train 5.719643 (lr=8.8324e-06) (hash(x)=150432537)
215
+ 7100 val loss 5.6688
216
+ 7100 val perplexity 289.6923
217
+ 7100 train 5.545722 (lr=8.3674e-06) (hash(x)=148277836)
218
+ 7200 val loss 5.6698
219
+ 7200 val perplexity 289.9790
220
+ 7200 train 5.664272 (lr=7.9261e-06) (hash(x)=168591659)
221
+ 7300 val loss 5.6615
222
+ 7300 val perplexity 287.5750
223
+ 7300 train 5.642782 (lr=7.5093e-06) (hash(x)=152764851)
224
+ 7400 val loss 5.6584
225
+ 7400 val perplexity 286.6996
226
+ 7400 train 5.621808 (lr=7.1174e-06) (hash(x)=142920131)
227
+ 7500 val loss 5.6559
228
+ 7500 val perplexity 285.9665
229
+ 7500 train 5.656323 (lr=6.7511e-06) (hash(x)=146954649)
230
+ 7600 val loss 5.6564
231
+ 7600 val perplexity 286.1283
232
+ 7600 train 5.434806 (lr=6.4109e-06) (hash(x)=144529566)
233
+ 7700 val loss 5.6503
234
+ 7700 val perplexity 284.3646
235
+ 7700 train 5.609683 (lr=6.0972e-06) (hash(x)=142669945)
236
+ 7800 val loss 5.6460
237
+ 7800 val perplexity 283.1537
238
+ 7800 train 5.704182 (lr=5.8107e-06) (hash(x)=161563764)
239
+ 7900 val loss 5.6443
240
+ 7900 val perplexity 282.6702
241
+ 7900 train 5.439746 (lr=5.5515e-06) (hash(x)=141516532)
242
+ 8000 val loss 5.6455
243
+ 8000 val perplexity 283.0064
244
+ 8000 train 5.515144 (lr=5.3203e-06) (hash(x)=152418457)
245
+ 8100 val loss 5.6398
246
+ 8100 val perplexity 281.4167
247
+ 8100 train 5.624706 (lr=5.1172e-06) (hash(x)=161109470)
248
+ 8200 val loss 5.6360
249
+ 8200 val perplexity 280.3446
250
+ 8200 train 5.758985 (lr=4.9425e-06) (hash(x)=154408822)
251
+ 8300 val loss 5.6342
252
+ 8300 val perplexity 279.8434
253
+ 8300 train 5.520042 (lr=4.7966e-06) (hash(x)=146431691)
254
+ 8400 val loss 5.6345
255
+ 8400 val perplexity 279.9092
256
+ 8400 train 5.412016 (lr=4.6796e-06) (hash(x)=150182527)
257
+ 8500 val loss 5.6286
258
+ 8500 val perplexity 278.2667
259
+ 8500 train 5.693805 (lr=4.5917e-06) (hash(x)=154828709)
260
+ 8600 val loss 5.6284
261
+ 8600 val perplexity 278.2239
262
+ 8600 train 5.539424 (lr=4.5330e-06) (hash(x)=154995428)
263
+ 8700 val loss 5.6285
264
+ 8700 val perplexity 278.2420
265
+ 8700 train 5.523919 (lr=4.5037e-06) (hash(x)=142111043)
266
+ 8749 val loss 5.6262
267
+ 8749 val perplexity 277.6001
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
attention_kindselective_n_heads4_seed1342/model_08749.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:50ff8744afab0bf33a1f184229c55f6f938e2f46b13c2d1aae2f9e9fc3be54cd
3
  size 92843394
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b784ac0421732561a9feaba30a9e9fef056d1e13bb74eb961fa766d7412a975
3
  size 92843394
attention_kindselective_n_heads4_seed1342/optimizer_08749.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51e2ecb9dfd6d5a8f8b29373903243ccda842f45d2f42561a1f91995ae236f24
3
  size 179406214
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a42419c28aa5e439c9bce28733a36c3eb5bf07c6395ea15a1c7f729048403a90
3
  size 179406214