andrew-healey commited on
Commit
3efed1e
·
verified ·
1 Parent(s): 29a86a3

Upload folder using huggingface_hub

Browse files
attention_kindselective_n_heads4_seed1341/args.json CHANGED
@@ -1 +1 @@
1
- {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_9/attention_kindselective_n_heads4_seed1341", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 4, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 8750, "warmup_steps": 500, "group": "wider_is_better_9", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1341, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 120, "total_batch_size": 61440, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 3e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "3e-5_61440_4_1341", "n_embd": 256}
 
1
+ {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_9/attention_kindselective_n_heads4_seed1341", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 4, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 8750, "warmup_steps": 500, "group": "wider_is_better_9", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1341, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 120, "total_batch_size": 61440, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 5.5e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "5.5e-5_61440_4_1341", "n_embd": 256}
attention_kindselective_n_heads4_seed1341/log2.txt CHANGED
@@ -1,534 +1,267 @@
1
  max_steps: 8750
2
- max_steps: 8750
3
- 0 val loss 11.3095
4
- 0 val perplexity 81594.2500
5
  0 val loss 11.3095
6
  0 val perplexity 81594.2500
7
- 0 train 11.313038 (lr=7.0000e-08) (hash(x)=145079536)
8
- 0 train 11.313038 (lr=6.0000e-08) (hash(x)=145079536)
9
- 100 val loss 10.1630
10
- 100 val perplexity 25926.1094
11
- 100 val loss 10.2343
12
- 100 val perplexity 27841.4590
13
- 100 train 10.156528 (lr=7.0700e-06) (hash(x)=155800595)
14
- 100 train 10.225678 (lr=6.0600e-06) (hash(x)=155800595)
15
- 200 val loss 9.7494
16
- 200 val perplexity 17144.4492
17
- 200 train 9.706524 (lr=1.4070e-05) (hash(x)=145606733)
18
- 200 val loss 9.7756
19
- 200 val perplexity 17599.2559
20
- 200 train 9.732876 (lr=1.2060e-05) (hash(x)=145606733)
21
- 300 val loss 8.8013
22
- 300 val perplexity 6643.1152
23
- 300 train 8.798988 (lr=2.1070e-05) (hash(x)=150367139)
24
- 300 val loss 8.9184
25
- 300 val perplexity 7468.2007
26
- 300 train 8.917364 (lr=1.8060e-05) (hash(x)=150367139)
27
- 400 val loss 7.9646
28
- 400 val perplexity 2877.3264
29
- 400 train 8.234877 (lr=2.8070e-05) (hash(x)=155747374)
30
- 400 val loss 7.9782
31
- 400 val perplexity 2916.8208
32
- 400 train 8.244305 (lr=2.4060e-05) (hash(x)=155747374)
33
- 500 val loss 7.6135
34
- 500 val perplexity 2025.2765
35
- 500 train 7.510717 (lr=3.5000e-05) (hash(x)=140604760)
36
- 500 val loss 7.5787
37
- 500 val perplexity 1955.9950
38
- 500 train 7.475321 (lr=3.0000e-05) (hash(x)=140604760)
39
- 600 val loss 7.4823
40
- 600 val perplexity 1776.2891
41
- 600 train 7.433376 (lr=3.4989e-05) (hash(x)=148404734)
42
- 600 val loss 7.4607
43
- 600 val perplexity 1738.3960
44
- 600 train 7.414741 (lr=2.9990e-05) (hash(x)=148404734)
45
- 700 val loss 7.4270
46
- 700 val perplexity 1680.7137
47
- 700 train 7.384828 (lr=3.4954e-05) (hash(x)=148115934)
48
- 700 val loss 7.3964
49
- 700 val perplexity 1630.0658
50
- 700 train 7.354086 (lr=2.9961e-05) (hash(x)=148115934)
51
- 800 val loss 7.3625
52
- 800 val perplexity 1575.7120
53
- 800 train 7.258959 (lr=3.4897e-05) (hash(x)=137464699)
54
- 800 val loss 7.3292
55
- 800 val perplexity 1524.1631
56
- 800 train 7.219800 (lr=2.9912e-05) (hash(x)=137464699)
57
- 900 val loss 7.3188
58
- 900 val perplexity 1508.4202
59
- 900 train 7.139495 (lr=3.4818e-05) (hash(x)=143886042)
60
- 900 val loss 7.2721
61
- 900 val perplexity 1439.5078
62
- 900 train 7.090878 (lr=2.9844e-05) (hash(x)=143886042)
63
- 1000 val loss 7.2571
64
- 1000 val perplexity 1418.1678
65
- 1000 train 7.259868 (lr=3.4715e-05) (hash(x)=163799796)
66
- 1000 val loss 7.2204
67
- 1000 val perplexity 1367.0923
68
- 1000 train 7.224873 (lr=2.9756e-05) (hash(x)=163799796)
69
- 1100 val loss 7.2004
70
- 1100 val perplexity 1339.9773
71
- 1100 train 7.151253 (lr=3.4591e-05) (hash(x)=144592844)
72
- 1100 val loss 7.1528
73
- 1100 val perplexity 1277.6487
74
- 1100 train 7.103980 (lr=2.9649e-05) (hash(x)=144592844)
75
- 1200 val loss 7.1496
76
- 1200 val perplexity 1273.5430
77
- 1200 train 7.510290 (lr=3.4444e-05) (hash(x)=204706354)
78
- 1200 val loss 7.0939
79
- 1200 val perplexity 1204.5594
80
- 1200 train 7.442509 (lr=2.9523e-05) (hash(x)=204706354)
81
- 1300 val loss 7.0922
82
- 1300 val perplexity 1202.5422
83
- 1300 train 7.020781 (lr=3.4275e-05) (hash(x)=150862210)
84
- 1300 val loss 7.0370
85
- 1300 val perplexity 1138.0204
86
- 1300 train 6.966258 (lr=2.9378e-05) (hash(x)=150862210)
87
- 1400 val loss 7.0167
88
- 1400 val perplexity 1115.1079
89
- 1400 train 6.995105 (lr=3.4084e-05) (hash(x)=147766811)
90
- 1400 val loss 6.9707
91
- 1400 val perplexity 1064.9824
92
- 1400 train 6.951100 (lr=2.9215e-05) (hash(x)=147766811)
93
- 1500 val loss 6.9622
94
- 1500 val perplexity 1056.0016
95
- 1500 train 6.773920 (lr=3.3872e-05) (hash(x)=135925327)
96
- 1500 val loss 6.9144
97
- 1500 val perplexity 1006.6741
98
- 1500 train 6.724929 (lr=2.9033e-05) (hash(x)=135925327)
99
- 1600 val loss 6.9194
100
- 1600 val perplexity 1011.7564
101
- 1600 train 6.889316 (lr=3.3638e-05) (hash(x)=160440642)
102
- 1600 val loss 6.8620
103
- 1600 val perplexity 955.2529
104
- 1600 train 6.835199 (lr=2.8833e-05) (hash(x)=160440642)
105
- 1700 val loss 6.8600
106
- 1700 val perplexity 953.3604
107
- 1700 train 6.953665 (lr=3.3384e-05) (hash(x)=151184106)
108
- 1700 val loss 6.7968
109
- 1700 val perplexity 894.9904
110
- 1700 train 6.890760 (lr=2.8615e-05) (hash(x)=151184106)
111
- 1800 val loss 6.8067
112
- 1800 val perplexity 903.8531
113
- 1800 train 6.697411 (lr=3.3109e-05) (hash(x)=148100580)
114
- 1800 val loss 6.7426
115
- 1800 val perplexity 847.7204
116
- 1800 train 6.634793 (lr=2.8379e-05) (hash(x)=148100580)
117
- 1900 val loss 6.7725
118
- 1900 val perplexity 873.4525
119
- 1900 train 6.604763 (lr=3.2814e-05) (hash(x)=149434659)
120
- 1900 val loss 6.6984
121
- 1900 val perplexity 811.1066
122
- 1900 train 6.530216 (lr=2.8127e-05) (hash(x)=149434659)
123
- 2000 val loss 6.7098
124
- 2000 val perplexity 820.4353
125
- 2000 train 6.604181 (lr=3.2500e-05) (hash(x)=152285486)
126
- 2000 val loss 6.6355
127
- 2000 val perplexity 761.6394
128
- 2000 train 6.533551 (lr=2.7857e-05) (hash(x)=152285486)
129
- 2100 val loss 6.6650
130
- 2100 val perplexity 784.4327
131
- 2100 train 6.413668 (lr=3.2166e-05) (hash(x)=144294295)
132
- 2100 val loss 6.5839
133
- 2100 val perplexity 723.3694
134
- 2100 train 6.331763 (lr=2.7571e-05) (hash(x)=144294295)
135
- 2200 val loss 6.6268
136
- 2200 val perplexity 755.0834
137
- 2200 train 6.659666 (lr=3.1813e-05) (hash(x)=175030215)
138
- 2200 val loss 6.5444
139
- 2200 val perplexity 695.3740
140
- 2200 train 6.575781 (lr=2.7269e-05) (hash(x)=175030215)
141
- 2300 val loss 6.5935
142
- 2300 val perplexity 730.3160
143
- 2300 train 6.395243 (lr=3.1443e-05) (hash(x)=150831428)
144
- 2300 val loss 6.5118
145
- 2300 val perplexity 673.0463
146
- 2300 train 6.314830 (lr=2.6951e-05) (hash(x)=150831428)
147
- 2400 val loss 6.5340
148
- 2400 val perplexity 688.1121
149
- 2400 train 6.689769 (lr=3.1054e-05) (hash(x)=140808297)
150
- 2400 val loss 6.4649
151
- 2400 val perplexity 642.2308
152
- 2400 train 6.624811 (lr=2.6618e-05) (hash(x)=140808297)
153
- 2500 val loss 6.4948
154
- 2500 val perplexity 661.7064
155
- 2500 train 6.484925 (lr=3.0649e-05) (hash(x)=153160275)
156
- 2500 val loss 6.4345
157
- 2500 val perplexity 622.9711
158
- 2500 train 6.418693 (lr=2.6270e-05) (hash(x)=153160275)
159
- 2600 val loss 6.4673
160
- 2600 val perplexity 643.7227
161
- 2600 train 6.227062 (lr=3.0227e-05) (hash(x)=133990623)
162
- 2600 val loss 6.4079
163
- 2600 val perplexity 606.6450
164
- 2600 train 6.165543 (lr=2.5909e-05) (hash(x)=133990623)
165
- 2700 val loss 6.4393
166
- 2700 val perplexity 625.9974
167
- 2700 train 6.414563 (lr=2.9789e-05) (hash(x)=142860944)
168
- 2700 val loss 6.3846
169
- 2700 val perplexity 592.6437
170
- 2700 train 6.360803 (lr=2.5533e-05) (hash(x)=142860944)
171
- 2800 val loss 6.4029
172
- 2800 val perplexity 603.5896
173
- 2800 train 6.432872 (lr=2.9336e-05) (hash(x)=137959511)
174
- 2800 val loss 6.3519
175
- 2800 val perplexity 573.5657
176
- 2800 train 6.378989 (lr=2.5145e-05) (hash(x)=137959511)
177
- 2900 val loss 6.3718
178
- 2900 val perplexity 585.1338
179
- 2900 train 6.311280 (lr=2.8868e-05) (hash(x)=147009873)
180
- 2900 val loss 6.3265
181
- 2900 val perplexity 559.1986
182
- 2900 train 6.269898 (lr=2.4744e-05) (hash(x)=147009873)
183
- 3000 val loss 6.3476
184
- 3000 val perplexity 571.1144
185
- 3000 train 6.218025 (lr=2.8386e-05) (hash(x)=158264841)
186
- 3000 val loss 6.3063
187
- 3000 val perplexity 548.0392
188
- 3000 train 6.173168 (lr=2.4331e-05) (hash(x)=158264841)
189
- 3100 val loss 6.3292
190
- 3100 val perplexity 560.6805
191
- 3100 train 6.188198 (lr=2.7891e-05) (hash(x)=139232251)
192
- 3100 val loss 6.2925
193
- 3100 val perplexity 540.5228
194
- 3100 train 6.155460 (lr=2.3906e-05) (hash(x)=139232251)
195
- 3200 val loss 6.2932
196
- 3200 val perplexity 540.8548
197
- 3200 train 6.341791 (lr=2.7383e-05) (hash(x)=153436104)
198
- 3200 val loss 6.2591
199
- 3200 val perplexity 522.7567
200
- 3200 train 6.305866 (lr=2.3471e-05) (hash(x)=153436104)
201
- 3300 val loss 6.2764
202
- 3300 val perplexity 531.8832
203
- 3300 train 6.092842 (lr=2.6864e-05) (hash(x)=149681831)
204
- 3300 val loss 6.2423
205
- 3300 val perplexity 514.0164
206
- 3300 train 6.057350 (lr=2.3026e-05) (hash(x)=149681831)
207
- 3400 val loss 6.2633
208
- 3400 val perplexity 524.9594
209
- 3400 train 6.166925 (lr=2.6333e-05) (hash(x)=168425516)
210
- 3400 val loss 6.2299
211
- 3400 val perplexity 507.6933
212
- 3400 train 6.136177 (lr=2.2572e-05) (hash(x)=168425516)
213
- 3500 val loss 6.2348
214
- 3500 val perplexity 510.2036
215
- 3500 train 6.263444 (lr=2.5793e-05) (hash(x)=163104338)
216
- 3500 val loss 6.2078
217
- 3500 val perplexity 496.5847
218
- 3500 train 6.231250 (lr=2.2108e-05) (hash(x)=163104338)
219
- 3600 val loss 6.2157
220
- 3600 val perplexity 500.5353
221
- 3600 train 6.190797 (lr=2.5243e-05) (hash(x)=165109772)
222
- 3600 val loss 6.1890
223
- 3600 val perplexity 487.3521
224
- 3600 train 6.160804 (lr=2.1637e-05) (hash(x)=165109772)
225
- 3700 val loss 6.2076
226
- 3700 val perplexity 496.5286
227
- 3700 train 5.940827 (lr=2.4684e-05) (hash(x)=153420306)
228
- 3700 val loss 6.1816
229
- 3700 val perplexity 483.7829
230
- 3700 train 5.910549 (lr=2.1158e-05) (hash(x)=153420306)
231
- 3800 val loss 6.1786
232
- 3800 val perplexity 482.2998
233
- 3800 train 6.204558 (lr=2.4117e-05) (hash(x)=160168863)
234
- 3800 val loss 6.1568
235
- 3800 val perplexity 471.9337
236
- 3800 train 6.183776 (lr=2.0672e-05) (hash(x)=160168863)
237
- 3900 val loss 6.1666
238
- 3900 val perplexity 476.5572
239
- 3900 train 6.095548 (lr=2.3543e-05) (hash(x)=153906073)
240
- 3900 val loss 6.1428
241
- 3900 val perplexity 465.3320
242
- 3900 train 6.070639 (lr=2.0180e-05) (hash(x)=153906073)
243
- 4000 val loss 6.1527
244
- 4000 val perplexity 469.9752
245
- 4000 train 6.081241 (lr=2.2963e-05) (hash(x)=151055067)
246
- 4000 val loss 6.1338
247
- 4000 val perplexity 461.1803
248
- 4000 train 6.064647 (lr=1.9683e-05) (hash(x)=151055067)
249
- 4100 val loss 6.1350
250
- 4100 val perplexity 461.7586
251
- 4100 train 6.116509 (lr=2.2378e-05) (hash(x)=149629830)
252
- 4100 val loss 6.1163
253
- 4100 val perplexity 453.1846
254
- 4100 train 6.090425 (lr=1.9181e-05) (hash(x)=149629830)
255
- 4200 val loss 6.1249
256
- 4200 val perplexity 457.1209
257
- 4200 train 6.003567 (lr=2.1788e-05) (hash(x)=143101381)
258
- 4200 val loss 6.1033
259
- 4200 val perplexity 447.3311
260
- 4200 train 5.984151 (lr=1.8675e-05) (hash(x)=143101381)
261
- 4300 val loss 6.1115
262
- 4300 val perplexity 451.0039
263
- 4300 train 6.010469 (lr=2.1194e-05) (hash(x)=149712044)
264
- 4300 val loss 6.0949
265
- 4300 val perplexity 443.5709
266
- 4300 train 5.993714 (lr=1.8166e-05) (hash(x)=149712044)
267
- 4400 val loss 6.0941
268
- 4400 val perplexity 443.2506
269
- 4400 train 6.020243 (lr=2.0598e-05) (hash(x)=153446449)
270
- 4400 val loss 6.0809
271
- 4400 val perplexity 437.4159
272
- 4400 train 6.007243 (lr=1.7655e-05) (hash(x)=153446449)
273
- 4500 val loss 6.0824
274
- 4500 val perplexity 438.0891
275
- 4500 train 6.027428 (lr=1.9999e-05) (hash(x)=146086947)
276
- 4500 val loss 6.0681
277
- 4500 val perplexity 431.8431
278
- 4500 train 6.013597 (lr=1.7142e-05) (hash(x)=146086947)
279
- 4600 val loss 6.0720
280
- 4600 val perplexity 433.5492
281
- 4600 train 5.910559 (lr=1.9400e-05) (hash(x)=153800173)
282
- 4600 val loss 6.0581
283
- 4600 val perplexity 427.5742
284
- 4600 train 5.898770 (lr=1.6629e-05) (hash(x)=153800173)
285
- 4700 val loss 6.0597
286
- 4700 val perplexity 428.2572
287
- 4700 train 6.110185 (lr=1.8800e-05) (hash(x)=155962726)
288
- 4700 val loss 6.0479
289
- 4700 val perplexity 423.2194
290
- 4700 train 6.100060 (lr=1.6114e-05) (hash(x)=155962726)
291
- 4800 val loss 6.0515
292
- 4800 val perplexity 424.7603
293
- 4800 train 6.275963 (lr=1.8201e-05) (hash(x)=142045616)
294
- 4800 val loss 6.0388
295
- 4800 val perplexity 419.4092
296
- 4800 train 6.270424 (lr=1.5601e-05) (hash(x)=142045616)
297
- 4900 val loss 6.0377
298
- 4900 val perplexity 418.9119
299
- 4900 train 6.005858 (lr=1.7604e-05) (hash(x)=143418248)
300
- 4900 val loss 6.0267
301
- 4900 val perplexity 414.3510
302
- 4900 train 5.989192 (lr=1.5089e-05) (hash(x)=143418248)
303
- 5000 val loss 6.0290
304
- 5000 val perplexity 415.2997
305
- 5000 train 5.954850 (lr=1.7009e-05) (hash(x)=145789790)
306
- 5000 val loss 6.0175
307
- 5000 val perplexity 410.5454
308
- 5000 train 5.938465 (lr=1.4579e-05) (hash(x)=145789790)
309
- 5100 val loss 6.0238
310
- 5100 val perplexity 413.1402
311
- 5100 train 5.980156 (lr=1.6417e-05) (hash(x)=137795633)
312
- 5100 val loss 6.0147
313
- 5100 val perplexity 409.3998
314
- 5100 train 5.970935 (lr=1.4071e-05) (hash(x)=137795633)
315
- 5200 val loss 6.0137
316
- 5200 val perplexity 408.9803
317
- 5200 train 6.178001 (lr=1.5829e-05) (hash(x)=148907132)
318
- 5200 val loss 6.0025
319
- 5200 val perplexity 404.4547
320
- 5200 train 6.166307 (lr=1.3568e-05) (hash(x)=148907132)
321
- 5300 val loss 6.0028
322
- 5300 val perplexity 404.5762
323
- 5300 train 5.882075 (lr=1.5246e-05) (hash(x)=152343580)
324
- 5300 val loss 5.9958
325
- 5300 val perplexity 401.7314
326
- 5300 train 5.877028 (lr=1.3068e-05) (hash(x)=152343580)
327
- 5400 val loss 5.9968
328
- 5400 val perplexity 402.1420
329
- 5400 train 5.843998 (lr=1.4669e-05) (hash(x)=148578264)
330
- 5400 val loss 5.9889
331
- 5400 val perplexity 398.9933
332
- 5400 train 5.840799 (lr=1.2573e-05) (hash(x)=148578264)
333
- 5500 val loss 5.9842
334
- 5500 val perplexity 397.1019
335
- 5500 train 6.141342 (lr=1.4099e-05) (hash(x)=145635833)
336
- 5500 val loss 5.9769
337
- 5500 val perplexity 394.1999
338
- 5500 train 6.139304 (lr=1.2085e-05) (hash(x)=145635833)
339
- 5600 val loss 5.9796
340
- 5600 val perplexity 395.2935
341
- 5600 train 5.955497 (lr=1.3536e-05) (hash(x)=156337844)
342
- 5600 val loss 5.9735
343
- 5600 val perplexity 392.8972
344
- 5600 train 5.955674 (lr=1.1602e-05) (hash(x)=156337844)
345
- 5700 val loss 5.9731
346
- 5700 val perplexity 392.7069
347
- 5700 train 5.856874 (lr=1.2981e-05) (hash(x)=147168506)
348
- 5700 val loss 5.9688
349
- 5700 val perplexity 391.0504
350
- 5700 train 5.857741 (lr=1.1127e-05) (hash(x)=147168506)
351
- 5800 val loss 5.9644
352
- 5800 val perplexity 389.3232
353
- 5800 train 6.032385 (lr=1.2436e-05) (hash(x)=159566920)
354
- 5800 val loss 5.9585
355
- 5800 val perplexity 387.0450
356
- 5800 train 6.023722 (lr=1.0659e-05) (hash(x)=159566920)
357
- 5900 val loss 5.9578
358
- 5900 val perplexity 386.7631
359
- 5900 train 6.025391 (lr=1.1900e-05) (hash(x)=158273929)
360
- 5900 val loss 5.9536
361
- 5900 val perplexity 385.1432
362
- 5900 train 6.027411 (lr=1.0200e-05) (hash(x)=158273929)
363
- 6000 val loss 5.9532
364
- 6000 val perplexity 384.9910
365
- 6000 train 5.921682 (lr=1.1375e-05) (hash(x)=156649749)
366
- 6000 val loss 5.9499
367
- 6000 val perplexity 383.6991
368
- 6000 train 5.916996 (lr=9.7500e-06) (hash(x)=156649749)
369
- 6100 val loss 5.9437
370
- 6100 val perplexity 381.3258
371
- 6100 train 5.860867 (lr=1.0861e-05) (hash(x)=146812388)
372
- 6100 val loss 5.9397
373
- 6100 val perplexity 379.8224
374
- 6100 train 5.858076 (lr=9.3098e-06) (hash(x)=146812388)
375
- 6200 val loss 5.9380
376
- 6200 val perplexity 379.1947
377
- 6200 train 5.760242 (lr=1.0360e-05) (hash(x)=143522146)
378
- 6200 val loss 5.9360
379
- 6200 val perplexity 378.4272
380
- 6200 train 5.755511 (lr=8.8800e-06) (hash(x)=143522146)
381
- 6300 val loss 5.9354
382
- 6300 val perplexity 378.1945
383
- 6300 train 5.772961 (lr=9.8715e-06) (hash(x)=150124474)
384
- 6300 val loss 5.9314
385
- 6300 val perplexity 376.6789
386
- 6300 train 5.768746 (lr=8.4613e-06) (hash(x)=150124474)
387
- 6400 val loss 5.9273
388
- 6400 val perplexity 375.1227
389
- 6400 train 5.950421 (lr=9.3966e-06) (hash(x)=141242117)
390
- 6400 val loss 5.9284
391
- 6400 val perplexity 375.5501
392
- 6400 train 5.955993 (lr=8.0542e-06) (hash(x)=141242117)
393
- 6500 val loss 5.9207
394
- 6500 val perplexity 372.6563
395
- 6500 train 5.801312 (lr=8.9359e-06) (hash(x)=143529762)
396
- 6500 val loss 5.9189
397
- 6500 val perplexity 371.9922
398
- 6500 train 5.801332 (lr=7.6594e-06) (hash(x)=143529762)
399
- 6600 val loss 5.9155
400
- 6600 val perplexity 370.7257
401
- 6600 train 5.697849 (lr=8.4903e-06) (hash(x)=136948374)
402
- 6600 val loss 5.9148
403
- 6600 val perplexity 370.4690
404
- 6600 train 5.696297 (lr=7.2774e-06) (hash(x)=136948374)
405
- 6700 val loss 5.9141
406
- 6700 val perplexity 370.2074
407
- 6700 train 5.733417 (lr=8.0602e-06) (hash(x)=146268592)
408
- 6700 val loss 5.9131
409
- 6700 val perplexity 369.8498
410
- 6700 train 5.730753 (lr=6.9087e-06) (hash(x)=146268592)
411
- 6800 val loss 5.9055
412
- 6800 val perplexity 367.0687
413
- 6800 train 5.868842 (lr=7.6463e-06) (hash(x)=152676836)
414
- 6800 val loss 5.9058
415
- 6800 val perplexity 367.1629
416
- 6800 train 5.863817 (lr=6.5540e-06) (hash(x)=152676836)
417
- 6900 val loss 5.8984
418
- 6900 val perplexity 364.4714
419
- 6900 train 5.812114 (lr=7.2493e-06) (hash(x)=134657776)
420
- 6900 val loss 5.8995
421
- 6900 val perplexity 364.8668
422
- 6900 train 5.811509 (lr=6.2137e-06) (hash(x)=134657776)
423
- 7000 val loss 5.8971
424
- 7000 val perplexity 363.9915
425
- 7000 train 5.862196 (lr=6.8697e-06) (hash(x)=166721861)
426
- 7000 val loss 5.8977
427
- 7000 val perplexity 364.2080
428
- 7000 train 5.862389 (lr=5.8883e-06) (hash(x)=166721861)
429
- 7100 val loss 5.8936
430
- 7100 val perplexity 362.6912
431
- 7100 train 5.697666 (lr=6.5080e-06) (hash(x)=135496702)
432
- 7100 val loss 5.8958
433
- 7100 val perplexity 363.5053
434
- 7100 train 5.697424 (lr=5.5783e-06) (hash(x)=135496702)
435
- 7200 val loss 5.8906
436
- 7200 val perplexity 361.6083
437
- 7200 train 5.993858 (lr=6.1648e-06) (hash(x)=155567461)
438
- 7200 val loss 5.8928
439
- 7200 val perplexity 362.4025
440
- 7200 train 5.994301 (lr=5.2841e-06) (hash(x)=155567461)
441
- 7300 val loss 5.8846
442
- 7300 val perplexity 359.4441
443
- 7300 train 5.676643 (lr=5.8405e-06) (hash(x)=142803829)
444
- 7300 val loss 5.8867
445
- 7300 val perplexity 360.1984
446
- 7300 train 5.685822 (lr=5.0062e-06) (hash(x)=142803829)
447
- 7400 val loss 5.8824
448
- 7400 val perplexity 358.6774
449
- 7400 train 5.699541 (lr=5.5357e-06) (hash(x)=145294178)
450
- 7400 val loss 5.8848
451
- 7400 val perplexity 359.5374
452
- 7400 train 5.707546 (lr=4.7449e-06) (hash(x)=145294178)
453
- 7500 val loss 5.8797
454
- 7500 val perplexity 357.6940
455
- 7500 train 5.642534 (lr=5.2508e-06) (hash(x)=150573713)
456
- 7500 val loss 5.8828
457
- 7500 val perplexity 358.8115
458
- 7500 train 5.639037 (lr=4.5007e-06) (hash(x)=150573713)
459
- 7600 val loss 5.8773
460
- 7600 val perplexity 356.8476
461
- 7600 train 5.904793 (lr=4.9862e-06) (hash(x)=142771511)
462
- 7600 val loss 5.8806
463
- 7600 val perplexity 358.0397
464
- 7600 train 5.905676 (lr=4.2739e-06) (hash(x)=142771511)
465
- 7700 val loss 5.8716
466
- 7700 val perplexity 354.8214
467
- 7700 train 5.766538 (lr=4.7423e-06) (hash(x)=143602175)
468
- 7700 val loss 5.8750
469
- 7700 val perplexity 356.0248
470
- 7700 train 5.774282 (lr=4.0648e-06) (hash(x)=143602175)
471
- 7800 val loss 5.8698
472
- 7800 val perplexity 354.1783
473
- 7800 train 5.892739 (lr=4.5194e-06) (hash(x)=152379862)
474
- 7800 val loss 5.8722
475
- 7800 val perplexity 355.0424
476
- 7800 train 5.895572 (lr=3.8738e-06) (hash(x)=152379862)
477
- 7900 val loss 5.8686
478
- 7900 val perplexity 353.7581
479
- 7900 train 5.703426 (lr=4.3179e-06) (hash(x)=146655921)
480
- 7900 val loss 5.8714
481
- 7900 val perplexity 354.7439
482
- 7900 train 5.703874 (lr=3.7010e-06) (hash(x)=146655921)
483
- 8000 val loss 5.8693
484
- 8000 val perplexity 353.9875
485
- 8000 train 5.982882 (lr=4.1380e-06) (hash(x)=148262482)
486
- 8000 val loss 5.8727
487
- 8000 val perplexity 355.2155
488
- 8000 train 5.985026 (lr=3.5468e-06) (hash(x)=148262482)
489
- 8100 val loss 5.8610
490
- 8100 val perplexity 351.0891
491
- 8100 train 5.792439 (lr=3.9800e-06) (hash(x)=147683655)
492
- 8100 val loss 5.8643
493
- 8100 val perplexity 352.2198
494
- 8100 train 5.793932 (lr=3.4114e-06) (hash(x)=147683655)
495
- 8200 val loss 5.8610
496
- 8200 val perplexity 351.0704
497
- 8200 train 5.954844 (lr=3.8442e-06) (hash(x)=157312987)
498
- 8200 val loss 5.8644
499
- 8200 val perplexity 352.2767
500
- 8200 train 5.959354 (lr=3.2950e-06) (hash(x)=157312987)
501
- 8300 val loss 5.8584
502
- 8300 val perplexity 350.1542
503
- 8300 train 5.827306 (lr=3.7307e-06) (hash(x)=141107543)
504
- 8300 val loss 5.8617
505
- 8300 val perplexity 351.3268
506
- 8300 train 5.833569 (lr=3.1977e-06) (hash(x)=141107543)
507
- 8400 val loss 5.8546
508
- 8400 val perplexity 348.8233
509
- 8400 train 5.829935 (lr=3.6397e-06) (hash(x)=141323024)
510
- 8400 val loss 5.8581
511
- 8400 val perplexity 350.0630
512
- 8400 train 5.836152 (lr=3.1197e-06) (hash(x)=141323024)
513
- 8500 val loss 5.8523
514
- 8500 val perplexity 348.0277
515
- 8500 train 5.828456 (lr=3.5713e-06) (hash(x)=150696521)
516
- 8500 val loss 5.8561
517
- 8500 val perplexity 349.3654
518
- 8500 train 5.834352 (lr=3.0611e-06) (hash(x)=150696521)
519
- 8600 val loss 5.8522
520
- 8600 val perplexity 348.0011
521
- 8600 train 5.811702 (lr=3.5257e-06) (hash(x)=162288191)
522
- 8600 val loss 5.8554
523
- 8600 val perplexity 349.1052
524
- 8600 train 5.809893 (lr=3.0220e-06) (hash(x)=162288191)
525
- 8700 val loss 5.8482
526
- 8700 val perplexity 346.6077
527
- 8700 train 5.718023 (lr=3.5029e-06) (hash(x)=152860941)
528
- 8700 val loss 5.8523
529
- 8700 val perplexity 348.0176
530
- 8700 train 5.726460 (lr=3.0024e-06) (hash(x)=152860941)
531
- 8749 val loss 5.8467
532
- 8749 val perplexity 346.0782
533
- 8749 val loss 5.8518
534
- 8749 val perplexity 347.8583
 
1
  max_steps: 8750
 
 
 
2
  0 val loss 11.3095
3
  0 val perplexity 81594.2500
4
+ 0 train 11.313040 (lr=1.1000e-07) (hash(x)=145079536)
5
+ 100 val loss 10.0285
6
+ 100 val perplexity 22663.5449
7
+ 100 train 10.022821 (lr=1.1110e-05) (hash(x)=155800595)
8
+ 200 val loss 9.3391
9
+ 200 val perplexity 11373.7100
10
+ 200 train 9.288088 (lr=2.2110e-05) (hash(x)=145606733)
11
+ 300 val loss 8.0597
12
+ 300 val perplexity 3164.1838
13
+ 300 train 8.040195 (lr=3.3110e-05) (hash(x)=150367139)
14
+ 400 val loss 7.5562
15
+ 400 val perplexity 1912.4974
16
+ 400 train 7.863613 (lr=4.4110e-05) (hash(x)=155747374)
17
+ 500 val loss 7.4240
18
+ 500 val perplexity 1675.6675
19
+ 500 train 7.315066 (lr=5.5000e-05) (hash(x)=140604760)
20
+ 600 val loss 7.3587
21
+ 600 val perplexity 1569.7710
22
+ 600 train 7.302488 (lr=5.4982e-05) (hash(x)=148404734)
23
+ 700 val loss 7.3047
24
+ 700 val perplexity 1487.2728
25
+ 700 train 7.262489 (lr=5.4928e-05) (hash(x)=148115934)
26
+ 800 val loss 7.2321
27
+ 800 val perplexity 1383.0587
28
+ 800 train 7.115292 (lr=5.4839e-05) (hash(x)=137464699)
29
+ 900 val loss 7.1679
30
+ 900 val perplexity 1297.1080
31
+ 900 train 6.982425 (lr=5.4713e-05) (hash(x)=143886042)
32
+ 1000 val loss 7.1077
33
+ 1000 val perplexity 1221.3217
34
+ 1000 train 7.107786 (lr=5.4553e-05) (hash(x)=163799796)
35
+ 1100 val loss 7.0422
36
+ 1100 val perplexity 1143.8781
37
+ 1100 train 6.998779 (lr=5.4357e-05) (hash(x)=144592844)
38
+ 1200 val loss 6.9882
39
+ 1200 val perplexity 1083.7599
40
+ 1200 train 7.327436 (lr=5.4126e-05) (hash(x)=204706354)
41
+ 1300 val loss 6.9446
42
+ 1300 val perplexity 1037.5281
43
+ 1300 train 6.878752 (lr=5.3860e-05) (hash(x)=150862210)
44
+ 1400 val loss 6.8865
45
+ 1400 val perplexity 978.9376
46
+ 1400 train 6.871954 (lr=5.3561e-05) (hash(x)=147766811)
47
+ 1500 val loss 6.8288
48
+ 1500 val perplexity 924.0453
49
+ 1500 train 6.639542 (lr=5.3227e-05) (hash(x)=135925327)
50
+ 1600 val loss 6.7740
51
+ 1600 val perplexity 874.8393
52
+ 1600 train 6.746453 (lr=5.2860e-05) (hash(x)=160440642)
53
+ 1700 val loss 6.7164
54
+ 1700 val perplexity 825.8420
55
+ 1700 train 6.815380 (lr=5.2461e-05) (hash(x)=151184106)
56
+ 1800 val loss 6.6585
57
+ 1800 val perplexity 779.3540
58
+ 1800 train 6.557573 (lr=5.2029e-05) (hash(x)=148100580)
59
+ 1900 val loss 6.6249
60
+ 1900 val perplexity 753.6406
61
+ 1900 train 6.457758 (lr=5.1565e-05) (hash(x)=149434659)
62
+ 2000 val loss 6.5724
63
+ 2000 val perplexity 715.0803
64
+ 2000 train 6.471416 (lr=5.1071e-05) (hash(x)=152285486)
65
+ 2100 val loss 6.5395
66
+ 2100 val perplexity 691.9631
67
+ 2100 train 6.288988 (lr=5.0547e-05) (hash(x)=144294295)
68
+ 2200 val loss 6.5062
69
+ 2200 val perplexity 669.3031
70
+ 2200 train 6.534094 (lr=4.9993e-05) (hash(x)=175030215)
71
+ 2300 val loss 6.4785
72
+ 2300 val perplexity 651.0085
73
+ 2300 train 6.280591 (lr=4.9410e-05) (hash(x)=150831428)
74
+ 2400 val loss 6.4330
75
+ 2400 val perplexity 622.0258
76
+ 2400 train 6.595540 (lr=4.8800e-05) (hash(x)=140808297)
77
+ 2500 val loss 6.4295
78
+ 2500 val perplexity 619.8776
79
+ 2500 train 6.419309 (lr=4.8162e-05) (hash(x)=153160275)
80
+ 2600 val loss 6.4029
81
+ 2600 val perplexity 603.5640
82
+ 2600 train 6.176099 (lr=4.7499e-05) (hash(x)=133990623)
83
+ 2700 val loss 6.3730
84
+ 2700 val perplexity 585.8351
85
+ 2700 train 6.347503 (lr=4.6811e-05) (hash(x)=142860944)
86
+ 2800 val loss 6.3438
87
+ 2800 val perplexity 568.9605
88
+ 2800 train 6.365329 (lr=4.6099e-05) (hash(x)=137959511)
89
+ 2900 val loss 6.3173
90
+ 2900 val perplexity 554.1024
91
+ 2900 train 6.263561 (lr=4.5364e-05) (hash(x)=147009873)
92
+ 3000 val loss 6.2962
93
+ 3000 val perplexity 542.5182
94
+ 3000 train 6.152777 (lr=4.4606e-05) (hash(x)=158264841)
95
+ 3100 val loss 6.2885
96
+ 3100 val perplexity 538.3243
97
+ 3100 train 6.133510 (lr=4.3828e-05) (hash(x)=139232251)
98
+ 3200 val loss 6.2569
99
+ 3200 val perplexity 521.6201
100
+ 3200 train 6.301063 (lr=4.3031e-05) (hash(x)=153436104)
101
+ 3300 val loss 6.2290
102
+ 3300 val perplexity 507.2550
103
+ 3300 train 6.049569 (lr=4.2215e-05) (hash(x)=149681831)
104
+ 3400 val loss 6.2229
105
+ 3400 val perplexity 504.1629
106
+ 3400 train 6.117395 (lr=4.1381e-05) (hash(x)=168425516)
107
+ 3500 val loss 6.1881
108
+ 3500 val perplexity 486.9261
109
+ 3500 train 6.212366 (lr=4.0532e-05) (hash(x)=163104338)
110
+ 3600 val loss 6.1671
111
+ 3600 val perplexity 476.8029
112
+ 3600 train 6.143349 (lr=3.9667e-05) (hash(x)=165109772)
113
+ 3700 val loss 6.1626
114
+ 3700 val perplexity 474.6586
115
+ 3700 train 5.890706 (lr=3.8789e-05) (hash(x)=153420306)
116
+ 3800 val loss 6.1309
117
+ 3800 val perplexity 459.8583
118
+ 3800 train 6.154771 (lr=3.7898e-05) (hash(x)=160168863)
119
+ 3900 val loss 6.1152
120
+ 3900 val perplexity 452.6844
121
+ 3900 train 6.032316 (lr=3.6996e-05) (hash(x)=153906073)
122
+ 4000 val loss 6.1059
123
+ 4000 val perplexity 448.4742
124
+ 4000 train 6.034504 (lr=3.6085e-05) (hash(x)=151055067)
125
+ 4100 val loss 6.0870
126
+ 4100 val perplexity 440.1180
127
+ 4100 train 6.061316 (lr=3.5165e-05) (hash(x)=149629830)
128
+ 4200 val loss 6.0718
129
+ 4200 val perplexity 433.4450
130
+ 4200 train 5.946362 (lr=3.4238e-05) (hash(x)=143101381)
131
+ 4300 val loss 6.0579
132
+ 4300 val perplexity 427.4625
133
+ 4300 train 5.960875 (lr=3.3305e-05) (hash(x)=149712044)
134
+ 4400 val loss 6.0403
135
+ 4400 val perplexity 419.9999
136
+ 4400 train 5.964753 (lr=3.2368e-05) (hash(x)=153446449)
137
+ 4500 val loss 6.0242
138
+ 4500 val perplexity 413.3047
139
+ 4500 train 5.972738 (lr=3.1428e-05) (hash(x)=146086947)
140
+ 4600 val loss 6.0203
141
+ 4600 val perplexity 411.6975
142
+ 4600 train 5.859571 (lr=3.0486e-05) (hash(x)=153800173)
143
+ 4700 val loss 6.0057
144
+ 4700 val perplexity 405.7249
145
+ 4700 train 6.063126 (lr=2.9543e-05) (hash(x)=155962726)
146
+ 4800 val loss 5.9915
147
+ 4800 val perplexity 400.0210
148
+ 4800 train 6.226341 (lr=2.8602e-05) (hash(x)=142045616)
149
+ 4900 val loss 5.9758
150
+ 4900 val perplexity 393.7772
151
+ 4900 train 5.946213 (lr=2.7663e-05) (hash(x)=143418248)
152
+ 5000 val loss 5.9665
153
+ 5000 val perplexity 390.1279
154
+ 5000 train 5.884785 (lr=2.6728e-05) (hash(x)=145789790)
155
+ 5100 val loss 5.9587
156
+ 5100 val perplexity 387.1258
157
+ 5100 train 5.927770 (lr=2.5798e-05) (hash(x)=137795633)
158
+ 5200 val loss 5.9446
159
+ 5200 val perplexity 381.7025
160
+ 5200 train 6.104008 (lr=2.4874e-05) (hash(x)=148907132)
161
+ 5300 val loss 5.9428
162
+ 5300 val perplexity 380.9910
163
+ 5300 train 5.824908 (lr=2.3958e-05) (hash(x)=152343580)
164
+ 5400 val loss 5.9278
165
+ 5400 val perplexity 375.3416
166
+ 5400 train 5.779144 (lr=2.3051e-05) (hash(x)=148578264)
167
+ 5500 val loss 5.9159
168
+ 5500 val perplexity 370.8882
169
+ 5500 train 6.081324 (lr=2.2155e-05) (hash(x)=145635833)
170
+ 5600 val loss 5.9097
171
+ 5600 val perplexity 368.5929
172
+ 5600 train 5.884354 (lr=2.1271e-05) (hash(x)=156337844)
173
+ 5700 val loss 5.9002
174
+ 5700 val perplexity 365.1134
175
+ 5700 train 5.783232 (lr=2.0399e-05) (hash(x)=147168506)
176
+ 5800 val loss 5.8906
177
+ 5800 val perplexity 361.6292
178
+ 5800 train 5.954978 (lr=1.9542e-05) (hash(x)=159566920)
179
+ 5900 val loss 5.8862
180
+ 5900 val perplexity 360.0167
181
+ 5900 train 5.948883 (lr=1.8700e-05) (hash(x)=158273929)
182
+ 6000 val loss 5.8794
183
+ 6000 val perplexity 357.5957
184
+ 6000 train 5.850380 (lr=1.7875e-05) (hash(x)=156649749)
185
+ 6100 val loss 5.8706
186
+ 6100 val perplexity 354.4674
187
+ 6100 train 5.786806 (lr=1.7068e-05) (hash(x)=146812388)
188
+ 6200 val loss 5.8642
189
+ 6200 val perplexity 352.1928
190
+ 6200 train 5.684584 (lr=1.6280e-05) (hash(x)=143522146)
191
+ 6300 val loss 5.8598
192
+ 6300 val perplexity 350.6573
193
+ 6300 train 5.697289 (lr=1.5512e-05) (hash(x)=150124474)
194
+ 6400 val loss 5.8556
195
+ 6400 val perplexity 349.1890
196
+ 6400 train 5.881581 (lr=1.4766e-05) (hash(x)=141242117)
197
+ 6500 val loss 5.8434
198
+ 6500 val perplexity 344.9429
199
+ 6500 train 5.724736 (lr=1.4042e-05) (hash(x)=143529762)
200
+ 6600 val loss 5.8384
201
+ 6600 val perplexity 343.2130
202
+ 6600 train 5.621650 (lr=1.3342e-05) (hash(x)=136948374)
203
+ 6700 val loss 5.8354
204
+ 6700 val perplexity 342.2034
205
+ 6700 train 5.649374 (lr=1.2666e-05) (hash(x)=146268592)
206
+ 6800 val loss 5.8289
207
+ 6800 val perplexity 339.9703
208
+ 6800 train 5.787588 (lr=1.2016e-05) (hash(x)=152676836)
209
+ 6900 val loss 5.8204
210
+ 6900 val perplexity 337.0951
211
+ 6900 train 5.736982 (lr=1.1392e-05) (hash(x)=134657776)
212
+ 7000 val loss 5.8177
213
+ 7000 val perplexity 336.1875
214
+ 7000 train 5.775075 (lr=1.0795e-05) (hash(x)=166721861)
215
+ 7100 val loss 5.8143
216
+ 7100 val perplexity 335.0538
217
+ 7100 train 5.619752 (lr=1.0227e-05) (hash(x)=135496702)
218
+ 7200 val loss 5.8106
219
+ 7200 val perplexity 333.8330
220
+ 7200 train 5.914602 (lr=9.6875e-06) (hash(x)=155567461)
221
+ 7300 val loss 5.8045
222
+ 7300 val perplexity 331.7833
223
+ 7300 train 5.601978 (lr=9.1780e-06) (hash(x)=142803829)
224
+ 7400 val loss 5.8025
225
+ 7400 val perplexity 331.1190
226
+ 7400 train 5.623451 (lr=8.6990e-06) (hash(x)=145294178)
227
+ 7500 val loss 5.7976
228
+ 7500 val perplexity 329.5175
229
+ 7500 train 5.558932 (lr=8.2513e-06) (hash(x)=150573713)
230
+ 7600 val loss 5.7957
231
+ 7600 val perplexity 328.8690
232
+ 7600 train 5.817007 (lr=7.8355e-06) (hash(x)=142771511)
233
+ 7700 val loss 5.7898
234
+ 7700 val perplexity 326.9609
235
+ 7700 train 5.689264 (lr=7.4522e-06) (hash(x)=143602175)
236
+ 7800 val loss 5.7861
237
+ 7800 val perplexity 325.7462
238
+ 7800 train 5.816852 (lr=7.1019e-06) (hash(x)=152379862)
239
+ 7900 val loss 5.7852
240
+ 7900 val perplexity 325.4419
241
+ 7900 train 5.620042 (lr=6.7852e-06) (hash(x)=146655921)
242
+ 8000 val loss 5.7856
243
+ 8000 val perplexity 325.5628
244
+ 8000 train 5.900315 (lr=6.5025e-06) (hash(x)=148262482)
245
+ 8100 val loss 5.7776
246
+ 8100 val perplexity 322.9937
247
+ 8100 train 5.713513 (lr=6.2543e-06) (hash(x)=147683655)
248
+ 8200 val loss 5.7774
249
+ 8200 val perplexity 322.9288
250
+ 8200 train 5.876583 (lr=6.0408e-06) (hash(x)=157312987)
251
+ 8300 val loss 5.7743
252
+ 8300 val perplexity 321.9338
253
+ 8300 train 5.736997 (lr=5.8625e-06) (hash(x)=141107543)
254
+ 8400 val loss 5.7698
255
+ 8400 val perplexity 320.4694
256
+ 8400 train 5.747867 (lr=5.7195e-06) (hash(x)=141323024)
257
+ 8500 val loss 5.7673
258
+ 8500 val perplexity 319.6715
259
+ 8500 train 5.737051 (lr=5.6121e-06) (hash(x)=150696521)
260
+ 8600 val loss 5.7659
261
+ 8600 val perplexity 319.2322
262
+ 8600 train 5.718998 (lr=5.5404e-06) (hash(x)=162288191)
263
+ 8700 val loss 5.7624
264
+ 8700 val perplexity 318.1103
265
+ 8700 train 5.633211 (lr=5.5045e-06) (hash(x)=152860941)
266
+ 8749 val loss 5.7618
267
+ 8749 val perplexity 317.9061
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
attention_kindselective_n_heads4_seed1341/model_08749.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:395f6caf00fc6931495f915a58e51ee5930a2f4ca2b3f84a36b624a6ae25a106
3
  size 92843394
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b7fd67aa3ad05060e49d991c1a5ce788f8dd22e7a1852efc3ad664cfb3e9e6a
3
  size 92843394
attention_kindselective_n_heads4_seed1341/optimizer_08749.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9059975da2d811a2ec619a9cd6f63d985d15ff163e8a4becb0434b514ed8062b
3
  size 179406214
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e26d19de11e659cfb1c05e7f92e8c844a6fa530d7e243d5fc97a27e28b66d8ec
3
  size 179406214