andrew-healey commited on
Commit
7cf62a7
·
verified ·
1 Parent(s): 2cc1690

Upload folder using huggingface_hub

Browse files
attention_kindselective_n_heads4_seed1343/args.json CHANGED
@@ -1 +1 @@
1
- {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_9/attention_kindselective_n_heads4_seed1343", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 4, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 8750, "warmup_steps": 500, "group": "wider_is_better_9", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1343, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 120, "total_batch_size": 61440, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 5e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "5e-5_61440_4_1343", "n_embd": 256}
 
1
+ {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_9/attention_kindselective_n_heads4_seed1343", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 4, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 8750, "warmup_steps": 500, "group": "wider_is_better_9", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1343, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 120, "total_batch_size": 61440, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 3.5e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "3.5e-5_61440_4_1343", "n_embd": 256}
attention_kindselective_n_heads4_seed1343/log2.txt CHANGED
@@ -1,470 +1,424 @@
1
  max_steps: 8750
 
 
 
2
  0 val loss 11.3014
3
- 0 val perplexity 80933.0312
4
- 2100 val loss 6.3224
5
- 2100 val perplexity 556.9330
6
- 2100 train 6.196806 (lr=3.6761e-05) (hash(x)=146254754)
7
- 2200 val loss 6.2845
8
- 2200 val perplexity 536.1783
9
- 0 train 11.305981 (lr=1.0000e-07) (hash(x)=153418542)
10
- 2200 train 6.313621 (lr=3.6358e-05) (hash(x)=150630461)
11
- 2300 val loss 6.2546
12
- 2300 val perplexity 520.3895
13
- 100 val loss 9.9368
14
- 100 val perplexity 20677.2930
15
- 2300 train 6.399120 (lr=3.5935e-05) (hash(x)=165208942)
16
- 100 train 9.942423 (lr=1.0100e-05) (hash(x)=143722001)
17
- 2400 val loss 6.2406
18
- 2400 val perplexity 513.1553
19
- 2400 train 6.163989 (lr=3.5491e-05) (hash(x)=153448706)
20
- 200 val loss 9.1738
21
- 200 val perplexity 9641.2969
22
- 200 train 9.168076 (lr=2.0100e-05) (hash(x)=152989689)
23
- 2500 val loss 6.2074
24
- 2500 val perplexity 496.4181
25
- 2500 train 6.221104 (lr=3.5027e-05) (hash(x)=148505056)
26
- 300 val loss 8.0313
27
- 300 val perplexity 3075.7539
28
- 300 train 8.040445 (lr=3.0100e-05) (hash(x)=150071018)
29
- 2600 val loss 6.1837
30
- 2600 val perplexity 484.7639
31
- 2600 train 6.135411 (lr=3.4545e-05) (hash(x)=142911960)
32
- 400 val loss 7.5503
33
- 400 val perplexity 1901.2579
34
- 400 train 7.585745 (lr=4.0100e-05) (hash(x)=153358238)
35
- 2700 val loss 6.1688
36
- 2700 val perplexity 477.6137
37
- 2700 train 6.176059 (lr=3.4044e-05) (hash(x)=149951660)
38
- 500 val loss 7.4193
39
- 500 val perplexity 1667.9182
40
- 500 train 7.292376 (lr=5.0000e-05) (hash(x)=148983354)
41
- 2800 val loss 6.1399
42
- 2800 val perplexity 464.0015
43
- 2800 train 6.131063 (lr=3.3526e-05) (hash(x)=152956713)
44
- 600 val loss 7.3269
45
- 600 val perplexity 1520.5901
46
- 600 train 7.215260 (lr=4.9984e-05) (hash(x)=150770333)
47
- 2900 val loss 6.1245
48
- 2900 val perplexity 456.9059
49
- 2900 train 6.038101 (lr=3.2992e-05) (hash(x)=147247056)
50
- 700 val loss 7.2152
51
- 700 val perplexity 1359.8782
52
- 700 train 7.133345 (lr=4.9935e-05) (hash(x)=145246201)
53
- 3000 val loss 6.1072
54
- 3000 val perplexity 449.0740
55
- 3000 train 5.915527 (lr=3.2441e-05) (hash(x)=146911716)
56
- 800 val loss 7.1187
57
- 800 val perplexity 1234.8700
58
- 800 train 6.945539 (lr=4.9853e-05) (hash(x)=142635842)
59
- 3100 val loss 6.0823
60
- 3100 val perplexity 438.0373
61
- 3100 train 6.100656 (lr=3.1875e-05) (hash(x)=153282809)
62
- 900 val loss 7.0203
63
- 900 val perplexity 1119.0918
64
- 900 train 7.039608 (lr=4.9739e-05) (hash(x)=148299832)
65
- 3200 val loss 6.0701
66
- 3200 val perplexity 432.7231
67
- 3200 train 5.970005 (lr=3.1295e-05) (hash(x)=152009984)
68
- 1000 val loss 6.9440
69
- 1000 val perplexity 1036.9138
70
- 1000 train 6.905874 (lr=4.9593e-05) (hash(x)=152887721)
71
- 3300 val loss 6.0525
72
- 3300 val perplexity 425.1593
73
- 3300 train 5.965965 (lr=3.0702e-05) (hash(x)=150012952)
74
- 1100 val loss 6.8704
75
- 1100 val perplexity 963.2864
76
- 1100 train 6.759805 (lr=4.9415e-05) (hash(x)=153721194)
77
- 3400 val loss 6.0426
78
- 3400 val perplexity 420.9893
79
- 3400 train 5.870891 (lr=3.0095e-05) (hash(x)=146217477)
80
- 1200 val loss 6.7894
81
- 1200 val perplexity 888.3673
82
- 1200 train 6.780546 (lr=4.9205e-05) (hash(x)=153047184)
83
- 3500 val loss 6.0158
84
- 3500 val perplexity 409.8569
85
- 3500 train 6.183609 (lr=2.9477e-05) (hash(x)=180156144)
86
- 1300 val loss 6.7008
87
- 1300 val perplexity 813.0482
88
- 1300 train 6.604268 (lr=4.8964e-05) (hash(x)=152466045)
89
- 3600 val loss 6.0103
90
- 3600 val perplexity 407.6101
91
- 3600 train 6.041433 (lr=2.8849e-05) (hash(x)=154243319)
92
- 1400 val loss 6.6371
93
- 1400 val perplexity 762.8847
94
- 1400 train 6.432257 (lr=4.8691e-05) (hash(x)=151991552)
95
- 3700 val loss 5.9959
96
- 3700 val perplexity 401.7838
97
- 3700 train 5.900227 (lr=2.8210e-05) (hash(x)=154330476)
98
- 1500 val loss 6.5722
99
- 1500 val perplexity 714.9579
100
- 1500 train 6.518350 (lr=4.8388e-05) (hash(x)=147892594)
101
- 3800 val loss 5.9745
102
- 3800 val perplexity 393.2638
103
- 3800 train 5.912947 (lr=2.7562e-05) (hash(x)=145988858)
104
- 1600 val loss 6.5054
105
- 1600 val perplexity 668.7148
106
- 1600 train 6.312996 (lr=4.8055e-05) (hash(x)=135678663)
107
- 3900 val loss 5.9611
108
- 3900 val perplexity 388.0560
109
- 3900 train 5.967092 (lr=2.6907e-05) (hash(x)=149648609)
110
- 1700 val loss 6.4774
111
- 1700 val perplexity 650.2755
112
- 1700 train 6.497375 (lr=4.7691e-05) (hash(x)=143909276)
113
- 4000 val loss 5.9522
114
- 4000 val perplexity 384.5835
115
- 4000 train 5.925228 (lr=2.6244e-05) (hash(x)=154149272)
116
- 1800 val loss 6.4415
117
- 1800 val perplexity 627.3212
118
- 1800 train 6.596472 (lr=4.7299e-05) (hash(x)=155393468)
119
- 4100 val loss 5.9394
120
- 4100 val perplexity 379.7084
121
- 4100 train 5.833369 (lr=2.5575e-05) (hash(x)=132608538)
122
- 1900 val loss 6.3915
123
- 1900 val perplexity 596.7308
124
- 1900 train 6.438015 (lr=4.6878e-05) (hash(x)=150957388)
125
- 4200 val loss 5.9195
126
- 4200 val perplexity 372.2381
127
- 4200 train 5.868464 (lr=2.4900e-05) (hash(x)=140443636)
128
- 2000 val loss 6.3512
129
- 2000 val perplexity 573.1802
130
- 2000 train 6.613527 (lr=4.6428e-05) (hash(x)=144030938)
131
- 4300 val loss 5.9172
132
- 4300 val perplexity 371.3643
133
- 4300 train 5.843295 (lr=2.4222e-05) (hash(x)=138919540)
134
- 2100 val loss 6.3281
135
- 2100 val perplexity 560.1060
136
- 2100 train 6.201053 (lr=4.5951e-05) (hash(x)=146254754)
137
- 4400 val loss 5.9048
138
- 4400 val perplexity 366.7889
139
- 4400 train 5.880773 (lr=2.3540e-05) (hash(x)=153594684)
140
- 2200 val loss 6.2873
141
- 2200 val perplexity 537.7227
142
- 2200 train 6.317572 (lr=4.5448e-05) (hash(x)=150630461)
143
- 4500 val loss 5.8904
144
- 4500 val perplexity 361.5636
145
- 4500 train 5.839415 (lr=2.2856e-05) (hash(x)=144084750)
146
- 2300 val loss 6.2593
147
- 2300 val perplexity 522.8305
148
- 2300 train 6.396164 (lr=4.4918e-05) (hash(x)=165208942)
149
- 4600 val loss 5.8777
150
- 4600 val perplexity 356.9865
151
- 4600 train 5.835000 (lr=2.2171e-05) (hash(x)=147423675)
152
- 2400 val loss 6.2430
153
- 2400 val perplexity 514.4112
154
- 2400 train 6.165377 (lr=4.4363e-05) (hash(x)=153448706)
155
- 4700 val loss 5.8663
156
- 4700 val perplexity 352.9321
157
- 4700 train 5.934175 (lr=2.1486e-05) (hash(x)=162157039)
158
- 2500 val loss 6.2069
159
- 2500 val perplexity 496.1466
160
- 2500 train 6.221903 (lr=4.3784e-05) (hash(x)=148505056)
161
- 4800 val loss 5.8548
162
- 4800 val perplexity 348.8984
163
- 4800 train 5.754379 (lr=2.0801e-05) (hash(x)=134970942)
164
- 2600 val loss 6.1822
165
- 2600 val perplexity 484.0474
166
- 2600 train 6.129616 (lr=4.3181e-05) (hash(x)=142911960)
167
- 4900 val loss 5.8505
168
- 4900 val perplexity 347.4015
169
- 4900 train 5.765707 (lr=2.0118e-05) (hash(x)=160093370)
170
- 2700 val loss 6.1644
171
- 2700 val perplexity 475.5042
172
- 2700 train 6.167647 (lr=4.2555e-05) (hash(x)=149951660)
173
- 5000 val loss 5.8379
174
- 5000 val perplexity 343.0621
175
- 5000 train 5.732998 (lr=1.9438e-05) (hash(x)=154986299)
176
- 2800 val loss 6.1327
177
- 2800 val perplexity 460.6697
178
- 2800 train 6.113997 (lr=4.1908e-05) (hash(x)=152956713)
179
- 5100 val loss 5.8256
180
- 5100 val perplexity 338.8670
181
- 5100 train 5.676945 (lr=1.8762e-05) (hash(x)=145406582)
182
- 2900 val loss 6.1145
183
- 2900 val perplexity 452.3685
184
- 2900 train 6.028293 (lr=4.1240e-05) (hash(x)=147247056)
185
- 5200 val loss 5.8219
186
- 5200 val perplexity 337.6047
187
- 5200 train 5.873819 (lr=1.8090e-05) (hash(x)=148029261)
188
- 3000 val loss 6.0926
189
- 3000 val perplexity 442.5560
190
- 3000 train 5.884184 (lr=4.0551e-05) (hash(x)=146911716)
191
- 5300 val loss 5.8068
192
- 5300 val perplexity 332.5408
193
- 5300 train 5.803960 (lr=1.7424e-05) (hash(x)=157909233)
194
- 3100 val loss 6.0708
195
- 3100 val perplexity 433.0459
196
- 3100 train 6.087581 (lr=3.9844e-05) (hash(x)=153282809)
197
- 5400 val loss 5.7990
198
- 5400 val perplexity 329.9706
199
- 5400 train 5.673873 (lr=1.6765e-05) (hash(x)=142250860)
200
- 3200 val loss 6.0537
201
- 3200 val perplexity 425.6975
202
- 3200 train 5.953282 (lr=3.9119e-05) (hash(x)=152009984)
203
- 5500 val loss 5.7910
204
- 5500 val perplexity 327.3319
205
- 5500 train 5.669307 (lr=1.6113e-05) (hash(x)=145694703)
206
- 3300 val loss 6.0375
207
- 3300 val perplexity 418.8448
208
- 3300 train 5.941113 (lr=3.8377e-05) (hash(x)=150012952)
209
- 5600 val loss 5.7805
210
- 5600 val perplexity 323.9262
211
- 5600 train 5.709619 (lr=1.5469e-05) (hash(x)=141788252)
212
- 3400 val loss 6.0252
213
- 3400 val perplexity 413.7304
214
- 3400 train 5.861111 (lr=3.7619e-05) (hash(x)=146217477)
215
- 5700 val loss 5.7769
216
- 5700 val perplexity 322.7441
217
- 5700 train 5.676227 (lr=1.4836e-05) (hash(x)=150886692)
218
- 3500 val loss 5.9987
219
- 3500 val perplexity 402.9040
220
- 3500 train 6.165073 (lr=3.6847e-05) (hash(x)=180156144)
221
- 5800 val loss 5.7699
222
- 5800 val perplexity 320.5020
223
- 5800 train 5.680188 (lr=1.4212e-05) (hash(x)=151814419)
224
- 3600 val loss 5.9841
225
- 3600 val perplexity 397.0735
226
- 3600 train 6.020372 (lr=3.6061e-05) (hash(x)=154243319)
227
- 5900 val loss 5.7615
228
- 5900 val perplexity 317.8253
229
- 5900 train 5.777674 (lr=1.3600e-05) (hash(x)=146972539)
230
- 3700 val loss 5.9718
231
- 3700 val perplexity 392.2184
232
- 3700 train 5.873587 (lr=3.5263e-05) (hash(x)=154330476)
233
- 6000 val loss 5.7513
234
- 6000 val perplexity 314.6142
235
- 6000 train 5.636987 (lr=1.3000e-05) (hash(x)=150219954)
236
- 3800 val loss 5.9490
237
- 3800 val perplexity 383.3524
238
- 3800 train 5.881207 (lr=3.4453e-05) (hash(x)=145988858)
239
- 6100 val loss 5.7502
240
- 6100 val perplexity 314.2680
241
- 6100 train 5.747046 (lr=1.2413e-05) (hash(x)=161131435)
242
- 3900 val loss 5.9333
243
- 3900 val perplexity 377.3957
244
- 3900 train 5.938285 (lr=3.3633e-05) (hash(x)=149648609)
245
- 6200 val loss 5.7391
246
- 6200 val perplexity 310.7870
247
- 6200 train 5.753590 (lr=1.1840e-05) (hash(x)=150226893)
248
- 4000 val loss 5.9220
249
- 4000 val perplexity 373.1478
250
- 4000 train 5.897403 (lr=3.2805e-05) (hash(x)=154149272)
251
- 6300 val loss 5.7344
252
- 6300 val perplexity 309.3321
253
- 6300 train 5.564880 (lr=1.1282e-05) (hash(x)=144389552)
254
- 4100 val loss 5.9014
255
- 4100 val perplexity 365.5590
256
- 4100 train 5.797142 (lr=3.1968e-05) (hash(x)=132608538)
257
- 6400 val loss 5.7310
258
- 6400 val perplexity 308.2737
259
- 6400 train 5.513679 (lr=1.0739e-05) (hash(x)=147757943)
260
- 4200 val loss 5.8845
261
- 4200 val perplexity 359.4143
262
- 4200 train 5.833921 (lr=3.1126e-05) (hash(x)=140443636)
263
- 6500 val loss 5.7216
264
- 6500 val perplexity 305.4032
265
- 6500 train 5.706445 (lr=1.0213e-05) (hash(x)=144616373)
266
- 4300 val loss 5.8805
267
- 4300 val perplexity 358.0006
268
- 4300 train 5.813493 (lr=3.0277e-05) (hash(x)=138919540)
269
- 6600 val loss 5.7210
270
- 6600 val perplexity 305.2003
271
- 6600 train 5.544097 (lr=9.7032e-06) (hash(x)=145266452)
272
- 4400 val loss 5.8668
273
- 4400 val perplexity 353.1261
274
- 4400 train 5.851140 (lr=2.9425e-05) (hash(x)=153594684)
275
- 6700 val loss 5.7158
276
- 6700 val perplexity 303.6178
277
- 6700 train 5.709424 (lr=9.2116e-06) (hash(x)=149952261)
278
- 4500 val loss 5.8488
279
- 4500 val perplexity 346.8007
280
- 4500 train 5.799411 (lr=2.8571e-05) (hash(x)=144084750)
281
- 6800 val loss 5.7100
282
- 6800 val perplexity 301.8742
283
- 6800 train 5.594436 (lr=8.7387e-06) (hash(x)=143351199)
284
- 4600 val loss 5.8407
285
- 4600 val perplexity 344.0316
286
- 4600 train 5.799302 (lr=2.7714e-05) (hash(x)=147423675)
287
- 6900 val loss 5.7100
288
- 6900 val perplexity 301.8849
289
- 6900 train 5.672192 (lr=8.2849e-06) (hash(x)=156055618)
290
- 4700 val loss 5.8252
291
- 4700 val perplexity 338.7215
292
- 4700 train 5.892113 (lr=2.6857e-05) (hash(x)=162157039)
293
- 7000 val loss 5.7022
294
- 7000 val perplexity 299.5153
295
- 7000 train 5.774765 (lr=7.8510e-06) (hash(x)=164290908)
296
- 4800 val loss 5.8125
297
- 4800 val perplexity 334.4376
298
- 4800 train 5.717739 (lr=2.6002e-05) (hash(x)=134970942)
299
- 7100 val loss 5.6998
300
- 7100 val perplexity 298.7951
301
- 7100 train 5.805682 (lr=7.4377e-06) (hash(x)=150263832)
302
- 4900 val loss 5.8041
303
- 4900 val perplexity 331.6604
304
- 4900 train 5.717675 (lr=2.5148e-05) (hash(x)=160093370)
305
- 7200 val loss 5.6960
306
- 7200 val perplexity 297.6701
307
- 7200 train 5.380092 (lr=7.0455e-06) (hash(x)=139219680)
308
- 5000 val loss 5.7910
309
- 5000 val perplexity 327.3266
310
- 5000 train 5.678819 (lr=2.4298e-05) (hash(x)=154986299)
311
- 7300 val loss 5.6911
312
- 7300 val perplexity 296.2145
313
- 7300 train 5.636185 (lr=6.6749e-06) (hash(x)=150662994)
314
- 5100 val loss 5.7785
315
- 5100 val perplexity 323.2635
316
- 5100 train 5.629937 (lr=2.3452e-05) (hash(x)=145406582)
317
- 7400 val loss 5.6884
318
- 7400 val perplexity 295.4274
319
- 7400 train 5.530379 (lr=6.3266e-06) (hash(x)=139007967)
320
- 5200 val loss 5.7709
321
- 5200 val perplexity 320.8177
322
- 5200 train 5.830273 (lr=2.2613e-05) (hash(x)=148029261)
323
- 7500 val loss 5.6865
324
- 7500 val perplexity 294.8652
325
- 7500 train 5.512374 (lr=6.0010e-06) (hash(x)=138142461)
326
- 5300 val loss 5.7569
327
- 5300 val perplexity 316.3814
328
- 5300 train 5.755656 (lr=2.1780e-05) (hash(x)=157909233)
329
- 7600 val loss 5.6826
330
- 7600 val perplexity 293.7216
331
- 7600 train 5.603275 (lr=5.6986e-06) (hash(x)=150023998)
332
- 5400 val loss 5.7469
333
- 5400 val perplexity 313.2240
334
- 5400 train 5.614817 (lr=2.0956e-05) (hash(x)=142250860)
335
- 7700 val loss 5.6784
336
- 7700 val perplexity 292.4821
337
- 7700 train 5.571332 (lr=5.4198e-06) (hash(x)=143393355)
338
- 5500 val loss 5.7398
339
- 5500 val perplexity 310.9943
340
- 5500 train 5.616222 (lr=2.0141e-05) (hash(x)=145694703)
341
- 7800 val loss 5.6787
342
- 7800 val perplexity 292.5575
343
- 7800 train 5.549252 (lr=5.1650e-06) (hash(x)=144916472)
344
- 5600 val loss 5.7284
345
- 5600 val perplexity 307.4781
346
- 5600 train 5.665291 (lr=1.9337e-05) (hash(x)=141788252)
347
- 7900 val loss 5.6739
348
- 7900 val perplexity 291.1699
349
- 7900 train 5.630789 (lr=4.9347e-06) (hash(x)=150236934)
350
- 5700 val loss 5.7235
351
- 5700 val perplexity 305.9635
352
- 5700 train 5.621311 (lr=1.8545e-05) (hash(x)=150886692)
353
- 8000 val loss 5.6710
354
- 8000 val perplexity 290.3376
355
- 8000 train 5.667097 (lr=4.7291e-06) (hash(x)=146536422)
356
- 5800 val loss 5.7146
357
- 5800 val perplexity 303.2628
358
- 5800 train 5.627031 (lr=1.7765e-05) (hash(x)=151814419)
359
- 8100 val loss 5.6706
360
- 8100 val perplexity 290.2078
361
- 8100 train 5.504349 (lr=4.5486e-06) (hash(x)=151300857)
362
- 5900 val loss 5.7072
363
- 5900 val perplexity 301.0217
364
- 5900 train 5.720672 (lr=1.7000e-05) (hash(x)=146972539)
365
- 8200 val loss 5.6692
366
- 8200 val perplexity 289.8167
367
- 8200 train 5.704734 (lr=4.3933e-06) (hash(x)=160351956)
368
- 6000 val loss 5.6969
369
- 6000 val perplexity 297.9379
370
- 6000 train 5.586535 (lr=1.6250e-05) (hash(x)=150219954)
371
- 8300 val loss 5.6657
372
- 8300 val perplexity 288.7936
373
- 8300 train 5.714310 (lr=4.2636e-06) (hash(x)=150514540)
374
- 6100 val loss 5.6921
375
- 6100 val perplexity 296.5178
376
- 6100 train 5.682611 (lr=1.5516e-05) (hash(x)=161131435)
377
- 8400 val loss 5.6639
378
- 8400 val perplexity 288.2700
379
- 8400 train 5.613809 (lr=4.1596e-06) (hash(x)=155904762)
380
- 6200 val loss 5.6831
381
- 6200 val perplexity 293.8677
382
- 6200 train 5.696242 (lr=1.4800e-05) (hash(x)=150226893)
383
- 8500 val loss 5.6650
384
- 8500 val perplexity 288.5907
385
- 8500 train 5.469418 (lr=4.0815e-06) (hash(x)=146923196)
386
- 6300 val loss 5.6761
387
- 6300 val perplexity 291.8135
388
- 6300 train 5.511406 (lr=1.4102e-05) (hash(x)=144389552)
389
- 8600 val loss 5.6607
390
- 8600 val perplexity 287.3493
391
- 8600 train 5.639455 (lr=4.0294e-06) (hash(x)=138977080)
392
- 6400 val loss 5.6739
393
- 6400 val perplexity 291.1631
394
- 6400 train 5.455398 (lr=1.3424e-05) (hash(x)=147757943)
395
- 8700 val loss 5.6583
396
- 8700 val perplexity 286.6700
397
- 8700 train 5.469374 (lr=4.0033e-06) (hash(x)=146442792)
398
- 6500 val loss 5.6628
399
- 6500 val perplexity 287.9458
400
- 6500 train 5.650785 (lr=1.2766e-05) (hash(x)=144616373)
401
- 8749 val loss 5.6585
402
- 8749 val perplexity 286.7288
403
- 6600 val loss 5.6621
404
- 6600 val perplexity 287.7638
405
- 6600 train 5.484093 (lr=1.2129e-05) (hash(x)=145266452)
406
- 6700 val loss 5.6560
407
- 6700 val perplexity 286.0123
408
- 6700 train 5.652499 (lr=1.1515e-05) (hash(x)=149952261)
409
- 6800 val loss 5.6499
410
- 6800 val perplexity 284.2496
411
- 6800 train 5.533232 (lr=1.0923e-05) (hash(x)=143351199)
412
- 6900 val loss 5.6497
413
- 6900 val perplexity 284.1957
414
- 6900 train 5.607444 (lr=1.0356e-05) (hash(x)=156055618)
415
- 7000 val loss 5.6418
416
- 7000 val perplexity 281.9726
417
- 7000 train 5.713253 (lr=9.8138e-06) (hash(x)=164290908)
418
- 7100 val loss 5.6384
419
- 7100 val perplexity 281.0004
420
- 7100 train 5.740729 (lr=9.2971e-06) (hash(x)=150263832)
421
- 7200 val loss 5.6352
422
- 7200 val perplexity 280.1169
423
- 7200 train 5.315210 (lr=8.8068e-06) (hash(x)=139219680)
424
- 7300 val loss 5.6297
425
- 7300 val perplexity 278.5819
426
- 7300 train 5.575299 (lr=8.3436e-06) (hash(x)=150662994)
427
- 7400 val loss 5.6261
428
- 7400 val perplexity 277.5779
429
- 7400 train 5.465163 (lr=7.9082e-06) (hash(x)=139007967)
430
- 7500 val loss 5.6249
431
- 7500 val perplexity 277.2398
432
- 7500 train 5.455156 (lr=7.5012e-06) (hash(x)=138142461)
433
- 7600 val loss 5.6196
434
- 7600 val perplexity 275.7852
435
- 7600 train 5.530750 (lr=7.1232e-06) (hash(x)=150023998)
436
- 7700 val loss 5.6152
437
- 7700 val perplexity 274.5555
438
- 7700 train 5.509050 (lr=6.7747e-06) (hash(x)=143393355)
439
- 7800 val loss 5.6149
440
- 7800 val perplexity 274.4759
441
- 7800 train 5.480126 (lr=6.4563e-06) (hash(x)=144916472)
442
- 7900 val loss 5.6097
443
- 7900 val perplexity 273.0717
444
- 7900 train 5.565450 (lr=6.1684e-06) (hash(x)=150236934)
445
- 8000 val loss 5.6071
446
- 8000 val perplexity 272.3544
447
- 8000 train 5.608926 (lr=5.9114e-06) (hash(x)=146536422)
448
- 8100 val loss 5.6069
449
- 8100 val perplexity 272.3043
450
- 8100 train 5.441198 (lr=5.6857e-06) (hash(x)=151300857)
451
- 8200 val loss 5.6050
452
- 8200 val perplexity 271.7760
453
- 8200 train 5.634646 (lr=5.4917e-06) (hash(x)=160351956)
454
- 8300 val loss 5.6014
455
- 8300 val perplexity 270.8028
456
- 8300 train 5.654285 (lr=5.3295e-06) (hash(x)=150514540)
457
- 8400 val loss 5.5999
458
- 8400 val perplexity 270.4036
459
- 8400 train 5.547650 (lr=5.1995e-06) (hash(x)=155904762)
460
- 8500 val loss 5.5998
461
- 8500 val perplexity 270.3614
462
- 8500 train 5.406527 (lr=5.1019e-06) (hash(x)=146923196)
463
- 8600 val loss 5.5961
464
- 8600 val perplexity 269.3655
465
- 8600 train 5.577782 (lr=5.0367e-06) (hash(x)=138977080)
466
- 8700 val loss 5.5944
467
- 8700 val perplexity 268.9079
468
- 8700 train 5.395451 (lr=5.0041e-06) (hash(x)=146442792)
469
- 8749 val loss 5.5935
470
- 8749 val perplexity 268.6727
 
1
  max_steps: 8750
2
+ max_steps: 8750
3
+ 0 val loss 11.3014
4
+ 0 val perplexity 80932.9531
5
  0 val loss 11.3014
6
+ 0 val perplexity 80932.9531
7
+ 0 train 11.305982 (lr=7.0000e-08) (hash(x)=153418542)
8
+ 0 train 11.305982 (lr=6.0000e-08) (hash(x)=153418542)
9
+ 100 val loss 10.0438
10
+ 100 val perplexity 23011.9707
11
+ 100 train 10.050365 (lr=7.0700e-06) (hash(x)=143722001)
12
+ 100 val loss 10.1073
13
+ 100 val perplexity 24520.4238
14
+ 100 train 10.113642 (lr=6.0600e-06) (hash(x)=143722001)
15
+ 200 val loss 9.5448
16
+ 200 val perplexity 13971.1826
17
+ 200 train 9.549261 (lr=1.4070e-05) (hash(x)=152989689)
18
+ 300 val loss 8.6011
19
+ 300 val perplexity 5437.8970
20
+ 300 train 8.619393 (lr=2.1070e-05) (hash(x)=150071018)
21
+ 200 val loss 9.6223
22
+ 200 val perplexity 15098.4424
23
+ 200 train 9.629188 (lr=1.2060e-05) (hash(x)=152989689)
24
+ 400 val loss 7.8269
25
+ 400 val perplexity 2507.1106
26
+ 400 train 7.854829 (lr=2.8070e-05) (hash(x)=153358238)
27
+ 500 val loss 7.5393
28
+ 500 val perplexity 1880.5006
29
+ 500 train 7.423524 (lr=3.5000e-05) (hash(x)=148983354)
30
+ 300 val loss 8.7160
31
+ 300 val perplexity 6099.9268
32
+ 300 train 8.734112 (lr=1.8060e-05) (hash(x)=150071018)
33
+ 600 val loss 7.4113
34
+ 600 val perplexity 1654.4934
35
+ 600 train 7.314175 (lr=3.4989e-05) (hash(x)=150770333)
36
+ 700 val loss 7.3267
37
+ 700 val perplexity 1520.3972
38
+ 400 val loss 7.8636
39
+ 400 val perplexity 2600.8687
40
+ 700 train 7.248207 (lr=3.4954e-05) (hash(x)=145246201)
41
+ 400 train 7.890481 (lr=2.4060e-05) (hash(x)=153358238)
42
+ 800 val loss 7.2515
43
+ 800 val perplexity 1410.1771
44
+ 800 train 7.083712 (lr=3.4897e-05) (hash(x)=142635842)
45
+ 500 val loss 7.5315
46
+ 500 val perplexity 1865.8441
47
+ 500 train 7.414758 (lr=3.0000e-05) (hash(x)=148983354)
48
+ 900 val loss 7.1575
49
+ 900 val perplexity 1283.6674
50
+ 900 train 7.189825 (lr=3.4818e-05) (hash(x)=148299832)
51
+ 1000 val loss 7.0754
52
+ 1000 val perplexity 1182.5190
53
+ 1000 train 7.035053 (lr=3.4715e-05) (hash(x)=152887721)
54
+ 600 val loss 7.4045
55
+ 600 val perplexity 1643.3911
56
+ 600 train 7.309084 (lr=2.9990e-05) (hash(x)=150770333)
57
+ 1100 val loss 7.0022
58
+ 1100 val perplexity 1099.0203
59
+ 1100 train 6.891657 (lr=3.4591e-05) (hash(x)=153721194)
60
+ 1200 val loss 6.9192
61
+ 1200 val perplexity 1011.5258
62
+ 1200 train 6.912433 (lr=3.4444e-05) (hash(x)=153047184)
63
+ 700 val loss 7.3257
64
+ 700 val perplexity 1518.8248
65
+ 700 train 7.243683 (lr=2.9961e-05) (hash(x)=145246201)
66
+ 1300 val loss 6.8385
67
+ 1300 val perplexity 933.1031
68
+ 1300 train 6.749180 (lr=3.4275e-05) (hash(x)=152466045)
69
+ 800 val loss 7.2468
70
+ 800 val perplexity 1403.6697
71
+ 1400 val loss 6.7694
72
+ 1400 val perplexity 870.7479
73
+ 800 train 7.078628 (lr=2.9912e-05) (hash(x)=142635842)
74
+ 1400 train 6.569815 (lr=3.4084e-05) (hash(x)=151991552)
75
+ 1500 val loss 6.6936
76
+ 1500 val perplexity 807.2139
77
+ 1500 train 6.642127 (lr=3.3872e-05) (hash(x)=147892594)
78
+ 900 val loss 7.1541
79
+ 900 val perplexity 1279.3289
80
+ 900 train 7.189307 (lr=2.9844e-05) (hash(x)=148299832)
81
+ 1600 val loss 6.6326
82
+ 1600 val perplexity 759.4801
83
+ 1600 train 6.438039 (lr=3.3638e-05) (hash(x)=135678663)
84
+ 1700 val loss 6.5863
85
+ 1700 val perplexity 725.0812
86
+ 1700 train 6.610061 (lr=3.3384e-05) (hash(x)=143909276)
87
+ 1000 val loss 7.0835
88
+ 1000 val perplexity 1192.1432
89
+ 1000 train 7.044007 (lr=2.9756e-05) (hash(x)=152887721)
90
+ 1800 val loss 6.5362
91
+ 1800 val perplexity 689.6570
92
+ 1800 train 6.684325 (lr=3.3109e-05) (hash(x)=155393468)
93
+ 1900 val loss 6.4801
94
+ 1900 val perplexity 652.0583
95
+ 1900 train 6.526745 (lr=3.2814e-05) (hash(x)=150957388)
96
+ 1100 val loss 7.0117
97
+ 1100 val perplexity 1109.5219
98
+ 1100 train 6.898890 (lr=2.9649e-05) (hash(x)=153721194)
99
+ 2000 val loss 6.4394
100
+ 2000 val perplexity 626.0261
101
+ 2000 train 6.688434 (lr=3.2500e-05) (hash(x)=144030938)
102
+ 1200 val loss 6.9380
103
+ 1200 val perplexity 1030.7054
104
+ 1200 train 6.929457 (lr=2.9523e-05) (hash(x)=153047184)
105
+ 2100 val loss 6.4148
106
+ 2100 val perplexity 610.8321
107
+ 2100 train 6.284521 (lr=3.2166e-05) (hash(x)=146254754)
108
+ 2200 val loss 6.3695
109
+ 2200 val perplexity 583.7899
110
+ 2200 train 6.399682 (lr=3.1813e-05) (hash(x)=150630461)
111
+ 1300 val loss 6.8630
112
+ 1300 val perplexity 956.2305
113
+ 1300 train 6.769284 (lr=2.9378e-05) (hash(x)=152466045)
114
+ 2300 val loss 6.3410
115
+ 2300 val perplexity 567.3669
116
+ 2300 train 6.493802 (lr=3.1443e-05) (hash(x)=165208942)
117
+ 2400 val loss 6.3212
118
+ 2400 val perplexity 556.2291
119
+ 2400 train 6.247215 (lr=3.1054e-05) (hash(x)=153448706)
120
+ 1400 val loss 6.8011
121
+ 1400 val perplexity 898.8173
122
+ 1400 train 6.607848 (lr=2.9215e-05) (hash(x)=151991552)
123
+ 2500 val loss 6.2890
124
+ 2500 val perplexity 538.6242
125
+ 2500 train 6.301908 (lr=3.0649e-05) (hash(x)=148505056)
126
+ 1500 val loss 6.7320
127
+ 1500 val perplexity 838.8151
128
+ 1500 train 6.682234 (lr=2.9033e-05) (hash(x)=147892594)
129
+ 2600 val loss 6.2645
130
+ 2600 val perplexity 525.5840
131
+ 2600 train 6.209645 (lr=3.0227e-05) (hash(x)=142911960)
132
+ 2700 val loss 6.2464
133
+ 2700 val perplexity 516.1746
134
+ 2700 train 6.249057 (lr=2.9789e-05) (hash(x)=149951660)
135
+ 1600 val loss 6.6699
136
+ 1600 val perplexity 788.3246
137
+ 1600 train 6.471530 (lr=2.8833e-05) (hash(x)=135678663)
138
+ 2800 val loss 6.2150
139
+ 2800 val perplexity 500.2123
140
+ 2800 train 6.209450 (lr=2.9336e-05) (hash(x)=152956713)
141
+ 2900 val loss 6.1961
142
+ 2900 val perplexity 490.8508
143
+ 2900 train 6.112799 (lr=2.8868e-05) (hash(x)=147247056)
144
+ 1700 val loss 6.6200
145
+ 1700 val perplexity 749.9150
146
+ 1700 train 6.636775 (lr=2.8615e-05) (hash(x)=143909276)
147
+ 3000 val loss 6.1776
148
+ 3000 val perplexity 481.8291
149
+ 3000 train 5.997092 (lr=2.8386e-05) (hash(x)=146911716)
150
+ 3100 val loss 6.1579
151
+ 3100 val perplexity 472.4378
152
+ 3100 train 6.180588 (lr=2.7891e-05) (hash(x)=153282809)
153
+ 1800 val loss 6.5746
154
+ 1800 val perplexity 716.6734
155
+ 1800 train 6.721464 (lr=2.8379e-05) (hash(x)=155393468)
156
+ 3200 val loss 6.1410
157
+ 3200 val perplexity 464.5390
158
+ 3200 train 6.036775 (lr=2.7383e-05) (hash(x)=152009984)
159
+ 1900 val loss 6.5198
160
+ 1900 val perplexity 678.4648
161
+ 1900 train 6.568262 (lr=2.8127e-05) (hash(x)=150957388)
162
+ 3300 val loss 6.1275
163
+ 3300 val perplexity 458.2952
164
+ 3300 train 6.036661 (lr=2.6864e-05) (hash(x)=150012952)
165
+ 3400 val loss 6.1178
166
+ 3400 val perplexity 453.8775
167
+ 3400 train 5.951714 (lr=2.6333e-05) (hash(x)=146217477)
168
+ 2000 val loss 6.4772
169
+ 2000 val perplexity 650.1710
170
+ 2000 train 6.721745 (lr=2.7857e-05) (hash(x)=144030938)
171
+ 3500 val loss 6.0906
172
+ 3500 val perplexity 441.6902
173
+ 3500 train 6.259603 (lr=2.5793e-05) (hash(x)=180156144)
174
+ 3600 val loss 6.0826
175
+ 3600 val perplexity 438.1469
176
+ 3600 train 6.120096 (lr=2.5243e-05) (hash(x)=154243319)
177
+ 2100 val loss 6.4546
178
+ 2100 val perplexity 635.6494
179
+ 2100 train 6.322638 (lr=2.7571e-05) (hash(x)=146254754)
180
+ 3700 val loss 6.0649
181
+ 3700 val perplexity 430.4960
182
+ 3700 train 5.963401 (lr=2.4684e-05) (hash(x)=154330476)
183
+ 2200 val loss 6.4164
184
+ 2200 val perplexity 611.7704
185
+ 2200 train 6.443453 (lr=2.7269e-05) (hash(x)=150630461)
186
+ 3800 val loss 6.0467
187
+ 3800 val perplexity 422.7079
188
+ 3800 train 5.975914 (lr=2.4117e-05) (hash(x)=145988858)
189
+ 3900 val loss 6.0322
190
+ 3900 val perplexity 416.6099
191
+ 3900 train 6.045941 (lr=2.3543e-05) (hash(x)=149648609)
192
+ 2300 val loss 6.3844
193
+ 2300 val perplexity 592.5361
194
+ 2300 train 6.536951 (lr=2.6951e-05) (hash(x)=165208942)
195
+ 4000 val loss 6.0218
196
+ 4000 val perplexity 412.3248
197
+ 4000 train 5.994764 (lr=2.2963e-05) (hash(x)=154149272)
198
+ 4100 val loss 6.0072
199
+ 4100 val perplexity 406.3483
200
+ 4100 train 5.903815 (lr=2.2378e-05) (hash(x)=132608538)
201
+ 2400 val loss 6.3635
202
+ 2400 val perplexity 580.2852
203
+ 2400 train 6.286649 (lr=2.6618e-05) (hash(x)=153448706)
204
+ 4200 val loss 5.9916
205
+ 4200 val perplexity 400.0361
206
+ 4200 train 5.939827 (lr=2.1788e-05) (hash(x)=140443636)
207
+ 2500 val loss 6.3339
208
+ 2500 val perplexity 563.3612
209
+ 2500 train 6.346560 (lr=2.6270e-05) (hash(x)=148505056)
210
+ 4300 val loss 5.9913
211
+ 4300 val perplexity 399.9310
212
+ 4300 train 5.913877 (lr=2.1194e-05) (hash(x)=138919540)
213
+ 4400 val loss 5.9748
214
+ 4400 val perplexity 393.4041
215
+ 4400 train 5.947241 (lr=2.0598e-05) (hash(x)=153594684)
216
+ 2600 val loss 6.3080
217
+ 2600 val perplexity 548.9651
218
+ 2600 train 6.246406 (lr=2.5909e-05) (hash(x)=142911960)
219
+ 4500 val loss 5.9610
220
+ 4500 val perplexity 388.0159
221
+ 4500 train 5.911044 (lr=1.9999e-05) (hash(x)=144084750)
222
+ 4600 val loss 5.9548
223
+ 4600 val perplexity 385.5923
224
+ 4600 train 5.910222 (lr=1.9400e-05) (hash(x)=147423675)
225
+ 2700 val loss 6.2882
226
+ 2700 val perplexity 538.1934
227
+ 2700 train 6.289870 (lr=2.5533e-05) (hash(x)=149951660)
228
+ 4700 val loss 5.9404
229
+ 4700 val perplexity 380.0801
230
+ 4700 train 6.012012 (lr=1.8800e-05) (hash(x)=162157039)
231
+ 4800 val loss 5.9313
232
+ 4800 val perplexity 376.6331
233
+ 4800 train 5.835851 (lr=1.8201e-05) (hash(x)=134970942)
234
+ 2800 val loss 6.2582
235
+ 2800 val perplexity 522.2596
236
+ 2800 train 6.255882 (lr=2.5145e-05) (hash(x)=152956713)
237
+ 4900 val loss 5.9238
238
+ 4900 val perplexity 373.8367
239
+ 4900 train 5.844675 (lr=1.7604e-05) (hash(x)=160093370)
240
+ 2900 val loss 6.2437
241
+ 2900 val perplexity 514.7744
242
+ 2900 train 6.157321 (lr=2.4744e-05) (hash(x)=147247056)
243
+ 5000 val loss 5.9154
244
+ 5000 val perplexity 370.6849
245
+ 5000 train 5.809307 (lr=1.7009e-05) (hash(x)=154986299)
246
+ 5100 val loss 5.9005
247
+ 5100 val perplexity 365.2234
248
+ 5100 train 5.757463 (lr=1.6417e-05) (hash(x)=145406582)
249
+ 3000 val loss 6.2251
250
+ 3000 val perplexity 505.2902
251
+ 3000 train 6.048134 (lr=2.4331e-05) (hash(x)=146911716)
252
+ 5200 val loss 5.8944
253
+ 5200 val perplexity 362.9891
254
+ 5200 train 5.941761 (lr=1.5829e-05) (hash(x)=148029261)
255
+ 5300 val loss 5.8838
256
+ 5300 val perplexity 359.1739
257
+ 5300 train 5.881489 (lr=1.5246e-05) (hash(x)=157909233)
258
+ 3100 val loss 6.2032
259
+ 3100 val perplexity 494.3125
260
+ 3100 train 6.222645 (lr=2.3906e-05) (hash(x)=153282809)
261
+ 5400 val loss 5.8742
262
+ 5400 val perplexity 355.7338
263
+ 5400 train 5.757402 (lr=1.4669e-05) (hash(x)=142250860)
264
+ 3200 val loss 6.1904
265
+ 3200 val perplexity 488.0625
266
+ 3200 train 6.085604 (lr=2.3471e-05) (hash(x)=152009984)
267
+ 5500 val loss 5.8667
268
+ 5500 val perplexity 353.0680
269
+ 5500 train 5.740597 (lr=1.4099e-05) (hash(x)=145694703)
270
+ 5600 val loss 5.8590
271
+ 5600 val perplexity 350.3583
272
+ 5600 train 5.788935 (lr=1.3536e-05) (hash(x)=141788252)
273
+ 3300 val loss 6.1772
274
+ 3300 val perplexity 481.6196
275
+ 3300 train 6.087312 (lr=2.3026e-05) (hash(x)=150012952)
276
+ 5700 val loss 5.8508
277
+ 5700 val perplexity 347.4989
278
+ 5700 train 5.752746 (lr=1.2981e-05) (hash(x)=150886692)
279
+ 5800 val loss 5.8472
280
+ 5800 val perplexity 346.2778
281
+ 5800 train 5.754390 (lr=1.2436e-05) (hash(x)=151814419)
282
+ 3400 val loss 6.1625
283
+ 3400 val perplexity 474.5927
284
+ 3400 train 5.998357 (lr=2.2572e-05) (hash(x)=146217477)
285
+ 5900 val loss 5.8375
286
+ 5900 val perplexity 342.9117
287
+ 5900 train 5.847322 (lr=1.1900e-05) (hash(x)=146972539)
288
+ 6000 val loss 5.8301
289
+ 6000 val perplexity 340.3773
290
+ 6000 train 5.711991 (lr=1.1375e-05) (hash(x)=150219954)
291
+ 3500 val loss 6.1421
292
+ 3500 val perplexity 465.0062
293
+ 3500 train 6.311218 (lr=2.2108e-05) (hash(x)=180156144)
294
+ 6100 val loss 5.8264
295
+ 6100 val perplexity 339.1409
296
+ 6100 train 5.822284 (lr=1.0861e-05) (hash(x)=161131435)
297
+ 3600 val loss 6.1281
298
+ 3600 val perplexity 458.5415
299
+ 3600 train 6.163641 (lr=2.1637e-05) (hash(x)=154243319)
300
+ 6200 val loss 5.8175
301
+ 6200 val perplexity 336.1375
302
+ 6200 train 5.829421 (lr=1.0360e-05) (hash(x)=150226893)
303
+ 6300 val loss 5.8113
304
+ 6300 val perplexity 334.0599
305
+ 6300 train 5.643476 (lr=9.8715e-06) (hash(x)=144389552)
306
+ 3700 val loss 6.1135
307
+ 3700 val perplexity 451.9110
308
+ 3700 train 6.011994 (lr=2.1158e-05) (hash(x)=154330476)
309
+ 6400 val loss 5.8087
310
+ 6400 val perplexity 333.1811
311
+ 6400 train 5.592588 (lr=9.3966e-06) (hash(x)=147757943)
312
+ 6500 val loss 5.7999
313
+ 6500 val perplexity 330.2697
314
+ 6500 train 5.782908 (lr=8.9359e-06) (hash(x)=144616373)
315
+ 3800 val loss 6.0951
316
+ 3800 val perplexity 443.6720
317
+ 3800 train 6.029699 (lr=2.0672e-05) (hash(x)=145988858)
318
+ 6600 val loss 5.7994
319
+ 6600 val perplexity 330.1003
320
+ 6600 train 5.625536 (lr=8.4903e-06) (hash(x)=145266452)
321
+ 3900 val loss 6.0815
322
+ 3900 val perplexity 437.6682
323
+ 3900 train 6.089600 (lr=2.0180e-05) (hash(x)=149648609)
324
+ 6700 val loss 5.7939
325
+ 6700 val perplexity 328.2831
326
+ 6700 train 5.785510 (lr=8.0602e-06) (hash(x)=149952261)
327
+ 6800 val loss 5.7885
328
+ 6800 val perplexity 326.5093
329
+ 6800 train 5.665899 (lr=7.6463e-06) (hash(x)=143351199)
330
+ 4000 val loss 6.0739
331
+ 4000 val perplexity 434.3786
332
+ 4000 train 6.048274 (lr=1.9683e-05) (hash(x)=154149272)
333
+ 6900 val loss 5.7881
334
+ 6900 val perplexity 326.3762
335
+ 6900 train 5.749453 (lr=7.2493e-06) (hash(x)=156055618)
336
+ 7000 val loss 5.7804
337
+ 7000 val perplexity 323.9043
338
+ 7000 train 5.854444 (lr=6.8697e-06) (hash(x)=164290908)
339
+ 4100 val loss 6.0584
340
+ 4100 val perplexity 427.6766
341
+ 4100 train 5.955194 (lr=1.9181e-05) (hash(x)=132608538)
342
+ 7100 val loss 5.7781
343
+ 7100 val perplexity 323.1483
344
+ 7100 train 5.874842 (lr=6.5080e-06) (hash(x)=150263832)
345
+ 4200 val loss 6.0428
346
+ 4200 val perplexity 421.0503
347
+ 4200 train 5.982690 (lr=1.8675e-05) (hash(x)=140443636)
348
+ 7200 val loss 5.7763
349
+ 7200 val perplexity 322.5544
350
+ 7200 train 5.465231 (lr=6.1648e-06) (hash(x)=139219680)
351
+ 7300 val loss 5.7695
352
+ 7300 val perplexity 320.3765
353
+ 7300 train 5.715026 (lr=5.8405e-06) (hash(x)=150662994)
354
+ 4300 val loss 6.0377
355
+ 4300 val perplexity 418.9107
356
+ 4300 train 5.956387 (lr=1.8166e-05) (hash(x)=138919540)
357
+ 7400 val loss 5.7668
358
+ 7400 val perplexity 319.5232
359
+ 7400 train 5.611770 (lr=5.5357e-06) (hash(x)=139007967)
360
+ 7500 val loss 5.7658
361
+ 7500 val perplexity 319.1935
362
+ 7500 train 5.595723 (lr=5.2508e-06) (hash(x)=138142461)
363
+ 4400 val loss 6.0255
364
+ 4400 val perplexity 413.8688
365
+ 4400 train 5.995091 (lr=1.7655e-05) (hash(x)=153594684)
366
+ 7600 val loss 5.7609
367
+ 7600 val perplexity 317.6342
368
+ 7600 train 5.675484 (lr=4.9862e-06) (hash(x)=150023998)
369
+ 7700 val loss 5.7574
370
+ 7700 val perplexity 316.5240
371
+ 7700 train 5.652620 (lr=4.7423e-06) (hash(x)=143393355)
372
+ 4500 val loss 6.0108
373
+ 4500 val perplexity 407.7993
374
+ 4500 train 5.956801 (lr=1.7142e-05) (hash(x)=144084750)
375
+ 7800 val loss 5.7580
376
+ 7800 val perplexity 316.7283
377
+ 7800 train 5.633855 (lr=4.5194e-06) (hash(x)=144916472)
378
+ 4600 val loss 6.0024
379
+ 4600 val perplexity 404.3789
380
+ 4600 train 5.959134 (lr=1.6629e-05) (hash(x)=147423675)
381
+ 7900 val loss 5.7531
382
+ 7900 val perplexity 315.1684
383
+ 7900 train 5.706028 (lr=4.3179e-06) (hash(x)=150236934)
384
+ 8000 val loss 5.7501
385
+ 8000 val perplexity 314.2092
386
+ 8000 train 5.744445 (lr=4.1380e-06) (hash(x)=146536422)
387
+ 4700 val loss 5.9911
388
+ 4700 val perplexity 399.8620
389
+ 4700 train 6.067215 (lr=1.6114e-05) (hash(x)=162157039)
390
+ 8100 val loss 5.7506
391
+ 8100 val perplexity 314.3856
392
+ 8100 train 5.577597 (lr=3.9800e-06) (hash(x)=151300857)
393
+ 8200 val loss 5.7486
394
+ 8200 val perplexity 313.7418
395
+ 8200 train 5.792922 (lr=3.8442e-06) (hash(x)=160351956)
396
+ 4800 val loss 5.9808
397
+ 4800 val perplexity 395.7435
398
+ 4800 train 5.874648 (lr=1.5601e-05) (hash(x)=134970942)
399
+ 8300 val loss 5.7449
400
+ 8300 val perplexity 312.5950
401
+ 8300 train 5.781292 (lr=3.7307e-06) (hash(x)=150514540)
402
+ 4900 val loss 5.9759
403
+ 4900 val perplexity 393.8117
404
+ 4900 train 5.892208 (lr=1.5089e-05) (hash(x)=160093370)
405
+ 8400 val loss 5.7426
406
+ 8400 val perplexity 311.8808
407
+ 8400 train 5.695298 (lr=3.6397e-06) (hash(x)=155904762)
408
+ 8500 val loss 5.7438
409
+ 8500 val perplexity 312.2379
410
+ 8500 train 5.549510 (lr=3.5713e-06) (hash(x)=146923196)
411
+ 5000 val loss 5.9667
412
+ 5000 val perplexity 390.2094
413
+ 5000 train 5.861556 (lr=1.4579e-05) (hash(x)=154986299)
414
+ 8600 val loss 5.7400
415
+ 8600 val perplexity 311.0759
416
+ 8600 train 5.715517 (lr=3.5257e-06) (hash(x)=138977080)
417
+ 8700 val loss 5.7382
418
+ 8700 val perplexity 310.5197
419
+ 8700 train 5.543046 (lr=3.5029e-06) (hash(x)=146442792)
420
+ 5100 val loss 5.9525
421
+ 5100 val perplexity 384.7192
422
+ 5100 train 5.802466 (lr=1.4071e-05) (hash(x)=145406582)
423
+ 8749 val loss 5.7376
424
+ 8749 val perplexity 310.3325
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
attention_kindselective_n_heads4_seed1343/model_08749.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79630c585ae01a12e0da13e054f4d31bdbcb07132557794724704b86dc5f2eb8
3
  size 92843394
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:165fb75f76ce3bebd99414ffb57b2a8b9c8d2cb74bc6b1efd62ef1362ad7c162
3
  size 92843394
attention_kindselective_n_heads4_seed1343/optimizer_08749.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:496cc404e3f894099d946fd893c46f5c9984530da6547230964b29331cc3799f
3
  size 179406214
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5fdb83ba5561770f331d41f2e12db5dd14567e223eba95905bc5b55fdec00cbd
3
  size 179406214