andrew-healey commited on
Commit
c845ecb
·
verified ·
1 Parent(s): f3831fe

Upload folder using huggingface_hub

Browse files
attention_kindselective_n_heads8_seed1339/args.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_6/attention_kindselective_n_heads8_seed1339", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 8, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 10000, "warmup_steps": 200, "group": "wider_is_better_6", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1339, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 40, "total_batch_size": 10240, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 2e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "2e-5_10240_8_1339", "n_embd": 512}
attention_kindselective_n_heads8_seed1339/dataloader_02500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b2ea67f78ff5a7970d0db044ff7ee527b3dc065f295fd30f588df4b44b568d0
3
+ size 964
attention_kindselective_n_heads8_seed1339/dataloader_05000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f03ed2ebf741f15e13c79e6cc1e9a19b308450d81cc3b4d8d0338c63d77ca59
3
+ size 964
attention_kindselective_n_heads8_seed1339/dataloader_07500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82590037fb2eecbec961f7967a8dd1b8d85515d31a252f66b92b8139858a8b7c
3
+ size 964
attention_kindselective_n_heads8_seed1339/dataloader_09999.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c544303717d12355a69b8ffb1eb109434e4fdccfd5a61279b6e8ba2e870d6700
3
+ size 964
attention_kindselective_n_heads8_seed1339/log2.txt ADDED
@@ -0,0 +1,1209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ max_steps: 10000
2
+ 0 val loss 11.0281
3
+ 0 val perplexity 61581.4062
4
+ 0 val loss 11.0281
5
+ 0 val perplexity 61581.6992
6
+ 0 val loss 11.0281
7
+ 0 val perplexity 61581.4062
8
+ 0 val loss 11.0281
9
+ 0 val perplexity 61582.3477
10
+ 0 train 11.019774 (lr=1.5000e-07) (hash(x)=55241167)
11
+ 0 train 11.019774 (lr=3.5000e-07) (hash(x)=55241167)
12
+ 0 train 11.019774 (lr=2.5000e-07) (hash(x)=55241167)
13
+ 0 train 11.019774 (lr=1.0000e-07) (hash(x)=55241167)
14
+ 100 val loss 9.7097
15
+ 100 val perplexity 16476.1328
16
+ 100 train 9.700400 (lr=1.5150e-05) (hash(x)=46387919)
17
+ 100 val loss 9.4717
18
+ 100 val perplexity 12986.8184
19
+ 100 val loss 9.5842
20
+ 100 val perplexity 14532.9561
21
+ 100 val loss 9.8417
22
+ 100 val perplexity 18802.3438
23
+ 100 train 9.463525 (lr=3.5350e-05) (hash(x)=46387919)
24
+ 100 train 9.577502 (lr=2.5250e-05) (hash(x)=46387919)
25
+ 100 train 9.837164 (lr=1.0100e-05) (hash(x)=46387919)
26
+ 200 val loss 9.0894
27
+ 200 val perplexity 8861.2422
28
+ 200 train 9.134879 (lr=3.0000e-05) (hash(x)=51276369)
29
+ 200 val loss 8.2561
30
+ 200 val perplexity 3851.0483
31
+ 200 train 8.325293 (lr=7.0000e-05) (hash(x)=51276369)
32
+ 200 val loss 8.1614
33
+ 200 val perplexity 3503.0369
34
+ 200 val loss 9.4203
35
+ 200 val perplexity 12335.9941
36
+ 200 train 8.220893 (lr=5.0000e-05) (hash(x)=51276369)
37
+ 200 train 9.459747 (lr=2.0000e-05) (hash(x)=51276369)
38
+ 300 val loss 8.1348
39
+ 300 val perplexity 3411.0425
40
+ 300 train 8.079185 (lr=2.9993e-05) (hash(x)=48572728)
41
+ 300 val loss 7.7588
42
+ 300 val perplexity 2342.0601
43
+ 300 train 7.683750 (lr=6.9984e-05) (hash(x)=48572728)
44
+ 300 val loss 8.6640
45
+ 300 val perplexity 5790.9355
46
+ 300 train 8.623919 (lr=1.9995e-05) (hash(x)=48572728)
47
+ 300 val loss 7.6377
48
+ 300 val perplexity 2075.0649
49
+ 300 train 7.547552 (lr=4.9988e-05) (hash(x)=48572728)
50
+ 400 val loss 7.7796
51
+ 400 val perplexity 2391.3811
52
+ 400 train 7.544574 (lr=2.9972e-05) (hash(x)=44943400)
53
+ 400 val loss 7.5697
54
+ 400 val perplexity 1938.5656
55
+ 400 train 7.291669 (lr=6.9935e-05) (hash(x)=44943400)
56
+ 400 val loss 8.0513
57
+ 400 val perplexity 3137.8838
58
+ 400 train 7.839612 (lr=1.9982e-05) (hash(x)=44943400)
59
+ 400 val loss 7.5305
60
+ 400 val perplexity 1864.0149
61
+ 400 train 7.253582 (lr=4.9954e-05) (hash(x)=44943400)
62
+ 500 val loss 7.6533
63
+ 500 val perplexity 2107.6501
64
+ 500 train 7.540147 (lr=2.9938e-05) (hash(x)=48434590)
65
+ 500 val loss 7.4707
66
+ 500 val perplexity 1755.8274
67
+ 500 train 7.338100 (lr=6.9854e-05) (hash(x)=48434590)
68
+ 500 val loss 7.7816
69
+ 500 val perplexity 2396.0164
70
+ 500 train 7.681769 (lr=1.9958e-05) (hash(x)=48434590)
71
+ 500 val loss 7.4639
72
+ 500 val perplexity 1743.9081
73
+ 500 train 7.340280 (lr=4.9896e-05) (hash(x)=48434590)
74
+ 600 val loss 7.5785
75
+ 600 val perplexity 1955.7003
76
+ 600 train 7.409949 (lr=2.9889e-05) (hash(x)=47442266)
77
+ 600 val loss 7.3962
78
+ 600 val perplexity 1629.8575
79
+ 600 val loss 7.6537
80
+ 600 val perplexity 2108.4563
81
+ 600 train 7.210290 (lr=6.9741e-05) (hash(x)=47442266)
82
+ 600 train 7.491193 (lr=1.9926e-05) (hash(x)=47442266)
83
+ 600 val loss 7.3922
84
+ 600 val perplexity 1623.2819
85
+ 600 train 7.205268 (lr=4.9815e-05) (hash(x)=47442266)
86
+ 700 val loss 7.5302
87
+ 700 val perplexity 1863.4373
88
+ 700 train 7.585056 (lr=2.9827e-05) (hash(x)=56495712)
89
+ 700 val loss 7.5849
90
+ 700 val perplexity 1968.2036
91
+ 700 val loss 7.3470
92
+ 700 val perplexity 1551.4893
93
+ 700 train 7.630278 (lr=1.9885e-05) (hash(x)=56495712)
94
+ 700 train 7.403419 (lr=6.9596e-05) (hash(x)=56495712)
95
+ 700 val loss 7.3310
96
+ 700 val perplexity 1526.8895
97
+ 700 train 7.399583 (lr=4.9712e-05) (hash(x)=56495712)
98
+ 800 val loss 7.4954
99
+ 800 val perplexity 1799.7959
100
+ 800 train 7.490317 (lr=2.9751e-05) (hash(x)=50093738)
101
+ 800 val loss 7.3051
102
+ 800 val perplexity 1487.8765
103
+ 800 val loss 7.5465
104
+ 800 val perplexity 1894.0831
105
+ 800 train 7.297331 (lr=6.9419e-05) (hash(x)=50093738)
106
+ 800 train 7.542716 (lr=1.9834e-05) (hash(x)=50093738)
107
+ 800 val loss 7.2942
108
+ 800 val perplexity 1471.7411
109
+ 800 train 7.273575 (lr=4.9585e-05) (hash(x)=50093738)
110
+ 900 val loss 7.4639
111
+ 900 val perplexity 1744.0095
112
+ 900 train 7.444237 (lr=2.9662e-05) (hash(x)=49436061)
113
+ 900 val loss 7.2642
114
+ 900 val perplexity 1428.2000
115
+ 900 val loss 7.5119
116
+ 900 val perplexity 1829.6874
117
+ 900 train 7.232059 (lr=6.9210e-05) (hash(x)=49436061)
118
+ 900 train 7.486966 (lr=1.9774e-05) (hash(x)=49436061)
119
+ 900 val loss 7.2523
120
+ 900 val perplexity 1411.3772
121
+ 900 train 7.216028 (lr=4.9436e-05) (hash(x)=49436061)
122
+ 1000 val loss 7.4476
123
+ 1000 val perplexity 1715.7721
124
+ 1000 train 7.375908 (lr=2.9558e-05) (hash(x)=48818282)
125
+ 1000 val loss 7.4869
126
+ 1000 val perplexity 1784.4542
127
+ 1000 val loss 7.2531
128
+ 1000 val perplexity 1412.5016
129
+ 1000 train 7.416676 (lr=1.9706e-05) (hash(x)=48818282)
130
+ 1000 train 7.184549 (lr=6.8970e-05) (hash(x)=48818282)
131
+ 1000 val loss 7.2196
132
+ 1000 val perplexity 1365.9194
133
+ 1000 train 7.146063 (lr=4.9264e-05) (hash(x)=48818282)
134
+ 1100 val loss 7.4264
135
+ 1100 val perplexity 1679.7290
136
+ 1100 train 7.496390 (lr=2.9442e-05) (hash(x)=55387425)
137
+ 1100 val loss 7.4593
138
+ 1100 val perplexity 1735.9408
139
+ 1100 val loss 7.2475
140
+ 1100 val perplexity 1404.6104
141
+ 1100 train 7.533914 (lr=1.9628e-05) (hash(x)=55387425)
142
+ 1100 train 7.310481 (lr=6.8698e-05) (hash(x)=55387425)
143
+ 1100 val loss 7.1829
144
+ 1100 val perplexity 1316.6659
145
+ 1100 train 7.246019 (lr=4.9070e-05) (hash(x)=55387425)
146
+ 1200 val loss 7.4108
147
+ 1200 val perplexity 1653.7014
148
+ 1200 train 7.301041 (lr=2.9312e-05) (hash(x)=48328378)
149
+ 1200 val loss 7.4398
150
+ 1200 val perplexity 1702.3777
151
+ 1200 train 7.347461 (lr=1.9542e-05) (hash(x)=48328378)
152
+ 1200 val loss 7.2117
153
+ 1200 val perplexity 1355.1605
154
+ 1200 train 7.083104 (lr=6.8395e-05) (hash(x)=48328378)
155
+ 1200 val loss 7.1377
156
+ 1200 val perplexity 1258.4778
157
+ 1200 train 6.990996 (lr=4.8854e-05) (hash(x)=48328378)
158
+ 1300 val loss 7.3916
159
+ 1300 val perplexity 1622.2303
160
+ 1300 train 7.309278 (lr=2.9169e-05) (hash(x)=48956717)
161
+ 1300 val loss 7.4174
162
+ 1300 val perplexity 1664.6868
163
+ 1300 train 7.348650 (lr=1.9446e-05) (hash(x)=48956717)
164
+ 1300 val loss 7.1916
165
+ 1300 val perplexity 1328.1700
166
+ 1300 train 7.095232 (lr=6.8062e-05) (hash(x)=48956717)
167
+ 1300 val loss 7.1050
168
+ 1300 val perplexity 1218.0630
169
+ 1300 train 6.989187 (lr=4.8616e-05) (hash(x)=48956717)
170
+ 1400 val loss 7.3837
171
+ 1400 val perplexity 1609.5293
172
+ 1400 train 7.285535 (lr=2.9013e-05) (hash(x)=49782571)
173
+ 1400 val loss 7.4088
174
+ 1400 val perplexity 1650.4393
175
+ 1400 train 7.314537 (lr=1.9342e-05) (hash(x)=49782571)
176
+ 1400 val loss 7.1905
177
+ 1400 val perplexity 1326.7339
178
+ 1400 train 7.081223 (lr=6.7698e-05) (hash(x)=49782571)
179
+ 1400 val loss 7.0826
180
+ 1400 val perplexity 1191.0359
181
+ 1400 train 6.962979 (lr=4.8356e-05) (hash(x)=49782571)
182
+ 1500 val loss 7.3719
183
+ 1500 val perplexity 1590.6541
184
+ 1500 train 7.180953 (lr=2.8845e-05) (hash(x)=51081994)
185
+ 1500 val loss 7.3955
186
+ 1500 val perplexity 1628.5710
187
+ 1500 train 7.199613 (lr=1.9230e-05) (hash(x)=51081994)
188
+ 1500 val loss 7.1761
189
+ 1500 val perplexity 1307.8132
190
+ 1500 train 6.960799 (lr=6.7304e-05) (hash(x)=51081994)
191
+ 1500 val loss 7.0469
192
+ 1500 val perplexity 1149.2792
193
+ 1500 train 6.789298 (lr=4.8074e-05) (hash(x)=51081994)
194
+ 1600 val loss 7.3619
195
+ 1600 val perplexity 1574.7723
196
+ 1600 train 7.437032 (lr=2.8663e-05) (hash(x)=48791204)
197
+ 1600 val loss 7.3823
198
+ 1600 val perplexity 1607.2921
199
+ 1600 train 7.458527 (lr=1.9109e-05) (hash(x)=48791204)
200
+ 1600 val loss 7.1631
201
+ 1600 val perplexity 1290.9561
202
+ 1600 train 7.228574 (lr=6.6881e-05) (hash(x)=48791204)
203
+ 1600 val loss 7.0259
204
+ 1600 val perplexity 1125.4299
205
+ 1600 train 7.080369 (lr=4.7772e-05) (hash(x)=48791204)
206
+ 1700 val loss 7.3512
207
+ 1700 val perplexity 1557.9872
208
+ 1700 train 7.064727 (lr=2.8469e-05) (hash(x)=46425011)
209
+ 1700 val loss 7.3648
210
+ 1700 val perplexity 1579.4250
211
+ 1700 train 7.079679 (lr=1.8979e-05) (hash(x)=46425011)
212
+ 1700 val loss 7.1553
213
+ 1700 val perplexity 1280.8689
214
+ 1700 train 6.839215 (lr=6.6428e-05) (hash(x)=46425011)
215
+ 1700 val loss 6.9937
216
+ 1700 val perplexity 1089.7976
217
+ 1700 train 6.657190 (lr=4.7448e-05) (hash(x)=46425011)
218
+ 1800 val loss 7.3486
219
+ 1800 val perplexity 1553.9711
220
+ 1800 train 7.003738 (lr=2.8263e-05) (hash(x)=44919141)
221
+ 1800 val loss 7.3539
222
+ 1800 val perplexity 1562.3319
223
+ 1800 train 7.010586 (lr=1.8842e-05) (hash(x)=44919141)
224
+ 1800 val loss 7.1675
225
+ 1800 val perplexity 1296.6541
226
+ 1800 train 6.812972 (lr=6.5947e-05) (hash(x)=44919141)
227
+ 1800 val loss 6.9646
228
+ 1800 val perplexity 1058.4592
229
+ 1800 train 6.574129 (lr=4.7105e-05) (hash(x)=44919141)
230
+ 1900 val loss 7.3377
231
+ 1900 val perplexity 1537.2053
232
+ 1900 train 7.431317 (lr=2.8044e-05) (hash(x)=48299675)
233
+ 1900 val loss 7.3345
234
+ 1900 val perplexity 1532.3342
235
+ 1900 train 7.430192 (lr=1.8696e-05) (hash(x)=48299675)
236
+ 1900 val loss 7.1367
237
+ 1900 val perplexity 1257.2877
238
+ 1900 val loss 6.9399
239
+ 1900 val perplexity 1032.6664
240
+ 1900 train 7.246999 (lr=6.5437e-05) (hash(x)=48299675)
241
+ 1900 train 7.059433 (lr=4.6741e-05) (hash(x)=48299675)
242
+ 2000 val loss 7.3206
243
+ 2000 val perplexity 1511.1768
244
+ 2000 train 8.112602 (lr=2.7814e-05) (hash(x)=61804797)
245
+ 2000 val loss 7.3156
246
+ 2000 val perplexity 1503.5886
247
+ 2000 train 8.138527 (lr=1.8543e-05) (hash(x)=61804797)
248
+ 2000 val loss 7.0982
249
+ 2000 val perplexity 1209.7937
250
+ 2000 val loss 6.9065
251
+ 2000 val perplexity 998.7697
252
+ 2000 train 7.933258 (lr=6.4900e-05) (hash(x)=61804797)
253
+ 2000 train 7.793611 (lr=4.6357e-05) (hash(x)=61804797)
254
+ 2100 val loss 7.3054
255
+ 2100 val perplexity 1488.3832
256
+ 2100 train 7.394943 (lr=2.7572e-05) (hash(x)=53638030)
257
+ 2100 val loss 7.2988
258
+ 2100 val perplexity 1478.5387
259
+ 2100 train 7.388790 (lr=1.8382e-05) (hash(x)=53638030)
260
+ 2100 val loss 7.0771
261
+ 2100 val perplexity 1184.5203
262
+ 2100 val loss 6.8708
263
+ 2100 val perplexity 963.7128
264
+ 2100 train 7.168970 (lr=6.4335e-05) (hash(x)=53638030)
265
+ 2100 train 6.969800 (lr=4.5954e-05) (hash(x)=53638030)
266
+ 2200 val loss 7.2858
267
+ 2200 val perplexity 1459.4324
268
+ 2200 train 7.543324 (lr=2.7319e-05) (hash(x)=56364593)
269
+ 2200 val loss 7.2805
270
+ 2200 val perplexity 1451.6970
271
+ 2200 train 7.541931 (lr=1.8213e-05) (hash(x)=56364593)
272
+ 2200 val loss 6.8395
273
+ 2200 val perplexity 934.0424
274
+ 2200 train 7.073711 (lr=4.5532e-05) (hash(x)=56364593)
275
+ 2200 val loss 7.0551
276
+ 2200 val perplexity 1158.7914
277
+ 2200 train 7.278820 (lr=6.3745e-05) (hash(x)=56364593)
278
+ 2300 val loss 7.2711
279
+ 2300 val perplexity 1438.1741
280
+ 2300 train 7.410802 (lr=2.7055e-05) (hash(x)=52945580)
281
+ 2300 val loss 7.2713
282
+ 2300 val perplexity 1438.3915
283
+ 2300 train 7.406998 (lr=1.8036e-05) (hash(x)=52945580)
284
+ 2300 val loss 6.8227
285
+ 2300 val perplexity 918.5023
286
+ 2300 train 6.977562 (lr=4.5091e-05) (hash(x)=52945580)
287
+ 2300 val loss 7.0360
288
+ 2300 val perplexity 1136.8781
289
+ 2300 train 7.169257 (lr=6.3128e-05) (hash(x)=52945580)
290
+ 2400 val loss 7.2607
291
+ 2400 val perplexity 1423.3226
292
+ 2400 train 7.313107 (lr=2.6780e-05) (hash(x)=54595211)
293
+ 2400 val loss 7.2538
294
+ 2400 val perplexity 1413.4213
295
+ 2400 train 7.296937 (lr=1.7853e-05) (hash(x)=54595211)
296
+ 2400 val loss 6.7938
297
+ 2400 val perplexity 892.2969
298
+ 2400 train 6.862321 (lr=4.4633e-05) (hash(x)=54595211)
299
+ 2400 val loss 7.0215
300
+ 2400 val perplexity 1120.5073
301
+ 2400 train 7.083766 (lr=6.2486e-05) (hash(x)=54595211)
302
+ 2500 val loss 7.2483
303
+ 2500 val perplexity 1405.7004
304
+ 2500 train 7.270072 (lr=2.6494e-05) (hash(x)=52323363)
305
+ 2500 val loss 7.2439
306
+ 2500 val perplexity 1399.5399
307
+ 2500 train 7.261842 (lr=1.7663e-05) (hash(x)=52323363)
308
+ 2500 val loss 6.7793
309
+ 2500 val perplexity 879.4120
310
+ 2500 val loss 7.0116
311
+ 2500 val perplexity 1109.4362
312
+ 2500 train 7.030132 (lr=6.1819e-05) (hash(x)=52323363)
313
+ 2500 train 6.792108 (lr=4.4156e-05) (hash(x)=52323363)
314
+ 2600 val loss 7.2389
315
+ 2600 val perplexity 1392.5264
316
+ 2600 train 6.972273 (lr=2.6198e-05) (hash(x)=42652707)
317
+ 2600 val loss 7.2260
318
+ 2600 val perplexity 1374.6908
319
+ 2600 train 6.964200 (lr=1.7465e-05) (hash(x)=42652707)
320
+ 2600 val loss 6.7583
321
+ 2600 val perplexity 861.1493
322
+ 2600 train 6.511856 (lr=4.3663e-05) (hash(x)=42652707)
323
+ 2600 val loss 7.0148
324
+ 2600 val perplexity 1113.0165
325
+ 2600 train 6.752173 (lr=6.1128e-05) (hash(x)=42652707)
326
+ 2700 val loss 7.2357
327
+ 2700 val perplexity 1388.0681
328
+ 2700 train 6.921242 (lr=2.5892e-05) (hash(x)=45687855)
329
+ 2700 val loss 7.2168
330
+ 2700 val perplexity 1362.0977
331
+ 2700 val loss 6.7482
332
+ 2700 val perplexity 852.5545
333
+ 2700 train 6.905418 (lr=1.7261e-05) (hash(x)=45687855)
334
+ 2700 train 6.414542 (lr=4.3153e-05) (hash(x)=45687855)
335
+ 2700 val loss 6.9994
336
+ 2700 val perplexity 1095.9656
337
+ 2700 train 6.681432 (lr=6.0414e-05) (hash(x)=45687855)
338
+ 2800 val loss 7.2221
339
+ 2800 val perplexity 1369.3673
340
+ 2800 train 7.194708 (lr=2.5576e-05) (hash(x)=49962830)
341
+ 2800 val loss 6.7228
342
+ 2800 val perplexity 831.1389
343
+ 2800 val loss 7.2040
344
+ 2800 val perplexity 1344.7562
345
+ 2800 train 6.732316 (lr=4.2627e-05) (hash(x)=49962830)
346
+ 2800 train 7.177933 (lr=1.7051e-05) (hash(x)=49962830)
347
+ 2800 val loss 7.0031
348
+ 2800 val perplexity 1099.9865
349
+ 2800 train 6.995296 (lr=5.9677e-05) (hash(x)=49962830)
350
+ 2900 val loss 7.2305
351
+ 2900 val perplexity 1380.9703
352
+ 2900 train 7.279706 (lr=2.5251e-05) (hash(x)=52958451)
353
+ 2900 val loss 6.7024
354
+ 2900 val perplexity 814.3538
355
+ 2900 val loss 7.1967
356
+ 2900 val perplexity 1335.0525
357
+ 2900 train 6.726287 (lr=4.2085e-05) (hash(x)=52958451)
358
+ 2900 train 7.253380 (lr=1.6834e-05) (hash(x)=52958451)
359
+ 2900 val loss 6.9826
360
+ 2900 val perplexity 1077.7458
361
+ 2900 train 7.006102 (lr=5.8919e-05) (hash(x)=52958451)
362
+ 3000 val loss 7.2079
363
+ 3000 val perplexity 1349.9924
364
+ 3000 train 7.143060 (lr=2.4917e-05) (hash(x)=48896990)
365
+ 3000 val loss 6.6892
366
+ 3000 val perplexity 803.7150
367
+ 3000 train 6.626854 (lr=4.1529e-05) (hash(x)=48896990)
368
+ 3000 val loss 7.1807
369
+ 3000 val perplexity 1313.8224
370
+ 3000 train 7.114826 (lr=1.6611e-05) (hash(x)=48896990)
371
+ 3000 val loss 6.9730
372
+ 3000 val perplexity 1067.4177
373
+ 3000 train 6.921494 (lr=5.8140e-05) (hash(x)=48896990)
374
+ 3100 val loss 7.2022
375
+ 3100 val perplexity 1342.3608
376
+ 3100 train 7.150681 (lr=2.4574e-05) (hash(x)=47064930)
377
+ 3100 val loss 6.6703
378
+ 3100 val perplexity 788.6352
379
+ 3100 train 6.664508 (lr=4.0957e-05) (hash(x)=47064930)
380
+ 3100 val loss 7.1689
381
+ 3100 val perplexity 1298.4149
382
+ 3100 train 7.118716 (lr=1.6383e-05) (hash(x)=47064930)
383
+ 3100 val loss 6.9743
384
+ 3100 val perplexity 1068.8025
385
+ 3100 train 6.925107 (lr=5.7340e-05) (hash(x)=47064930)
386
+ 3200 val loss 7.1856
387
+ 3200 val perplexity 1320.2545
388
+ 3200 train 6.912687 (lr=2.4224e-05) (hash(x)=45145921)
389
+ 3200 val loss 6.6590
390
+ 3200 val perplexity 779.7889
391
+ 3200 train 6.334053 (lr=4.0373e-05) (hash(x)=45145921)
392
+ 3200 val loss 7.1583
393
+ 3200 val perplexity 1284.7445
394
+ 3200 train 6.887141 (lr=1.6149e-05) (hash(x)=45145921)
395
+ 3200 val loss 6.9708
396
+ 3200 val perplexity 1065.0846
397
+ 3200 train 6.685108 (lr=5.6522e-05) (hash(x)=45145921)
398
+ 3300 val loss 7.1754
399
+ 3300 val perplexity 1306.8682
400
+ 3300 train 7.026076 (lr=2.3865e-05) (hash(x)=48085018)
401
+ 3300 val loss 6.6375
402
+ 3300 val perplexity 763.1765
403
+ 3300 train 6.467580 (lr=3.9775e-05) (hash(x)=48085018)
404
+ 3300 val loss 7.1463
405
+ 3300 val perplexity 1269.4426
406
+ 3300 train 6.998851 (lr=1.5910e-05) (hash(x)=48085018)
407
+ 3300 val loss 6.9690
408
+ 3300 val perplexity 1063.1316
409
+ 3300 train 6.808588 (lr=5.5684e-05) (hash(x)=48085018)
410
+ 3400 val loss 7.1656
411
+ 3400 val perplexity 1294.1870
412
+ 3400 train 7.203237 (lr=2.3498e-05) (hash(x)=49596053)
413
+ 3400 val loss 6.6357
414
+ 3400 val perplexity 761.8221
415
+ 3400 train 6.642577 (lr=3.9164e-05) (hash(x)=49596053)
416
+ 3400 val loss 7.1407
417
+ 3400 val perplexity 1262.3193
418
+ 3400 train 7.180227 (lr=1.5666e-05) (hash(x)=49596053)
419
+ 3400 val loss 6.9654
420
+ 3400 val perplexity 1059.3287
421
+ 3400 train 6.969168 (lr=5.4829e-05) (hash(x)=49596053)
422
+ 3500 val loss 7.1442
423
+ 3500 val perplexity 1266.8002
424
+ 3500 train 6.707533 (lr=2.3125e-05) (hash(x)=41135634)
425
+ 3500 val loss 6.6047
426
+ 3500 val perplexity 738.5925
427
+ 3500 train 6.213636 (lr=3.8541e-05) (hash(x)=41135634)
428
+ 3500 val loss 7.1232
429
+ 3500 val perplexity 1240.4121
430
+ 3500 train 6.685036 (lr=1.5416e-05) (hash(x)=41135634)
431
+ 3500 val loss 6.9724
432
+ 3500 val perplexity 1066.7339
433
+ 3500 train 6.579717 (lr=5.3958e-05) (hash(x)=41135634)
434
+ 3600 val loss 7.1358
435
+ 3600 val perplexity 1256.1060
436
+ 3600 train 7.137975 (lr=2.2744e-05) (hash(x)=54286330)
437
+ 3600 val loss 6.5831
438
+ 3600 val perplexity 722.7491
439
+ 3600 train 6.582256 (lr=3.7907e-05) (hash(x)=54286330)
440
+ 3600 val loss 7.1126
441
+ 3600 val perplexity 1227.3097
442
+ 3600 train 7.123320 (lr=1.5163e-05) (hash(x)=54286330)
443
+ 3600 val loss 6.9504
444
+ 3600 val perplexity 1043.5963
445
+ 3600 train 6.965904 (lr=5.3070e-05) (hash(x)=54286330)
446
+ 3700 val loss 7.1273
447
+ 3700 val perplexity 1245.5331
448
+ 3700 train 7.047629 (lr=2.2357e-05) (hash(x)=57061993)
449
+ 3700 val loss 6.5909
450
+ 3700 val perplexity 728.4212
451
+ 3700 train 6.439983 (lr=3.7262e-05) (hash(x)=57061993)
452
+ 3700 val loss 7.1064
453
+ 3700 val perplexity 1219.7939
454
+ 3700 train 7.037267 (lr=1.4905e-05) (hash(x)=57061993)
455
+ 3700 val loss 6.9561
456
+ 3700 val perplexity 1049.5469
457
+ 3700 train 6.876348 (lr=5.2167e-05) (hash(x)=57061993)
458
+ 3800 val loss 7.1073
459
+ 3800 val perplexity 1220.8978
460
+ 3800 train 6.682832 (lr=2.1965e-05) (hash(x)=46544884)
461
+ 3800 val loss 6.5705
462
+ 3800 val perplexity 713.7065
463
+ 3800 train 6.108221 (lr=3.6608e-05) (hash(x)=46544884)
464
+ 3800 val loss 7.0976
465
+ 3800 val perplexity 1209.0325
466
+ 3800 train 6.682347 (lr=1.4643e-05) (hash(x)=46544884)
467
+ 3800 val loss 6.9326
468
+ 3800 val perplexity 1025.1139
469
+ 3800 train 6.517674 (lr=5.1251e-05) (hash(x)=46544884)
470
+ 3900 val loss 7.0991
471
+ 3900 val perplexity 1210.8584
472
+ 3900 train 6.966492 (lr=2.1566e-05) (hash(x)=48937435)
473
+ 3900 val loss 6.5600
474
+ 3900 val perplexity 706.3019
475
+ 3900 train 6.408988 (lr=3.5944e-05) (hash(x)=48937435)
476
+ 3900 val loss 7.0855
477
+ 3900 val perplexity 1194.5229
478
+ 3900 train 6.953877 (lr=1.4377e-05) (hash(x)=48937435)
479
+ 3900 val loss 6.9300
480
+ 3900 val perplexity 1022.4562
481
+ 3900 train 6.793073 (lr=5.0321e-05) (hash(x)=48937435)
482
+ 4000 val loss 7.0941
483
+ 4000 val perplexity 1204.8864
484
+ 4000 train 7.138962 (lr=2.1162e-05) (hash(x)=54466186)
485
+ 4000 val loss 6.5605
486
+ 4000 val perplexity 706.6361
487
+ 4000 train 6.604020 (lr=3.5271e-05) (hash(x)=54466186)
488
+ 4000 val loss 7.0805
489
+ 4000 val perplexity 1188.5492
490
+ 4000 train 7.115824 (lr=1.4108e-05) (hash(x)=54466186)
491
+ 4000 val loss 6.9326
492
+ 4000 val perplexity 1025.1711
493
+ 4000 train 6.948184 (lr=4.9379e-05) (hash(x)=54466186)
494
+ 4100 val loss 7.0807
495
+ 4100 val perplexity 1188.8593
496
+ 4100 train 6.931328 (lr=2.0754e-05) (hash(x)=51079995)
497
+ 4100 val loss 6.5394
498
+ 4100 val perplexity 691.8721
499
+ 4100 train 6.342807 (lr=3.4590e-05) (hash(x)=51079995)
500
+ 4100 val loss 7.0660
501
+ 4100 val perplexity 1171.4478
502
+ 4100 train 6.912895 (lr=1.3836e-05) (hash(x)=51079995)
503
+ 4100 val loss 6.9133
504
+ 4100 val perplexity 1005.5554
505
+ 4100 train 6.757079 (lr=4.8426e-05) (hash(x)=51079995)
506
+ 4200 val loss 7.0770
507
+ 4200 val perplexity 1184.3700
508
+ 4200 train 7.238811 (lr=2.0341e-05) (hash(x)=56495581)
509
+ 4200 val loss 6.5314
510
+ 4200 val perplexity 686.3465
511
+ 4200 train 6.662952 (lr=3.3902e-05) (hash(x)=56495581)
512
+ 4200 val loss 7.0606
513
+ 4200 val perplexity 1165.1755
514
+ 4200 train 7.224622 (lr=1.3561e-05) (hash(x)=56495581)
515
+ 4200 val loss 6.9211
516
+ 4200 val perplexity 1013.4227
517
+ 4200 train 7.085524 (lr=4.7463e-05) (hash(x)=56495581)
518
+ 4300 val loss 7.0596
519
+ 4300 val perplexity 1163.9338
520
+ 4300 train 6.913306 (lr=1.9924e-05) (hash(x)=45095478)
521
+ 4300 val loss 6.5033
522
+ 4300 val perplexity 667.3305
523
+ 4300 train 6.326633 (lr=3.3207e-05) (hash(x)=45095478)
524
+ 4300 val loss 7.0338
525
+ 4300 val perplexity 1134.3417
526
+ 4300 train 6.887545 (lr=1.3283e-05) (hash(x)=45095478)
527
+ 4300 val loss 6.8885
528
+ 4300 val perplexity 980.9670
529
+ 4300 train 6.741586 (lr=4.6490e-05) (hash(x)=45095478)
530
+ 4400 val loss 7.0466
531
+ 4400 val perplexity 1148.9017
532
+ 4400 train 6.998871 (lr=1.9504e-05) (hash(x)=49954028)
533
+ 4400 val loss 6.4871
534
+ 4400 val perplexity 656.6105
535
+ 4400 train 6.504465 (lr=3.2507e-05) (hash(x)=49954028)
536
+ 4400 val loss 7.0249
537
+ 4400 val perplexity 1124.2563
538
+ 4400 train 6.952001 (lr=1.3003e-05) (hash(x)=49954028)
539
+ 4400 val loss 6.8805
540
+ 4400 val perplexity 973.1015
541
+ 4400 train 6.864935 (lr=4.5509e-05) (hash(x)=49954028)
542
+ 4500 val loss 7.0395
543
+ 4500 val perplexity 1140.8114
544
+ 4500 train 6.956896 (lr=1.9081e-05) (hash(x)=48975821)
545
+ 4500 val loss 6.4814
546
+ 4500 val perplexity 652.8572
547
+ 4500 train 6.403913 (lr=3.1801e-05) (hash(x)=48975821)
548
+ 4500 val loss 7.0197
549
+ 4500 val perplexity 1118.4393
550
+ 4500 train 6.932141 (lr=1.2720e-05) (hash(x)=48975821)
551
+ 4500 val loss 6.8784
552
+ 4500 val perplexity 971.0259
553
+ 4500 train 6.792545 (lr=4.4521e-05) (hash(x)=48975821)
554
+ 4600 val loss 7.0202
555
+ 4600 val perplexity 1119.0405
556
+ 4600 train 7.099740 (lr=1.8655e-05) (hash(x)=49293028)
557
+ 4600 val loss 6.4562
558
+ 4600 val perplexity 636.6337
559
+ 4600 train 6.527642 (lr=3.1091e-05) (hash(x)=49293028)
560
+ 4600 val loss 7.0010
561
+ 4600 val perplexity 1097.6952
562
+ 4600 train 7.073569 (lr=1.2436e-05) (hash(x)=49293028)
563
+ 4600 val loss 6.8440
564
+ 4600 val perplexity 938.2322
565
+ 4600 train 6.921262 (lr=4.3527e-05) (hash(x)=49293028)
566
+ 4700 val loss 7.0042
567
+ 4700 val perplexity 1101.2261
568
+ 4700 train 6.936911 (lr=1.8226e-05) (hash(x)=48077753)
569
+ 4700 val loss 6.4481
570
+ 4700 val perplexity 631.5083
571
+ 4700 train 6.353138 (lr=3.0377e-05) (hash(x)=48077753)
572
+ 4700 val loss 6.9942
573
+ 4700 val perplexity 1090.2617
574
+ 4700 train 6.922954 (lr=1.2151e-05) (hash(x)=48077753)
575
+ 4800 val loss 6.9923
576
+ 4800 val perplexity 1088.2677
577
+ 4800 train 6.816248 (lr=1.7796e-05) (hash(x)=45037879)
578
+ 4700 val loss 6.8457
579
+ 4700 val perplexity 939.7895
580
+ 4700 train 6.764469 (lr=4.2528e-05) (hash(x)=48077753)
581
+ 4800 val loss 6.4412
582
+ 4800 val perplexity 627.1833
583
+ 4800 train 6.349098 (lr=2.9661e-05) (hash(x)=45037879)
584
+ 4800 val loss 6.9846
585
+ 4800 val perplexity 1079.8929
586
+ 4800 train 6.818547 (lr=1.1864e-05) (hash(x)=45037879)
587
+ 4900 val loss 6.9870
588
+ 4900 val perplexity 1082.4578
589
+ 4900 train 6.892083 (lr=1.7365e-05) (hash(x)=50320154)
590
+ 4800 val loss 6.8367
591
+ 4800 val perplexity 931.3828
592
+ 4800 train 6.679591 (lr=4.1525e-05) (hash(x)=45037879)
593
+ 4900 val loss 6.4299
594
+ 4900 val perplexity 620.1333
595
+ 4900 train 6.340251 (lr=2.8942e-05) (hash(x)=50320154)
596
+ 4900 val loss 6.9748
597
+ 4900 val perplexity 1069.3352
598
+ 4900 train 6.880073 (lr=1.1577e-05) (hash(x)=50320154)
599
+ 5000 val loss 6.9746
600
+ 5000 val perplexity 1069.1670
601
+ 4900 val loss 6.8297
602
+ 4900 val perplexity 924.9001
603
+ 4900 train 6.744113 (lr=4.0518e-05) (hash(x)=50320154)
604
+ 5000 train 7.202687 (lr=1.6933e-05) (hash(x)=55024523)
605
+ 5000 val loss 6.4126
606
+ 5000 val perplexity 609.4899
607
+ 5000 train 6.674364 (lr=2.8221e-05) (hash(x)=55024523)
608
+ 5000 val loss 6.9676
609
+ 5000 val perplexity 1061.6569
610
+ 5000 train 7.191324 (lr=1.1288e-05) (hash(x)=55024523)
611
+ 5100 val loss 6.9586
612
+ 5100 val perplexity 1052.1821
613
+ 5100 train 6.892559 (lr=1.6500e-05) (hash(x)=47570607)
614
+ 5000 val loss 6.8270
615
+ 5000 val perplexity 922.4503
616
+ 5000 train 7.052451 (lr=3.9510e-05) (hash(x)=55024523)
617
+ 5100 val loss 6.3987
618
+ 5100 val perplexity 601.0562
619
+ 5100 train 6.313329 (lr=2.7500e-05) (hash(x)=47570607)
620
+ 5100 val loss 6.9547
621
+ 5100 val perplexity 1048.0336
622
+ 5100 train 6.890707 (lr=1.1000e-05) (hash(x)=47570607)
623
+ 5200 val loss 6.9542
624
+ 5200 val perplexity 1047.5540
625
+ 5200 train 7.065962 (lr=1.6067e-05) (hash(x)=54311349)
626
+ 5100 val loss 6.8143
627
+ 5100 val perplexity 910.7801
628
+ 5100 train 6.739725 (lr=3.8500e-05) (hash(x)=47570607)
629
+ 5200 val loss 6.3979
630
+ 5200 val perplexity 600.5731
631
+ 5200 train 6.529141 (lr=2.6779e-05) (hash(x)=54311349)
632
+ 5200 val loss 6.9452
633
+ 5200 val perplexity 1038.1858
634
+ 5200 train 7.059914 (lr=1.0712e-05) (hash(x)=54311349)
635
+ 5300 val loss 6.9438
636
+ 5300 val perplexity 1036.7140
637
+ 5300 train 6.836552 (lr=1.5635e-05) (hash(x)=47178524)
638
+ 5200 val loss 6.8222
639
+ 5200 val perplexity 917.9576
640
+ 5200 train 6.950159 (lr=3.7490e-05) (hash(x)=54311349)
641
+ 5300 val loss 6.3854
642
+ 5300 val perplexity 593.1091
643
+ 5300 train 6.271163 (lr=2.6058e-05) (hash(x)=47178524)
644
+ 5400 val loss 6.9342
645
+ 5400 val perplexity 1026.7831
646
+ 5300 val loss 6.9379
647
+ 5300 val perplexity 1030.6195
648
+ 5400 train 6.933376 (lr=1.5204e-05) (hash(x)=49425088)
649
+ 5300 train 6.834307 (lr=1.0423e-05) (hash(x)=47178524)
650
+ 5300 val loss 6.8130
651
+ 5300 val perplexity 909.6308
652
+ 5300 train 6.703873 (lr=3.6482e-05) (hash(x)=47178524)
653
+ 5400 val loss 6.3748
654
+ 5400 val perplexity 586.8486
655
+ 5400 train 6.402876 (lr=2.5339e-05) (hash(x)=49425088)
656
+ 5500 val loss 6.9271
657
+ 5500 val perplexity 1019.5517
658
+ 5500 train 6.762579 (lr=1.4774e-05) (hash(x)=46383189)
659
+ 5400 val loss 6.9295
660
+ 5400 val perplexity 1021.9454
661
+ 5400 train 6.937221 (lr=1.0136e-05) (hash(x)=49425088)
662
+ 5400 val loss 6.7971
663
+ 5400 val perplexity 895.2819
664
+ 5400 train 6.790603 (lr=3.5475e-05) (hash(x)=49425088)
665
+ 5500 val loss 6.3650
666
+ 5500 val perplexity 581.1469
667
+ 5500 train 6.185537 (lr=2.4623e-05) (hash(x)=46383189)
668
+ 5600 val loss 6.9271
669
+ 5600 val perplexity 1019.5128
670
+ 5600 train 7.123457 (lr=1.4345e-05) (hash(x)=54084990)
671
+ 5500 val loss 6.9233
672
+ 5500 val perplexity 1015.6611
673
+ 5500 train 6.753473 (lr=9.8491e-06) (hash(x)=46383189)
674
+ 5500 val loss 6.8003
675
+ 5500 val perplexity 898.1262
676
+ 5500 train 6.640832 (lr=3.4472e-05) (hash(x)=46383189)
677
+ 5600 val loss 6.3704
678
+ 5600 val perplexity 584.3188
679
+ 5600 train 6.532243 (lr=2.3909e-05) (hash(x)=54084990)
680
+ 5700 val loss 6.9151
681
+ 5700 val perplexity 1007.4103
682
+ 5700 train 6.593955 (lr=1.3919e-05) (hash(x)=47384182)
683
+ 5600 val loss 6.9220
684
+ 5600 val perplexity 1014.3364
685
+ 5600 train 7.129646 (lr=9.5636e-06) (hash(x)=54084990)
686
+ 5600 val loss 6.8021
687
+ 5600 val perplexity 899.7040
688
+ 5600 train 7.001111 (lr=3.3473e-05) (hash(x)=54084990)
689
+ 5700 val loss 6.3494
690
+ 5700 val perplexity 572.1433
691
+ 5700 train 6.033447 (lr=2.3199e-05) (hash(x)=47384182)
692
+ 5800 val loss 6.9078
693
+ 5800 val perplexity 1000.0802
694
+ 5800 train 6.791445 (lr=1.3496e-05) (hash(x)=51683744)
695
+ 5700 val loss 6.9151
696
+ 5700 val perplexity 1007.3843
697
+ 5700 train 6.592525 (lr=9.2796e-06) (hash(x)=47384182)
698
+ 5700 val loss 6.7996
699
+ 5700 val perplexity 897.4965
700
+ 5700 train 6.484940 (lr=3.2479e-05) (hash(x)=47384182)
701
+ 5800 val loss 6.3438
702
+ 5800 val perplexity 568.9437
703
+ 5800 train 6.210127 (lr=2.2493e-05) (hash(x)=51683744)
704
+ 5900 val loss 6.9021
705
+ 5900 val perplexity 994.3759
706
+ 5900 train 6.890203 (lr=1.3076e-05) (hash(x)=52798836)
707
+ 5800 val loss 6.9085
708
+ 5800 val perplexity 1000.7233
709
+ 5800 train 6.792042 (lr=8.9973e-06) (hash(x)=51683744)
710
+ 5800 val loss 6.7971
711
+ 5800 val perplexity 895.2123
712
+ 5800 train 6.675380 (lr=3.1491e-05) (hash(x)=51683744)
713
+ 5900 val loss 6.3346
714
+ 5900 val perplexity 563.7436
715
+ 5900 train 6.348949 (lr=2.1793e-05) (hash(x)=52798836)
716
+ 6000 val loss 6.8993
717
+ 6000 val perplexity 991.6287
718
+ 6000 train 6.782576 (lr=1.2659e-05) (hash(x)=50945000)
719
+ 5900 val loss 6.9018
720
+ 5900 val perplexity 994.1043
721
+ 5900 train 6.886106 (lr=8.7171e-06) (hash(x)=52798836)
722
+ 5900 val loss 6.7984
723
+ 5900 val perplexity 896.4216
724
+ 5900 train 6.783849 (lr=3.0510e-05) (hash(x)=52798836)
725
+ 6000 val loss 6.3380
726
+ 6000 val perplexity 565.6832
727
+ 6000 train 6.199138 (lr=2.1098e-05) (hash(x)=50945000)
728
+ 6100 val loss 6.8996
729
+ 6100 val perplexity 991.9120
730
+ 6100 train 6.640050 (lr=1.2246e-05) (hash(x)=48964427)
731
+ 6000 val loss 6.8999
732
+ 6000 val perplexity 992.2005
733
+ 6000 train 6.782024 (lr=8.4393e-06) (hash(x)=50945000)
734
+ 6000 val loss 6.7892
735
+ 6000 val perplexity 888.1768
736
+ 6000 train 6.678681 (lr=2.9537e-05) (hash(x)=50945000)
737
+ 6100 val loss 6.3275
738
+ 6100 val perplexity 559.7340
739
+ 6100 train 6.017313 (lr=2.0410e-05) (hash(x)=48964427)
740
+ 6200 val loss 6.8946
741
+ 6200 val perplexity 986.9628
742
+ 6200 train 6.686383 (lr=1.1838e-05) (hash(x)=47148610)
743
+ 6100 val loss 6.8952
744
+ 6100 val perplexity 987.5494
745
+ 6100 train 6.639012 (lr=8.1640e-06) (hash(x)=48964427)
746
+ 6100 val loss 6.7925
747
+ 6100 val perplexity 891.1420
748
+ 6100 train 6.528467 (lr=2.8574e-05) (hash(x)=48964427)
749
+ 6200 val loss 6.3256
750
+ 6200 val perplexity 558.7103
751
+ 6200 train 6.094476 (lr=1.9729e-05) (hash(x)=47148610)
752
+ 6300 val loss 6.8914
753
+ 6300 val perplexity 983.7978
754
+ 6300 train 6.825513 (lr=1.1434e-05) (hash(x)=50860553)
755
+ 6200 val loss 6.8907
756
+ 6200 val perplexity 983.0869
757
+ 6200 train 6.685598 (lr=7.8917e-06) (hash(x)=47148610)
758
+ 6200 val loss 6.7949
759
+ 6200 val perplexity 893.2441
760
+ 6200 train 6.577856 (lr=2.7621e-05) (hash(x)=47148610)
761
+ 6300 val loss 6.3159
762
+ 6300 val perplexity 553.2820
763
+ 6300 train 6.251698 (lr=1.9056e-05) (hash(x)=50860553)
764
+ 6400 val loss 6.8844
765
+ 6400 val perplexity 976.9618
766
+ 6400 train 6.688293 (lr=1.1035e-05) (hash(x)=50553992)
767
+ 6300 val loss 6.8884
768
+ 6300 val perplexity 980.7968
769
+ 6300 train 6.821770 (lr=7.6226e-06) (hash(x)=50860553)
770
+ 6300 val loss 6.7858
771
+ 6300 val perplexity 885.1630
772
+ 6300 train 6.732771 (lr=2.6679e-05) (hash(x)=50860553)
773
+ 6400 val loss 6.3137
774
+ 6400 val perplexity 552.0735
775
+ 6400 train 6.103390 (lr=1.8392e-05) (hash(x)=50553992)
776
+ 6500 val loss 6.8819
777
+ 6500 val perplexity 974.4667
778
+ 6500 train 6.751459 (lr=1.0643e-05) (hash(x)=52521332)
779
+ 6400 val loss 6.8830
780
+ 6400 val perplexity 975.5029
781
+ 6400 train 6.691499 (lr=7.3569e-06) (hash(x)=50553992)
782
+ 6400 val loss 6.7848
783
+ 6400 val perplexity 884.3045
784
+ 6400 train 6.593803 (lr=2.5749e-05) (hash(x)=50553992)
785
+ 6500 val loss 6.3159
786
+ 6500 val perplexity 553.3113
787
+ 6500 train 6.148011 (lr=1.7738e-05) (hash(x)=52521332)
788
+ 6600 val loss 6.8706
789
+ 6600 val perplexity 963.4868
790
+ 6600 train 7.219240 (lr=1.0256e-05) (hash(x)=52609843)
791
+ 6500 val loss 6.8780
792
+ 6500 val perplexity 970.6671
793
+ 6500 train 6.759769 (lr=7.0950e-06) (hash(x)=52521332)
794
+ 6500 val loss 6.7864
795
+ 6500 val perplexity 885.7056
796
+ 6500 train 6.640779 (lr=2.4833e-05) (hash(x)=52521332)
797
+ 6600 val loss 6.2923
798
+ 6600 val perplexity 540.3746
799
+ 6600 train 6.663771 (lr=1.7093e-05) (hash(x)=52609843)
800
+ 6700 val loss 6.8572
801
+ 6700 val perplexity 950.7396
802
+ 6700 train 6.936516 (lr=9.8753e-06) (hash(x)=53632957)
803
+ 6600 val loss 6.8663
804
+ 6600 val perplexity 959.3960
805
+ 6600 train 7.209294 (lr=6.8372e-06) (hash(x)=52609843)
806
+ 6600 val loss 6.7701
807
+ 6600 val perplexity 871.4236
808
+ 6600 train 7.132092 (lr=2.3930e-05) (hash(x)=52609843)
809
+ 6700 val loss 6.2836
810
+ 6700 val perplexity 535.7344
811
+ 6700 train 6.338120 (lr=1.6459e-05) (hash(x)=53632957)
812
+ 6800 val loss 6.8541
813
+ 6800 val perplexity 947.7671
814
+ 6800 train 6.899356 (lr=9.5017e-06) (hash(x)=46927608)
815
+ 6700 val loss 6.8581
816
+ 6700 val perplexity 951.5292
817
+ 6700 train 6.933838 (lr=6.5835e-06) (hash(x)=53632957)
818
+ 6700 val loss 6.7589
819
+ 6700 val perplexity 861.6755
820
+ 6700 train 6.818271 (lr=2.3042e-05) (hash(x)=53632957)
821
+ 6800 val loss 6.2808
822
+ 6800 val perplexity 534.2219
823
+ 6800 train 6.318456 (lr=1.5836e-05) (hash(x)=46927608)
824
+ 6900 val loss 6.8460
825
+ 6900 val perplexity 940.1440
826
+ 6900 train 6.898056 (lr=9.1353e-06) (hash(x)=54483057)
827
+ 6800 val loss 6.8557
828
+ 6800 val perplexity 949.2742
829
+ 6800 train 6.893238 (lr=6.3345e-06) (hash(x)=46927608)
830
+ 6800 val loss 6.7582
831
+ 6800 val perplexity 861.1172
832
+ 6800 train 6.799270 (lr=2.2171e-05) (hash(x)=46927608)
833
+ 6900 val loss 6.2722
834
+ 6900 val perplexity 529.6454
835
+ 6900 train 6.280756 (lr=1.5225e-05) (hash(x)=54483057)
836
+ 7000 val loss 6.8406
837
+ 7000 val perplexity 935.0188
838
+ 7000 train 7.017616 (lr=8.7764e-06) (hash(x)=51493884)
839
+ 6900 val loss 6.8503
840
+ 6900 val perplexity 944.1887
841
+ 6900 train 6.903790 (lr=6.0902e-06) (hash(x)=54483057)
842
+ 6900 val loss 6.7468
843
+ 6900 val perplexity 851.3662
844
+ 6900 train 6.786250 (lr=2.1316e-05) (hash(x)=54483057)
845
+ 7000 val loss 6.2622
846
+ 7000 val perplexity 524.3632
847
+ 7000 train 6.384635 (lr=1.4627e-05) (hash(x)=51493884)
848
+ 7100 val loss 6.8341
849
+ 7100 val perplexity 928.9899
850
+ 7100 train 6.986980 (lr=8.4255e-06) (hash(x)=53303341)
851
+ 7000 val loss 6.8478
852
+ 7000 val perplexity 941.8038
853
+ 7000 train 7.024706 (lr=5.8510e-06) (hash(x)=51493884)
854
+ 7000 val loss 6.7401
855
+ 7000 val perplexity 845.6735
856
+ 7000 train 6.893335 (lr=2.0478e-05) (hash(x)=51493884)
857
+ 7100 val loss 6.2533
858
+ 7100 val perplexity 519.7358
859
+ 7100 train 6.410233 (lr=1.4043e-05) (hash(x)=53303341)
860
+ 7200 val loss 6.8317
861
+ 7200 val perplexity 926.7962
862
+ 7200 train 6.506523 (lr=8.0829e-06) (hash(x)=45272178)
863
+ 7100 val loss 6.8382
864
+ 7100 val perplexity 932.8452
865
+ 7100 train 6.992830 (lr=5.6170e-06) (hash(x)=53303341)
866
+ 7100 val loss 6.7416
867
+ 7100 val perplexity 846.8945
868
+ 7100 train 6.887737 (lr=1.9660e-05) (hash(x)=53303341)
869
+ 7200 val loss 6.2521
870
+ 7200 val perplexity 519.1124
871
+ 7200 train 5.923408 (lr=1.3471e-05) (hash(x)=45272178)
872
+ 7300 val loss 6.8276
873
+ 7300 val perplexity 922.9598
874
+ 7300 train 6.765221 (lr=7.7489e-06) (hash(x)=50389872)
875
+ 7200 val loss 6.8349
876
+ 7200 val perplexity 929.7535
877
+ 7200 train 6.511398 (lr=5.3886e-06) (hash(x)=45272178)
878
+ 7200 val loss 6.7386
879
+ 7200 val perplexity 844.3676
880
+ 7200 train 6.421773 (lr=1.8860e-05) (hash(x)=45272178)
881
+ 7300 val loss 6.2463
882
+ 7300 val perplexity 516.0956
883
+ 7300 train 6.172292 (lr=1.2915e-05) (hash(x)=50389872)
884
+ 7400 val loss 6.8232
885
+ 7400 val perplexity 918.9180
886
+ 7400 train 6.427763 (lr=7.4239e-06) (hash(x)=43796301)
887
+ 7300 val loss 6.8288
888
+ 7300 val perplexity 924.1048
889
+ 7300 train 6.762436 (lr=5.1659e-06) (hash(x)=50389872)
890
+ 7300 val loss 6.7364
891
+ 7300 val perplexity 842.5316
892
+ 7300 train 6.675917 (lr=1.8081e-05) (hash(x)=50389872)
893
+ 7400 val loss 6.2405
894
+ 7400 val perplexity 513.1138
895
+ 7400 train 5.857545 (lr=1.2373e-05) (hash(x)=43796301)
896
+ 7500 val loss 6.8176
897
+ 7500 val perplexity 913.7742
898
+ 7500 train 6.729806 (lr=7.1083e-06) (hash(x)=47808686)
899
+ 7400 val loss 6.8260
900
+ 7400 val perplexity 921.4642
901
+ 7400 train 6.432824 (lr=4.9493e-06) (hash(x)=43796301)
902
+ 7400 val loss 6.7280
903
+ 7400 val perplexity 835.4955
904
+ 7400 train 6.339929 (lr=1.7323e-05) (hash(x)=43796301)
905
+ 7500 val loss 6.2347
906
+ 7500 val perplexity 510.1292
907
+ 7500 train 6.186483 (lr=1.1847e-05) (hash(x)=47808686)
908
+ 7600 val loss 6.8152
909
+ 7600 val perplexity 911.5635
910
+ 7600 train 6.492714 (lr=6.8023e-06) (hash(x)=41936898)
911
+ 7500 val loss 6.8228
912
+ 7500 val perplexity 918.5723
913
+ 7500 train 6.743688 (lr=4.7389e-06) (hash(x)=47808686)
914
+ 7500 val loss 6.7312
915
+ 7500 val perplexity 838.1662
916
+ 7500 train 6.651219 (lr=1.6586e-05) (hash(x)=47808686)
917
+ 7600 val loss 6.2323
918
+ 7600 val perplexity 508.9238
919
+ 7600 train 5.926254 (lr=1.1337e-05) (hash(x)=41936898)
920
+ 7700 val loss 6.8156
921
+ 7700 val perplexity 911.9543
922
+ 7700 train 7.035799 (lr=6.5062e-06) (hash(x)=57550318)
923
+ 7600 val loss 6.8198
924
+ 7600 val perplexity 915.7760
925
+ 7600 train 6.489659 (lr=4.5349e-06) (hash(x)=41936898)
926
+ 7600 val loss 6.7244
927
+ 7600 val perplexity 832.4652
928
+ 7600 train 6.405721 (lr=1.5872e-05) (hash(x)=41936898)
929
+ 7700 val loss 6.2320
930
+ 7700 val perplexity 508.7766
931
+ 7700 train 6.498176 (lr=1.0844e-05) (hash(x)=57550318)
932
+ 7800 val loss 6.8077
933
+ 7800 val perplexity 904.7552
934
+ 7800 train 6.667345 (lr=6.2205e-06) (hash(x)=47485210)
935
+ 7700 val loss 6.8177
936
+ 7700 val perplexity 913.8815
937
+ 7700 train 7.047784 (lr=4.3375e-06) (hash(x)=57550318)
938
+ 7700 val loss 6.7234
939
+ 7700 val perplexity 831.6507
940
+ 7700 train 6.950782 (lr=1.5181e-05) (hash(x)=57550318)
941
+ 7800 val loss 6.2312
942
+ 7800 val perplexity 508.3616
943
+ 7800 train 6.076318 (lr=1.0367e-05) (hash(x)=47485210)
944
+ 7900 val loss 6.8062
945
+ 7900 val perplexity 903.4692
946
+ 7900 train 7.151995 (lr=5.9453e-06) (hash(x)=53228688)
947
+ 7800 val loss 6.8134
948
+ 7800 val perplexity 909.9406
949
+ 7800 train 6.677879 (lr=4.1470e-06) (hash(x)=47485210)
950
+ 7800 val loss 6.7184
951
+ 7800 val perplexity 827.5126
952
+ 7800 train 6.569107 (lr=1.4514e-05) (hash(x)=47485210)
953
+ 7900 val loss 6.2215
954
+ 7900 val perplexity 503.4525
955
+ 7900 train 6.591218 (lr=9.9088e-06) (hash(x)=53228688)
956
+ 8000 val loss 6.8044
957
+ 8000 val perplexity 901.7992
958
+ 8000 train 6.852820 (lr=5.6809e-06) (hash(x)=52018673)
959
+ 7900 val loss 6.8108
960
+ 7900 val perplexity 907.6062
961
+ 7900 train 7.159123 (lr=3.9635e-06) (hash(x)=53228688)
962
+ 7900 val loss 6.7168
963
+ 7900 val perplexity 826.1689
964
+ 7900 train 7.056360 (lr=1.3872e-05) (hash(x)=53228688)
965
+ 8000 val loss 6.2179
966
+ 8000 val perplexity 501.6553
967
+ 8000 train 6.293939 (lr=9.4682e-06) (hash(x)=52018673)
968
+ 8100 val loss 6.7997
969
+ 8100 val perplexity 897.6198
970
+ 8100 train 6.593906 (lr=5.4277e-06) (hash(x)=47079349)
971
+ 8000 val loss 6.8103
972
+ 8000 val perplexity 907.1532
973
+ 8000 train 6.872995 (lr=3.7873e-06) (hash(x)=52018673)
974
+ 8000 val loss 6.7181
975
+ 8000 val perplexity 827.2534
976
+ 8000 train 6.770758 (lr=1.3255e-05) (hash(x)=52018673)
977
+ 8100 val loss 6.2129
978
+ 8100 val perplexity 499.1537
979
+ 8100 train 6.023364 (lr=9.0461e-06) (hash(x)=47079349)
980
+ 8200 val loss 6.7971
981
+ 8200 val perplexity 895.2606
982
+ 8200 train 6.860888 (lr=5.1858e-06) (hash(x)=57921563)
983
+ 8100 val loss 6.8062
984
+ 8100 val perplexity 903.4011
985
+ 8100 train 6.600097 (lr=3.6184e-06) (hash(x)=47079349)
986
+ 8100 val loss 6.7129
987
+ 8100 val perplexity 822.9833
988
+ 8100 train 6.511360 (lr=1.2665e-05) (hash(x)=47079349)
989
+ 8200 val loss 6.2123
990
+ 8200 val perplexity 498.8536
991
+ 8200 train 6.274234 (lr=8.6430e-06) (hash(x)=57921563)
992
+ 8300 val loss 6.7956
993
+ 8300 val perplexity 893.9357
994
+ 8300 train 6.333303 (lr=4.9556e-06) (hash(x)=45038933)
995
+ 8200 val loss 6.8044
996
+ 8200 val perplexity 901.7893
997
+ 8200 train 6.886234 (lr=3.4572e-06) (hash(x)=57921563)
998
+ 8200 val loss 6.7100
999
+ 8200 val perplexity 820.5785
1000
+ 8200 train 6.749637 (lr=1.2100e-05) (hash(x)=57921563)
1001
+ 8300 val loss 6.2076
1002
+ 8300 val perplexity 496.5031
1003
+ 8300 train 5.727176 (lr=8.2593e-06) (hash(x)=45038933)
1004
+ 8400 val loss 6.8014
1005
+ 8400 val perplexity 899.1366
1006
+ 8400 train 6.681006 (lr=4.7372e-06) (hash(x)=47763246)
1007
+ 8300 val loss 6.8026
1008
+ 8300 val perplexity 900.1713
1009
+ 8300 train 6.344654 (lr=3.3037e-06) (hash(x)=45038933)
1010
+ 8400 val loss 6.2179
1011
+ 8400 val perplexity 501.6299
1012
+ 8400 train 6.083429 (lr=7.8953e-06) (hash(x)=47763246)
1013
+ 8300 val loss 6.7081
1014
+ 8300 val perplexity 818.9875
1015
+ 8300 train 6.249794 (lr=1.1563e-05) (hash(x)=45038933)
1016
+ 8500 val loss 6.7949
1017
+ 8500 val perplexity 893.3076
1018
+ 8500 train 6.825691 (lr=4.5309e-06) (hash(x)=56176595)
1019
+ 8400 val loss 6.8055
1020
+ 8400 val perplexity 902.7965
1021
+ 8400 train 6.680042 (lr=3.1581e-06) (hash(x)=47763246)
1022
+ 8500 val loss 6.2090
1023
+ 8500 val perplexity 497.1946
1024
+ 8500 train 6.305939 (lr=7.5515e-06) (hash(x)=56176595)
1025
+ 8400 val loss 6.7128
1026
+ 8400 val perplexity 822.8507
1027
+ 8400 train 6.586521 (lr=1.1053e-05) (hash(x)=47763246)
1028
+ 8600 val loss 6.7920
1029
+ 8600 val perplexity 890.7125
1030
+ 8600 train 6.854806 (lr=4.3369e-06) (hash(x)=55184249)
1031
+ 8500 val loss 6.8023
1032
+ 8500 val perplexity 899.9142
1033
+ 8500 train 6.834250 (lr=3.0206e-06) (hash(x)=56176595)
1034
+ 8600 val loss 6.2066
1035
+ 8600 val perplexity 496.0045
1036
+ 8600 train 6.306992 (lr=7.2282e-06) (hash(x)=55184249)
1037
+ 8500 val loss 6.7070
1038
+ 8500 val perplexity 818.1206
1039
+ 8500 train 6.731531 (lr=1.0572e-05) (hash(x)=56176595)
1040
+ 8700 val loss 6.7905
1041
+ 8700 val perplexity 889.3621
1042
+ 8700 train 6.757795 (lr=4.1554e-06) (hash(x)=46471646)
1043
+ 8600 val loss 6.8010
1044
+ 8600 val perplexity 898.7457
1045
+ 8600 train 6.868478 (lr=2.8913e-06) (hash(x)=55184249)
1046
+ 8700 val loss 6.2044
1047
+ 8700 val perplexity 494.9356
1048
+ 8700 train 6.142774 (lr=6.9257e-06) (hash(x)=46471646)
1049
+ 8600 val loss 6.7051
1050
+ 8600 val perplexity 816.5574
1051
+ 8600 train 6.766881 (lr=1.0119e-05) (hash(x)=55184249)
1052
+ 8800 val loss 6.7900
1053
+ 8800 val perplexity 888.9067
1054
+ 8800 train 6.528963 (lr=3.9866e-06) (hash(x)=46233162)
1055
+ 8700 val loss 6.8006
1056
+ 8700 val perplexity 898.3956
1057
+ 8700 train 6.761759 (lr=2.7703e-06) (hash(x)=46471646)
1058
+ 8800 val loss 6.2019
1059
+ 8800 val perplexity 493.6650
1060
+ 8800 train 5.902849 (lr=6.6444e-06) (hash(x)=46233162)
1061
+ 8700 val loss 6.7049
1062
+ 8700 val perplexity 816.4258
1063
+ 8700 train 6.663841 (lr=9.6960e-06) (hash(x)=46471646)
1064
+ 8900 val loss 6.7892
1065
+ 8900 val perplexity 888.2064
1066
+ 8900 train 6.687468 (lr=3.8307e-06) (hash(x)=47233684)
1067
+ 8800 val loss 6.8005
1068
+ 8800 val perplexity 898.3215
1069
+ 8800 train 6.544139 (lr=2.6577e-06) (hash(x)=46233162)
1070
+ 8900 val loss 6.2001
1071
+ 8900 val perplexity 492.8117
1072
+ 8900 train 6.111488 (lr=6.3845e-06) (hash(x)=47233684)
1073
+ 8800 val loss 6.7042
1074
+ 8800 val perplexity 815.7993
1075
+ 8800 train 6.450424 (lr=9.3021e-06) (hash(x)=46233162)
1076
+ 9000 val loss 6.7840
1077
+ 9000 val perplexity 883.5842
1078
+ 9000 train 6.694007 (lr=3.6877e-06) (hash(x)=48374529)
1079
+ 8900 val loss 6.7976
1080
+ 8900 val perplexity 895.6607
1081
+ 8900 train 6.697267 (lr=2.5538e-06) (hash(x)=47233684)
1082
+ 9000 val loss 6.1962
1083
+ 9000 val perplexity 490.8607
1084
+ 9000 train 6.080988 (lr=6.1462e-06) (hash(x)=48374529)
1085
+ 9100 val loss 6.7769
1086
+ 9100 val perplexity 877.3354
1087
+ 9100 train 6.733949 (lr=3.5580e-06) (hash(x)=48065371)
1088
+ 8900 val loss 6.7280
1089
+ 8900 val perplexity 835.4911
1090
+ 8900 train 6.629285 (lr=8.9382e-06) (hash(x)=47233684)
1091
+ 9000 val loss 6.7929
1092
+ 9000 val perplexity 891.4957
1093
+ 9000 train 6.712595 (lr=2.4585e-06) (hash(x)=48374529)
1094
+ 9100 val loss 6.1904
1095
+ 9100 val perplexity 488.0458
1096
+ 9100 train 6.196475 (lr=5.9300e-06) (hash(x)=48065371)
1097
+ 9200 val loss 6.7719
1098
+ 9200 val perplexity 872.9341
1099
+ 9200 train 6.709245 (lr=3.4415e-06) (hash(x)=47408078)
1100
+ 9000 val loss 6.7004
1101
+ 9000 val perplexity 812.7133
1102
+ 9000 train 6.625999 (lr=8.6047e-06) (hash(x)=48374529)
1103
+ 9100 val loss 6.7882
1104
+ 9100 val perplexity 887.2844
1105
+ 9100 train 6.745033 (lr=2.3720e-06) (hash(x)=48065371)
1106
+ 9200 val loss 6.1849
1107
+ 9200 val perplexity 485.3546
1108
+ 9200 train 6.187281 (lr=5.7359e-06) (hash(x)=47408078)
1109
+ 9300 val loss 6.7691
1110
+ 9300 val perplexity 870.5515
1111
+ 9300 train 6.798889 (lr=3.3385e-06) (hash(x)=50749781)
1112
+ 9100 val loss 6.6936
1113
+ 9100 val perplexity 807.1893
1114
+ 9100 train 6.639514 (lr=8.3020e-06) (hash(x)=48065371)
1115
+ 9200 val loss 6.7833
1116
+ 9200 val perplexity 882.9709
1117
+ 9200 train 6.725305 (lr=2.2943e-06) (hash(x)=47408078)
1118
+ 9300 val loss 6.1818
1119
+ 9300 val perplexity 483.8589
1120
+ 9300 train 6.174564 (lr=5.5641e-06) (hash(x)=50749781)
1121
+ 9400 val loss 6.7662
1122
+ 9400 val perplexity 868.0317
1123
+ 9400 train 6.935981 (lr=3.2490e-06) (hash(x)=48560169)
1124
+ 9200 val loss 6.6890
1125
+ 9200 val perplexity 803.4981
1126
+ 9200 train 6.640819 (lr=8.0302e-06) (hash(x)=47408078)
1127
+ 9300 val loss 6.7810
1128
+ 9300 val perplexity 880.9481
1129
+ 9300 train 6.805525 (lr=2.2256e-06) (hash(x)=50749781)
1130
+ 9400 val loss 6.1777
1131
+ 9400 val perplexity 481.8840
1132
+ 9400 train 6.386855 (lr=5.4149e-06) (hash(x)=48560169)
1133
+ 9500 val loss 6.7656
1134
+ 9500 val perplexity 867.4702
1135
+ 9500 train 6.858753 (lr=3.1730e-06) (hash(x)=50936392)
1136
+ 9300 val loss 6.6884
1137
+ 9300 val perplexity 803.0623
1138
+ 9300 train 6.706632 (lr=7.7898e-06) (hash(x)=50749781)
1139
+ 9400 val loss 6.7790
1140
+ 9400 val perplexity 879.1718
1141
+ 9400 train 6.955186 (lr=2.1660e-06) (hash(x)=48560169)
1142
+ 9500 val loss 6.1773
1143
+ 9500 val perplexity 481.6834
1144
+ 9500 train 6.323741 (lr=5.2884e-06) (hash(x)=50936392)
1145
+ 9600 val loss 6.7621
1146
+ 9600 val perplexity 864.4913
1147
+ 9600 train 6.770996 (lr=3.1108e-06) (hash(x)=50651714)
1148
+ 9400 val loss 6.6865
1149
+ 9400 val perplexity 801.5171
1150
+ 9400 train 6.862773 (lr=7.5809e-06) (hash(x)=48560169)
1151
+ 9500 val loss 6.7787
1152
+ 9500 val perplexity 878.9538
1153
+ 9500 train 6.867536 (lr=2.1154e-06) (hash(x)=50936392)
1154
+ 9600 val loss 6.1737
1155
+ 9600 val perplexity 479.9369
1156
+ 9600 train 6.183331 (lr=5.1847e-06) (hash(x)=50651714)
1157
+ 9700 val loss 6.7618
1158
+ 9700 val perplexity 864.1599
1159
+ 9700 train 6.810016 (lr=3.0624e-06) (hash(x)=47311384)
1160
+ 9500 val loss 6.6858
1161
+ 9500 val perplexity 800.9467
1162
+ 9500 train 6.774101 (lr=7.4038e-06) (hash(x)=50936392)
1163
+ 9600 val loss 6.7750
1164
+ 9600 val perplexity 875.6749
1165
+ 9600 train 6.782158 (lr=2.0739e-06) (hash(x)=50651714)
1166
+ 9700 val loss 6.1739
1167
+ 9700 val perplexity 480.0685
1168
+ 9700 train 6.242620 (lr=5.1040e-06) (hash(x)=47311384)
1169
+ 9800 val loss 6.7606
1170
+ 9800 val perplexity 863.1872
1171
+ 9800 train 6.851887 (lr=3.0277e-06) (hash(x)=50921139)
1172
+ 9600 val loss 6.6821
1173
+ 9600 val perplexity 798.0090
1174
+ 9600 train 6.700386 (lr=7.2586e-06) (hash(x)=50651714)
1175
+ 9700 val loss 6.7748
1176
+ 9700 val perplexity 875.5066
1177
+ 9700 train 6.824248 (lr=2.0416e-06) (hash(x)=47311384)
1178
+ 9700 val loss 6.6819
1179
+ 9700 val perplexity 797.8264
1180
+ 9700 train 6.707674 (lr=7.1456e-06) (hash(x)=47311384)
1181
+ 9800 val loss 6.1714
1182
+ 9800 val perplexity 478.8529
1183
+ 9900 val loss 6.7642
1184
+ 9900 val perplexity 866.2962
1185
+ 9800 train 6.277335 (lr=5.0462e-06) (hash(x)=50921139)
1186
+ 9900 train 6.832372 (lr=3.0069e-06) (hash(x)=48142455)
1187
+ 9800 val loss 6.7740
1188
+ 9800 val perplexity 874.7659
1189
+ 9800 train 6.852812 (lr=2.0185e-06) (hash(x)=50921139)
1190
+ 9800 val loss 6.6834
1191
+ 9800 val perplexity 799.0583
1192
+ 9800 train 6.780736 (lr=7.0647e-06) (hash(x)=50921139)
1193
+ 9900 val loss 6.7763
1194
+ 9900 val perplexity 876.8536
1195
+ 9900 train 6.845608 (lr=2.0046e-06) (hash(x)=48142455)
1196
+ 9900 val loss 6.6816
1197
+ 9900 val perplexity 797.6224
1198
+ 9900 train 6.752887 (lr=7.0162e-06) (hash(x)=48142455)
1199
+ 9999 val loss 6.7563
1200
+ 9999 val perplexity 859.4526
1201
+ 9900 val loss 6.1866
1202
+ 9900 val perplexity 486.1883
1203
+ 9900 train 6.339453 (lr=5.0116e-06) (hash(x)=48142455)
1204
+ 9999 val loss 6.6808
1205
+ 9999 val perplexity 796.9465
1206
+ 9999 val loss 6.7708
1207
+ 9999 val perplexity 872.0459
1208
+ 9999 val loss 6.1683
1209
+ 9999 val perplexity 477.3951
attention_kindselective_n_heads8_seed1339/model_02500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8cc97daafc5f6ded218c83e1b779eb3d5e97ba3598ab35305df344c8869610e8
3
+ size 257976706
attention_kindselective_n_heads8_seed1339/model_05000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2cb8ebeeccc09ab5d42dc52154f3bc963855ad035936bf90089675a51ce9c382
3
+ size 257976706
attention_kindselective_n_heads8_seed1339/model_07500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80a4b48de41d0bff6b808707d66b00458018754a102bccc0170d3dabd0a8017a
3
+ size 257976706
attention_kindselective_n_heads8_seed1339/model_09999.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1382134aa91d192f7966000d70b476917de03bd95cdf60fc4eaeb69c0ae25798
3
+ size 257976706
attention_kindselective_n_heads8_seed1339/optimizer_02500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eecd2916e9fcb62023114ef3d168d16a02311d6db718b9ed79b44949ec4fdbf0
3
+ size 509672838
attention_kindselective_n_heads8_seed1339/optimizer_05000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3688091768d19206e9078a5e8894f0b8c17e39cea9c6e9bbf6c10abb705e06a0
3
+ size 509672838
attention_kindselective_n_heads8_seed1339/optimizer_07500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a603a4dc1aa4548b8438e55cbd52bf233b711699976c0fba008d5aaee5ba4c8e
3
+ size 509672838
attention_kindselective_n_heads8_seed1339/optimizer_09999.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:206e13efb97150555c1ca2b999f82739aba7206cd7894c146f5a95fbec5bed4f
3
+ size 509672838