andrew-healey commited on
Commit
f68044c
·
verified ·
1 Parent(s): e38c4b5

Upload folder using huggingface_hub

Browse files
attention_kindselective_n_heads8_seed1338/args.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_6/attention_kindselective_n_heads8_seed1338", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 8, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 10000, "warmup_steps": 200, "group": "wider_is_better_6", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1338, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 40, "total_batch_size": 10240, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 2e-05, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "2e-5_10240_8_1338", "n_embd": 512}
attention_kindselective_n_heads8_seed1338/dataloader_02500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b2ea67f78ff5a7970d0db044ff7ee527b3dc065f295fd30f588df4b44b568d0
3
+ size 964
attention_kindselective_n_heads8_seed1338/dataloader_05000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f03ed2ebf741f15e13c79e6cc1e9a19b308450d81cc3b4d8d0338c63d77ca59
3
+ size 964
attention_kindselective_n_heads8_seed1338/dataloader_07500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:82590037fb2eecbec961f7967a8dd1b8d85515d31a252f66b92b8139858a8b7c
3
+ size 964
attention_kindselective_n_heads8_seed1338/dataloader_09999.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c544303717d12355a69b8ffb1eb109434e4fdccfd5a61279b6e8ba2e870d6700
3
+ size 964
attention_kindselective_n_heads8_seed1338/log2.txt ADDED
@@ -0,0 +1,1209 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ max_steps: 10000
2
+ 0 val loss 11.0412
3
+ 0 val perplexity 62392.4570
4
+ 0 val loss 11.0412
5
+ 0 val perplexity 62392.5195
6
+ 0 val loss 11.0412
7
+ 0 val perplexity 62392.4570
8
+ 0 val loss 11.0412
9
+ 0 val perplexity 62392.4570
10
+ 0 train 11.030708 (lr=2.5000e-07) (hash(x)=50671684)
11
+ 0 train 11.030660 (lr=3.5000e-07) (hash(x)=50671684)
12
+ 0 train 11.030660 (lr=1.0000e-07) (hash(x)=50671684)
13
+ 0 train 11.030662 (lr=1.5000e-07) (hash(x)=50671684)
14
+ 100 val loss 9.5150
15
+ 100 val perplexity 13562.2197
16
+ 100 train 9.475113 (lr=3.5350e-05) (hash(x)=52740221)
17
+ 100 val loss 9.8877
18
+ 100 val perplexity 19685.8281
19
+ 100 val loss 9.6106
20
+ 100 val perplexity 14921.5723
21
+ 100 train 9.853711 (lr=1.0100e-05) (hash(x)=52740221)
22
+ 100 val loss 9.7397
23
+ 100 val perplexity 16979.0996
24
+ 100 train 9.567675 (lr=2.5250e-05) (hash(x)=52740221)
25
+ 100 train 9.698705 (lr=1.5150e-05) (hash(x)=52740221)
26
+ 200 val loss 8.0793
27
+ 200 val perplexity 3227.0042
28
+ 200 train 8.077329 (lr=7.0000e-05) (hash(x)=49034180)
29
+ 200 val loss 9.1855
30
+ 200 val perplexity 9755.0391
31
+ 200 val loss 9.4434
32
+ 200 val perplexity 12623.9707
33
+ 200 train 9.460796 (lr=2.0000e-05) (hash(x)=49034180)
34
+ 200 train 9.197133 (lr=3.0000e-05) (hash(x)=49034180)
35
+ 200 val loss 8.5470
36
+ 200 val perplexity 5151.4980
37
+ 200 train 8.548553 (lr=5.0000e-05) (hash(x)=49034180)
38
+ 300 val loss 7.5602
39
+ 300 val perplexity 1920.1787
40
+ 300 train 7.826581 (lr=6.9984e-05) (hash(x)=63180688)
41
+ 300 val loss 8.8301
42
+ 300 val perplexity 6837.0103
43
+ 300 train 8.877916 (lr=1.9995e-05) (hash(x)=63180688)
44
+ 300 val loss 8.2489
45
+ 300 val perplexity 3823.5215
46
+ 300 train 8.357185 (lr=2.9993e-05) (hash(x)=63180688)
47
+ 300 val loss 7.8357
48
+ 300 val perplexity 2529.2393
49
+ 300 train 8.017459 (lr=4.9988e-05) (hash(x)=63180688)
50
+ 400 val loss 7.4376
51
+ 400 val perplexity 1698.6493
52
+ 400 train 7.438178 (lr=6.9935e-05) (hash(x)=50373500)
53
+ 400 val loss 8.2621
54
+ 400 val perplexity 3874.1113
55
+ 400 train 8.178881 (lr=1.9982e-05) (hash(x)=50373500)
56
+ 400 val loss 7.8110
57
+ 400 val perplexity 2467.4846
58
+ 400 train 7.777402 (lr=2.9972e-05) (hash(x)=50373500)
59
+ 400 val loss 7.6415
60
+ 400 val perplexity 2082.9680
61
+ 400 train 7.627953 (lr=4.9954e-05) (hash(x)=50373500)
62
+ 500 val loss 7.3646
63
+ 500 val perplexity 1579.0923
64
+ 500 train 7.238733 (lr=6.9854e-05) (hash(x)=44547422)
65
+ 500 val loss 7.9960
66
+ 500 val perplexity 2969.1821
67
+ 500 train 7.944740 (lr=1.9958e-05) (hash(x)=44547422)
68
+ 500 val loss 7.6391
69
+ 500 val perplexity 2077.9395
70
+ 500 train 7.550472 (lr=2.9938e-05) (hash(x)=44547422)
71
+ 500 val loss 7.5464
72
+ 500 val perplexity 1893.9757
73
+ 500 train 7.441219 (lr=4.9896e-05) (hash(x)=44547422)
74
+ 600 val loss 7.3356
75
+ 600 val perplexity 1533.9551
76
+ 600 train 7.176397 (lr=6.9741e-05) (hash(x)=47184699)
77
+ 600 val loss 7.8623
78
+ 600 val perplexity 2597.5964
79
+ 600 train 7.796846 (lr=1.9926e-05) (hash(x)=47184699)
80
+ 600 val loss 7.5444
81
+ 600 val perplexity 1890.1512
82
+ 600 train 7.428701 (lr=2.9889e-05) (hash(x)=47184699)
83
+ 600 val loss 7.4974
84
+ 600 val perplexity 1803.3756
85
+ 600 train 7.370130 (lr=4.9815e-05) (hash(x)=47184699)
86
+ 700 val loss 7.3017
87
+ 700 val perplexity 1482.8025
88
+ 700 train 7.180538 (lr=6.9596e-05) (hash(x)=51374582)
89
+ 700 val loss 7.7705
90
+ 700 val perplexity 2369.7173
91
+ 700 train 7.703903 (lr=1.9885e-05) (hash(x)=51374582)
92
+ 700 val loss 7.4869
93
+ 700 val perplexity 1784.5436
94
+ 700 train 7.378650 (lr=2.9827e-05) (hash(x)=51374582)
95
+ 700 val loss 7.4491
96
+ 700 val perplexity 1718.3898
97
+ 700 train 7.325911 (lr=4.9712e-05) (hash(x)=51374582)
98
+ 800 val loss 7.2514
99
+ 800 val perplexity 1410.0104
100
+ 800 train 6.968596 (lr=6.9419e-05) (hash(x)=46264805)
101
+ 800 val loss 7.6907
102
+ 800 val perplexity 2187.8003
103
+ 800 train 7.486768 (lr=1.9834e-05) (hash(x)=46264805)
104
+ 800 val loss 7.4399
105
+ 800 val perplexity 1702.6342
106
+ 800 train 7.190279 (lr=2.9751e-05) (hash(x)=46264805)
107
+ 800 val loss 7.4127
108
+ 800 val perplexity 1656.8792
109
+ 800 train 7.154434 (lr=4.9585e-05) (hash(x)=46264805)
110
+ 900 val loss 7.2094
111
+ 900 val perplexity 1352.0565
112
+ 900 train 7.378638 (lr=6.9210e-05) (hash(x)=61178712)
113
+ 900 val loss 7.6276
114
+ 900 val perplexity 2054.2126
115
+ 900 train 7.740030 (lr=1.9774e-05) (hash(x)=61178712)
116
+ 900 val loss 7.4194
117
+ 900 val perplexity 1668.0503
118
+ 900 train 7.569212 (lr=2.9662e-05) (hash(x)=61178712)
119
+ 900 val loss 7.3820
120
+ 900 val perplexity 1606.8147
121
+ 900 train 7.542928 (lr=4.9436e-05) (hash(x)=61178712)
122
+ 1000 val loss 7.1676
123
+ 1000 val perplexity 1296.7246
124
+ 1000 train 7.039244 (lr=6.8970e-05) (hash(x)=50886520)
125
+ 1000 val loss 7.5767
126
+ 1000 val perplexity 1952.1151
127
+ 1000 train 7.549294 (lr=1.9706e-05) (hash(x)=50886520)
128
+ 1000 val loss 7.3707
129
+ 1000 val perplexity 1588.7726
130
+ 1000 train 7.311100 (lr=2.9558e-05) (hash(x)=50886520)
131
+ 1000 val loss 7.3440
132
+ 1000 val perplexity 1546.8590
133
+ 1000 train 7.258212 (lr=4.9264e-05) (hash(x)=50886520)
134
+ 1100 val loss 7.1469
135
+ 1100 val perplexity 1270.1753
136
+ 1100 train 6.893501 (lr=6.8698e-05) (hash(x)=48600099)
137
+ 1100 val loss 7.5429
138
+ 1100 val perplexity 1887.3206
139
+ 1100 train 7.344742 (lr=1.9628e-05) (hash(x)=48600099)
140
+ 1100 val loss 7.3492
141
+ 1100 val perplexity 1554.8961
142
+ 1100 train 7.104578 (lr=2.9442e-05) (hash(x)=48600099)
143
+ 1100 val loss 7.3139
144
+ 1100 val perplexity 1501.0520
145
+ 1100 train 7.054069 (lr=4.9070e-05) (hash(x)=48600099)
146
+ 1200 val loss 7.1126
147
+ 1200 val perplexity 1227.3606
148
+ 1200 train 6.809518 (lr=6.8395e-05) (hash(x)=50146792)
149
+ 1200 val loss 7.5094
150
+ 1200 val perplexity 1825.0875
151
+ 1200 train 7.275321 (lr=1.9542e-05) (hash(x)=50146792)
152
+ 1200 val loss 7.3220
153
+ 1200 val perplexity 1513.1711
154
+ 1200 train 7.054025 (lr=2.9312e-05) (hash(x)=50146792)
155
+ 1200 val loss 7.2835
156
+ 1200 val perplexity 1456.0861
157
+ 1200 train 7.006647 (lr=4.8854e-05) (hash(x)=50146792)
158
+ 1300 val loss 7.1075
159
+ 1300 val perplexity 1221.0498
160
+ 1300 train 6.973897 (lr=6.8062e-05) (hash(x)=52617313)
161
+ 1300 val loss 7.4832
162
+ 1300 val perplexity 1777.9160
163
+ 1300 train 7.413532 (lr=1.9446e-05) (hash(x)=52617313)
164
+ 1300 val loss 7.2895
165
+ 1300 val perplexity 1464.8091
166
+ 1300 train 7.192160 (lr=2.9169e-05) (hash(x)=52617313)
167
+ 1300 val loss 7.2547
168
+ 1300 val perplexity 1414.8009
169
+ 1300 train 7.144293 (lr=4.8616e-05) (hash(x)=52617313)
170
+ 1400 val loss 7.0632
171
+ 1400 val perplexity 1168.1980
172
+ 1400 train 6.699810 (lr=6.7698e-05) (hash(x)=49794446)
173
+ 1400 val loss 7.4626
174
+ 1400 val perplexity 1741.7556
175
+ 1400 train 7.178441 (lr=1.9342e-05) (hash(x)=49794446)
176
+ 1400 val loss 7.2588
177
+ 1400 val perplexity 1420.5990
178
+ 1400 train 6.923851 (lr=2.9013e-05) (hash(x)=49794446)
179
+ 1400 val loss 7.2191
180
+ 1400 val perplexity 1365.2207
181
+ 1400 train 6.870090 (lr=4.8356e-05) (hash(x)=49794446)
182
+ 1500 val loss 7.0495
183
+ 1500 val perplexity 1152.2533
184
+ 1500 train 6.669672 (lr=6.7304e-05) (hash(x)=50766317)
185
+ 1500 val loss 7.4453
186
+ 1500 val perplexity 1711.8087
187
+ 1500 train 7.146240 (lr=1.9230e-05) (hash(x)=50766317)
188
+ 1500 val loss 7.2383
189
+ 1500 val perplexity 1391.7863
190
+ 1500 val loss 7.1959
191
+ 1500 val perplexity 1333.9630
192
+ 1500 train 6.877825 (lr=2.8845e-05) (hash(x)=50766317)
193
+ 1500 train 6.816164 (lr=4.8074e-05) (hash(x)=50766317)
194
+ 1600 val loss 7.0410
195
+ 1600 val perplexity 1142.4799
196
+ 1600 train 6.882990 (lr=6.6881e-05) (hash(x)=55551175)
197
+ 1600 val loss 7.4334
198
+ 1600 val perplexity 1691.6099
199
+ 1600 train 7.326027 (lr=1.9109e-05) (hash(x)=55551175)
200
+ 1600 val loss 7.2227
201
+ 1600 val perplexity 1370.1851
202
+ 1600 train 7.092454 (lr=2.8663e-05) (hash(x)=55551175)
203
+ 1600 val loss 7.1825
204
+ 1600 val perplexity 1316.2428
205
+ 1600 train 7.034845 (lr=4.7772e-05) (hash(x)=55551175)
206
+ 1700 val loss 6.9983
207
+ 1700 val perplexity 1094.7711
208
+ 1700 train 7.202261 (lr=6.6428e-05) (hash(x)=56717172)
209
+ 1700 val loss 7.4104
210
+ 1700 val perplexity 1653.1425
211
+ 1700 train 7.585999 (lr=1.8979e-05) (hash(x)=56717172)
212
+ 1700 val loss 7.1886
213
+ 1700 val perplexity 1324.2910
214
+ 1700 train 7.378125 (lr=2.8469e-05) (hash(x)=56717172)
215
+ 1700 val loss 7.1430
216
+ 1700 val perplexity 1265.2288
217
+ 1700 train 7.331745 (lr=4.7448e-05) (hash(x)=56717172)
218
+ 1800 val loss 6.9486
219
+ 1800 val perplexity 1041.6569
220
+ 1800 train 7.104018 (lr=6.5947e-05) (hash(x)=55376447)
221
+ 1800 val loss 7.3684
222
+ 1800 val perplexity 1585.1343
223
+ 1800 train 7.653419 (lr=1.8842e-05) (hash(x)=55376447)
224
+ 1800 val loss 7.1570
225
+ 1800 val perplexity 1283.0609
226
+ 1800 train 7.384525 (lr=2.8263e-05) (hash(x)=55376447)
227
+ 1800 val loss 7.1026
228
+ 1800 val perplexity 1215.0824
229
+ 1800 train 7.298141 (lr=4.7105e-05) (hash(x)=55376447)
230
+ 1900 val loss 6.9468
231
+ 1900 val perplexity 1039.7795
232
+ 1900 train 6.644609 (lr=6.5437e-05) (hash(x)=43810837)
233
+ 1900 val loss 7.3462
234
+ 1900 val perplexity 1550.2446
235
+ 1900 train 7.057503 (lr=1.8696e-05) (hash(x)=43810837)
236
+ 1900 val loss 7.1444
237
+ 1900 val perplexity 1266.9542
238
+ 1900 train 6.842966 (lr=2.8044e-05) (hash(x)=43810837)
239
+ 1900 val loss 7.1072
240
+ 1900 val perplexity 1220.6906
241
+ 1900 train 6.805607 (lr=4.6741e-05) (hash(x)=43810837)
242
+ 2000 val loss 6.9125
243
+ 2000 val perplexity 1004.7640
244
+ 2000 train 6.955251 (lr=6.4900e-05) (hash(x)=50881655)
245
+ 2000 val loss 7.3188
246
+ 2000 val perplexity 1508.4669
247
+ 2000 train 7.377332 (lr=1.8543e-05) (hash(x)=50881655)
248
+ 2000 val loss 7.1095
249
+ 2000 val perplexity 1223.4854
250
+ 2000 train 7.146984 (lr=2.7814e-05) (hash(x)=50881655)
251
+ 2000 val loss 7.0598
252
+ 2000 val perplexity 1164.1548
253
+ 2000 train 7.097054 (lr=4.6357e-05) (hash(x)=50881655)
254
+ 2100 val loss 6.8942
255
+ 2100 val perplexity 986.5238
256
+ 2100 train 6.770719 (lr=6.4335e-05) (hash(x)=49386015)
257
+ 2100 val loss 7.2984
258
+ 2100 val perplexity 1478.0065
259
+ 2100 train 7.180274 (lr=1.8382e-05) (hash(x)=49386015)
260
+ 2100 val loss 7.0916
261
+ 2100 val perplexity 1201.8446
262
+ 2100 train 6.963319 (lr=2.7572e-05) (hash(x)=49386015)
263
+ 2100 val loss 7.0394
264
+ 2100 val perplexity 1140.6602
265
+ 2100 train 6.919005 (lr=4.5954e-05) (hash(x)=49386015)
266
+ 2200 val loss 6.8933
267
+ 2200 val perplexity 985.6652
268
+ 2200 train 6.780797 (lr=6.3745e-05) (hash(x)=48572079)
269
+ 2200 val loss 7.2870
270
+ 2200 val perplexity 1461.1439
271
+ 2200 train 7.191564 (lr=1.8213e-05) (hash(x)=48572079)
272
+ 2200 val loss 7.0769
273
+ 2200 val perplexity 1184.3407
274
+ 2200 train 6.976396 (lr=2.7319e-05) (hash(x)=48572079)
275
+ 2200 val loss 7.0254
276
+ 2200 val perplexity 1124.8413
277
+ 2200 train 6.915151 (lr=4.5532e-05) (hash(x)=48572079)
278
+ 2300 val loss 6.8545
279
+ 2300 val perplexity 948.1169
280
+ 2300 train 6.901963 (lr=6.3128e-05) (hash(x)=54950719)
281
+ 2300 val loss 7.2620
282
+ 2300 val perplexity 1425.1230
283
+ 2300 train 7.301581 (lr=1.8036e-05) (hash(x)=54950719)
284
+ 2300 val loss 7.0468
285
+ 2300 val perplexity 1149.1204
286
+ 2300 train 7.070806 (lr=2.7055e-05) (hash(x)=54950719)
287
+ 2300 val loss 6.9928
288
+ 2300 val perplexity 1088.7826
289
+ 2300 train 7.020673 (lr=4.5091e-05) (hash(x)=54950719)
290
+ 2400 val loss 6.8428
291
+ 2400 val perplexity 937.0630
292
+ 2400 train 6.468015 (lr=6.2486e-05) (hash(x)=42190240)
293
+ 2400 val loss 7.2473
294
+ 2400 val perplexity 1404.2675
295
+ 2400 train 6.883503 (lr=1.7853e-05) (hash(x)=42190240)
296
+ 2400 val loss 7.0281
297
+ 2400 val perplexity 1127.8716
298
+ 2400 train 6.654352 (lr=2.6780e-05) (hash(x)=42190240)
299
+ 2400 val loss 6.9754
300
+ 2400 val perplexity 1069.9851
301
+ 2400 train 6.604122 (lr=4.4633e-05) (hash(x)=42190240)
302
+ 2500 val loss 6.8213
303
+ 2500 val perplexity 917.2199
304
+ 2500 train 6.914586 (lr=6.1819e-05) (hash(x)=45223539)
305
+ 2500 val loss 7.2317
306
+ 2500 val perplexity 1382.6361
307
+ 2500 train 7.320289 (lr=1.7663e-05) (hash(x)=45223539)
308
+ 2500 val loss 7.0148
309
+ 2500 val perplexity 1112.9597
310
+ 2500 val loss 6.9591
311
+ 2500 val perplexity 1052.7030
312
+ 2500 train 7.101954 (lr=2.6494e-05) (hash(x)=45223539)
313
+ 2500 train 7.038404 (lr=4.4156e-05) (hash(x)=45223539)
314
+ 2600 val loss 6.8081
315
+ 2600 val perplexity 905.1906
316
+ 2600 train 6.809896 (lr=6.1128e-05) (hash(x)=54037353)
317
+ 2600 val loss 7.2149
318
+ 2600 val perplexity 1359.4930
319
+ 2600 train 7.229406 (lr=1.7465e-05) (hash(x)=54037353)
320
+ 2600 val loss 7.0019
321
+ 2600 val perplexity 1098.6744
322
+ 2600 train 7.021422 (lr=2.6198e-05) (hash(x)=54037353)
323
+ 2600 val loss 6.9374
324
+ 2600 val perplexity 1030.1311
325
+ 2600 train 6.954820 (lr=4.3663e-05) (hash(x)=54037353)
326
+ 2700 val loss 6.7813
327
+ 2700 val perplexity 881.1724
328
+ 2700 train 7.265512 (lr=6.0414e-05) (hash(x)=59131616)
329
+ 2700 val loss 7.1999
330
+ 2700 val perplexity 1339.3392
331
+ 2700 train 7.691195 (lr=1.7261e-05) (hash(x)=59131616)
332
+ 2700 val loss 6.9835
333
+ 2700 val perplexity 1078.6578
334
+ 2700 train 7.498905 (lr=2.5892e-05) (hash(x)=59131616)
335
+ 2700 val loss 6.9108
336
+ 2700 val perplexity 1003.0886
337
+ 2700 train 7.397624 (lr=4.3153e-05) (hash(x)=59131616)
338
+ 2800 val loss 6.7650
339
+ 2800 val perplexity 866.9603
340
+ 2800 train 6.601818 (lr=5.9677e-05) (hash(x)=45882743)
341
+ 2800 val loss 7.1851
342
+ 2800 val perplexity 1319.6088
343
+ 2800 train 7.032475 (lr=1.7051e-05) (hash(x)=45882743)
344
+ 2800 val loss 6.9720
345
+ 2800 val perplexity 1066.3402
346
+ 2800 val loss 6.8969
347
+ 2800 val perplexity 989.2418
348
+ 2800 train 6.810520 (lr=2.5576e-05) (hash(x)=45882743)
349
+ 2800 train 6.740892 (lr=4.2627e-05) (hash(x)=45882743)
350
+ 2900 val loss 6.7591
351
+ 2900 val perplexity 861.8439
352
+ 2900 train 6.378328 (lr=5.8919e-05) (hash(x)=43758910)
353
+ 2900 val loss 7.1679
354
+ 2900 val perplexity 1297.1414
355
+ 2900 train 6.796428 (lr=1.6834e-05) (hash(x)=43758910)
356
+ 2900 val loss 6.8926
357
+ 2900 val perplexity 984.9496
358
+ 2900 train 6.507185 (lr=4.2085e-05) (hash(x)=43758910)
359
+ 2900 val loss 6.9604
360
+ 2900 val perplexity 1054.0331
361
+ 2900 train 6.588078 (lr=2.5251e-05) (hash(x)=43758910)
362
+ 3000 val loss 6.7508
363
+ 3000 val perplexity 854.7387
364
+ 3000 train 6.644414 (lr=5.8140e-05) (hash(x)=47965974)
365
+ 3000 val loss 7.1723
366
+ 3000 val perplexity 1302.8431
367
+ 3000 train 7.095241 (lr=1.6611e-05) (hash(x)=47965974)
368
+ 3000 val loss 6.8833
369
+ 3000 val perplexity 975.8807
370
+ 3000 train 6.776263 (lr=4.1529e-05) (hash(x)=47965974)
371
+ 3100 val loss 6.7373
372
+ 3100 val perplexity 843.3165
373
+ 3100 train 6.585240 (lr=5.7340e-05) (hash(x)=48205243)
374
+ 3000 val loss 6.9506
375
+ 3000 val perplexity 1043.7854
376
+ 3000 train 6.852409 (lr=2.4917e-05) (hash(x)=47965974)
377
+ 3100 val loss 7.1504
378
+ 3100 val perplexity 1274.6481
379
+ 3100 train 7.029405 (lr=1.6383e-05) (hash(x)=48205243)
380
+ 3200 val loss 6.7279
381
+ 3200 val perplexity 835.3999
382
+ 3200 train 6.822510 (lr=5.6522e-05) (hash(x)=54511383)
383
+ 3100 val loss 6.8696
384
+ 3100 val perplexity 962.5995
385
+ 3100 train 6.728429 (lr=4.0957e-05) (hash(x)=48205243)
386
+ 3100 val loss 6.9418
387
+ 3100 val perplexity 1034.6399
388
+ 3100 train 6.792998 (lr=2.4574e-05) (hash(x)=48205243)
389
+ 3200 val loss 7.1385
390
+ 3200 val perplexity 1259.5211
391
+ 3200 train 7.222794 (lr=1.6149e-05) (hash(x)=54511383)
392
+ 3300 val loss 6.7237
393
+ 3300 val perplexity 831.9065
394
+ 3300 train 6.682970 (lr=5.5684e-05) (hash(x)=54428388)
395
+ 3200 val loss 6.8627
396
+ 3200 val perplexity 955.9105
397
+ 3200 train 6.959624 (lr=4.0373e-05) (hash(x)=54511383)
398
+ 3200 val loss 6.9223
399
+ 3200 val perplexity 1014.6286
400
+ 3200 train 7.015221 (lr=2.4224e-05) (hash(x)=54511383)
401
+ 3300 val loss 7.1309
402
+ 3300 val perplexity 1249.9888
403
+ 3300 train 7.094438 (lr=1.5910e-05) (hash(x)=54428388)
404
+ 3400 val loss 6.7185
405
+ 3400 val perplexity 827.5817
406
+ 3400 train 6.686805 (lr=5.4829e-05) (hash(x)=48115990)
407
+ 3300 val loss 6.8569
408
+ 3300 val perplexity 950.3951
409
+ 3300 train 6.825317 (lr=3.9775e-05) (hash(x)=54428388)
410
+ 3300 val loss 6.9152
411
+ 3300 val perplexity 1007.5222
412
+ 3300 train 6.874907 (lr=2.3865e-05) (hash(x)=54428388)
413
+ 3400 val loss 7.1204
414
+ 3400 val perplexity 1236.9988
415
+ 3400 train 7.180803 (lr=1.5666e-05) (hash(x)=48115990)
416
+ 3500 val loss 6.7021
417
+ 3500 val perplexity 814.1271
418
+ 3500 train 6.281488 (lr=5.3958e-05) (hash(x)=41137345)
419
+ 3400 val loss 6.8489
420
+ 3400 val perplexity 942.8759
421
+ 3400 train 6.851781 (lr=3.9164e-05) (hash(x)=48115990)
422
+ 3400 val loss 6.9011
423
+ 3400 val perplexity 993.3831
424
+ 3400 train 6.906279 (lr=2.3498e-05) (hash(x)=48115990)
425
+ 3500 val loss 7.1075
426
+ 3500 val perplexity 1221.1249
427
+ 3500 train 6.737636 (lr=1.5416e-05) (hash(x)=41137345)
428
+ 3600 val loss 6.6806
429
+ 3600 val perplexity 796.7698
430
+ 3600 train 6.564756 (lr=5.3070e-05) (hash(x)=55186224)
431
+ 3500 val loss 6.8443
432
+ 3500 val perplexity 938.5548
433
+ 3500 train 6.456430 (lr=3.8541e-05) (hash(x)=41137345)
434
+ 3500 val loss 6.8925
435
+ 3500 val perplexity 984.8782
436
+ 3500 train 6.500405 (lr=2.3125e-05) (hash(x)=41137345)
437
+ 3600 val loss 7.0997
438
+ 3600 val perplexity 1211.5752
439
+ 3600 train 6.996099 (lr=1.5163e-05) (hash(x)=55186224)
440
+ 3700 val loss 6.6738
441
+ 3700 val perplexity 791.3663
442
+ 3700 train 6.579067 (lr=5.2167e-05) (hash(x)=54990049)
443
+ 3600 val loss 6.8360
444
+ 3600 val perplexity 930.7857
445
+ 3600 train 6.722075 (lr=3.7907e-05) (hash(x)=55186224)
446
+ 3600 val loss 6.8866
447
+ 3600 val perplexity 979.0352
448
+ 3600 train 6.769399 (lr=2.2744e-05) (hash(x)=55186224)
449
+ 3700 val loss 7.0944
450
+ 3700 val perplexity 1205.2024
451
+ 3700 train 6.992174 (lr=1.4905e-05) (hash(x)=54990049)
452
+ 3800 val loss 6.6671
453
+ 3800 val perplexity 786.1088
454
+ 3800 train 6.389765 (lr=5.1251e-05) (hash(x)=46288812)
455
+ 3700 val loss 6.8334
456
+ 3700 val perplexity 928.3057
457
+ 3700 train 6.733560 (lr=3.7262e-05) (hash(x)=54990049)
458
+ 3700 val loss 6.8801
459
+ 3700 val perplexity 972.7336
460
+ 3700 train 6.785053 (lr=2.2357e-05) (hash(x)=54990049)
461
+ 3800 val loss 7.0916
462
+ 3800 val perplexity 1201.8423
463
+ 3800 train 6.850882 (lr=1.4643e-05) (hash(x)=46288812)
464
+ 3900 val loss 6.6555
465
+ 3900 val perplexity 777.0841
466
+ 3900 train 6.301198 (lr=5.0321e-05) (hash(x)=45829773)
467
+ 3800 val loss 6.8309
468
+ 3800 val perplexity 925.9794
469
+ 3800 train 6.575776 (lr=3.6608e-05) (hash(x)=46288812)
470
+ 3800 val loss 6.8744
471
+ 3800 val perplexity 967.1729
472
+ 3800 train 6.619145 (lr=2.1965e-05) (hash(x)=46288812)
473
+ 3900 val loss 7.0761
474
+ 3900 val perplexity 1183.3990
475
+ 3900 train 6.760256 (lr=1.4377e-05) (hash(x)=45829773)
476
+ 4000 val loss 6.6570
477
+ 4000 val perplexity 778.1891
478
+ 4000 train 6.411054 (lr=4.9379e-05) (hash(x)=52499943)
479
+ 3900 val loss 6.8091
480
+ 3900 val perplexity 906.0413
481
+ 3900 val loss 6.8577
482
+ 3900 val perplexity 951.1495
483
+ 3900 train 6.470855 (lr=3.5944e-05) (hash(x)=45829773)
484
+ 3900 train 6.522695 (lr=2.1566e-05) (hash(x)=45829773)
485
+ 4000 val loss 7.0684
486
+ 4000 val perplexity 1174.2996
487
+ 4000 train 6.878099 (lr=1.4108e-05) (hash(x)=52499943)
488
+ 4100 val loss 6.6296
489
+ 4100 val perplexity 757.1706
490
+ 4100 train 6.538190 (lr=4.8426e-05) (hash(x)=48563796)
491
+ 4000 val loss 6.8547
492
+ 4000 val perplexity 948.2933
493
+ 4000 val loss 6.8092
494
+ 4000 val perplexity 906.1381
495
+ 4000 train 6.591424 (lr=3.5271e-05) (hash(x)=52499943)
496
+ 4000 train 6.649190 (lr=2.1162e-05) (hash(x)=52499943)
497
+ 4100 val loss 7.0474
498
+ 4100 val perplexity 1149.8827
499
+ 4100 train 6.979160 (lr=1.3836e-05) (hash(x)=48563796)
500
+ 4200 val loss 6.6076
501
+ 4200 val perplexity 740.7054
502
+ 4200 train 6.571188 (lr=4.7463e-05) (hash(x)=49165143)
503
+ 4100 val loss 6.7875
504
+ 4100 val perplexity 886.7177
505
+ 4100 train 6.698511 (lr=3.4590e-05) (hash(x)=48563796)
506
+ 4100 val loss 6.8290
507
+ 4100 val perplexity 924.2396
508
+ 4100 train 6.757933 (lr=2.0754e-05) (hash(x)=48563796)
509
+ 4200 val loss 7.0299
510
+ 4200 val perplexity 1129.8665
511
+ 4200 train 7.013832 (lr=1.3561e-05) (hash(x)=49165143)
512
+ 4300 val loss 6.5900
513
+ 4300 val perplexity 727.8115
514
+ 4300 train 6.669829 (lr=4.6490e-05) (hash(x)=50973176)
515
+ 4200 val loss 6.8093
516
+ 4200 val perplexity 906.2409
517
+ 4200 train 6.791944 (lr=2.0341e-05) (hash(x)=49165143)
518
+ 4200 val loss 6.7663
519
+ 4200 val perplexity 868.0909
520
+ 4200 train 6.749839 (lr=3.3902e-05) (hash(x)=49165143)
521
+ 4300 val loss 7.0162
522
+ 4300 val perplexity 1114.5115
523
+ 4300 train 7.075418 (lr=1.3283e-05) (hash(x)=50973176)
524
+ 4400 val loss 6.5760
525
+ 4400 val perplexity 717.6350
526
+ 4400 train 6.643890 (lr=4.5509e-05) (hash(x)=55275124)
527
+ 4300 val loss 6.7969
528
+ 4300 val perplexity 895.0617
529
+ 4300 train 6.871970 (lr=1.9924e-05) (hash(x)=50973176)
530
+ 4300 val loss 6.7452
531
+ 4300 val perplexity 849.9814
532
+ 4300 train 6.811849 (lr=3.3207e-05) (hash(x)=50973176)
533
+ 4400 val loss 7.0096
534
+ 4400 val perplexity 1107.1816
535
+ 4400 train 7.059990 (lr=1.3003e-05) (hash(x)=55275124)
536
+ 4500 val loss 6.5617
537
+ 4500 val perplexity 707.4891
538
+ 4500 train 6.930486 (lr=4.4521e-05) (hash(x)=58646505)
539
+ 4400 val loss 6.7828
540
+ 4400 val perplexity 882.5753
541
+ 4400 train 6.836937 (lr=1.9504e-05) (hash(x)=55275124)
542
+ 4400 val loss 6.7274
543
+ 4400 val perplexity 834.9734
544
+ 4500 val loss 6.9973
545
+ 4500 val perplexity 1093.6769
546
+ 4400 train 6.781178 (lr=3.2507e-05) (hash(x)=55275124)
547
+ 4500 train 7.363664 (lr=1.2720e-05) (hash(x)=58646505)
548
+ 4600 val loss 6.5481
549
+ 4600 val perplexity 697.8899
550
+ 4600 train 6.364978 (lr=4.3527e-05) (hash(x)=42554666)
551
+ 4500 val loss 6.7737
552
+ 4500 val perplexity 874.5845
553
+ 4500 train 7.154090 (lr=1.9081e-05) (hash(x)=58646505)
554
+ 4600 val loss 6.9881
555
+ 4600 val perplexity 1083.6766
556
+ 4600 train 6.796409 (lr=1.2436e-05) (hash(x)=42554666)
557
+ 4500 val loss 6.7220
558
+ 4500 val perplexity 830.4634
559
+ 4500 train 7.101217 (lr=3.1801e-05) (hash(x)=58646505)
560
+ 4700 val loss 6.5355
561
+ 4700 val perplexity 689.1530
562
+ 4700 train 6.470394 (lr=4.2528e-05) (hash(x)=47846764)
563
+ 4600 val loss 6.7566
564
+ 4600 val perplexity 859.7034
565
+ 4600 train 6.551023 (lr=1.8655e-05) (hash(x)=42554666)
566
+ 4700 val loss 6.9803
567
+ 4700 val perplexity 1075.2849
568
+ 4700 train 6.901748 (lr=1.2151e-05) (hash(x)=47846764)
569
+ 4600 val loss 6.7128
570
+ 4600 val perplexity 822.8683
571
+ 4600 train 6.511578 (lr=3.1091e-05) (hash(x)=42554666)
572
+ 4800 val loss 6.5178
573
+ 4800 val perplexity 677.0767
574
+ 4800 train 7.064293 (lr=4.1525e-05) (hash(x)=58239019)
575
+ 4700 val loss 6.7513
576
+ 4700 val perplexity 855.1525
577
+ 4800 val loss 6.9707
578
+ 4800 val perplexity 1064.9286
579
+ 4700 train 6.683542 (lr=1.8226e-05) (hash(x)=47846764)
580
+ 4800 train 7.519922 (lr=1.1864e-05) (hash(x)=58239019)
581
+ 4700 val loss 6.6994
582
+ 4700 val perplexity 811.9506
583
+ 4700 train 6.628958 (lr=3.0377e-05) (hash(x)=47846764)
584
+ 4900 val loss 6.5181
585
+ 4900 val perplexity 677.2886
586
+ 4900 train 6.572011 (lr=4.0518e-05) (hash(x)=50711220)
587
+ 4800 val loss 6.7374
588
+ 4800 val perplexity 843.3644
589
+ 4900 val loss 6.9609
590
+ 4900 val perplexity 1054.5725
591
+ 4800 train 7.297200 (lr=1.7796e-05) (hash(x)=58239019)
592
+ 4900 train 7.010784 (lr=1.1577e-05) (hash(x)=50711220)
593
+ 4800 val loss 6.6917
594
+ 4800 val perplexity 805.7076
595
+ 4800 train 7.249681 (lr=2.9661e-05) (hash(x)=58239019)
596
+ 5000 val loss 6.5144
597
+ 5000 val perplexity 674.8202
598
+ 5000 train 6.457778 (lr=3.9510e-05) (hash(x)=45994194)
599
+ 5000 val loss 6.9575
600
+ 5000 val perplexity 1050.9631
601
+ 4900 val loss 6.7300
602
+ 4900 val perplexity 837.1061
603
+ 4900 train 6.796998 (lr=1.7365e-05) (hash(x)=50711220)
604
+ 4900 val loss 6.6740
605
+ 4900 val perplexity 791.5267
606
+ 4900 train 6.739973 (lr=2.8942e-05) (hash(x)=50711220)
607
+ 5000 train 6.903171 (lr=1.1288e-05) (hash(x)=45994194)
608
+ 5100 val loss 6.4941
609
+ 5100 val perplexity 661.2037
610
+ 5100 train 6.372785 (lr=3.8500e-05) (hash(x)=48659050)
611
+ 5100 val loss 6.9444
612
+ 5100 val perplexity 1037.3356
613
+ 5100 train 6.808917 (lr=1.1000e-05) (hash(x)=48659050)
614
+ 5000 val loss 6.7196
615
+ 5000 val perplexity 828.4827
616
+ 5000 val loss 6.6801
617
+ 5000 val perplexity 796.3593
618
+ 5000 train 6.678499 (lr=1.6933e-05) (hash(x)=45994194)
619
+ 5000 train 6.628725 (lr=2.8221e-05) (hash(x)=45994194)
620
+ 5200 val loss 6.4883
621
+ 5200 val perplexity 657.3944
622
+ 5200 train 6.391860 (lr=3.7490e-05) (hash(x)=49369682)
623
+ 5200 val loss 6.9390
624
+ 5200 val perplexity 1031.7484
625
+ 5200 train 6.868454 (lr=1.0712e-05) (hash(x)=49369682)
626
+ 5100 val loss 6.7082
627
+ 5100 val perplexity 819.0988
628
+ 5100 val loss 6.6555
629
+ 5100 val perplexity 777.0222
630
+ 5100 train 6.584312 (lr=1.6500e-05) (hash(x)=48659050)
631
+ 5100 train 6.527529 (lr=2.7500e-05) (hash(x)=48659050)
632
+ 5300 val loss 6.4765
633
+ 5300 val perplexity 649.6841
634
+ 5300 train 6.899568 (lr=3.6482e-05) (hash(x)=57787700)
635
+ 5300 val loss 6.9323
636
+ 5300 val perplexity 1024.8837
637
+ 5300 train 7.331203 (lr=1.0423e-05) (hash(x)=57787700)
638
+ 5200 val loss 6.6987
639
+ 5200 val perplexity 811.3156
640
+ 5200 val loss 6.6475
641
+ 5200 val perplexity 770.8427
642
+ 5200 train 6.620082 (lr=1.6067e-05) (hash(x)=49369682)
643
+ 5200 train 6.554370 (lr=2.6779e-05) (hash(x)=49369682)
644
+ 5400 val loss 6.4795
645
+ 5400 val perplexity 651.6630
646
+ 5400 train 6.409245 (lr=3.5475e-05) (hash(x)=49365400)
647
+ 5400 val loss 6.9261
648
+ 5400 val perplexity 1018.4915
649
+ 5400 train 6.856396 (lr=1.0136e-05) (hash(x)=49365400)
650
+ 5300 val loss 6.6931
651
+ 5300 val perplexity 806.8045
652
+ 5300 train 7.125628 (lr=1.5635e-05) (hash(x)=57787700)
653
+ 5300 val loss 6.6388
654
+ 5300 val perplexity 764.1794
655
+ 5300 train 7.059060 (lr=2.6058e-05) (hash(x)=57787700)
656
+ 5500 val loss 6.4668
657
+ 5500 val perplexity 643.4189
658
+ 5500 train 6.418372 (lr=3.4472e-05) (hash(x)=48720412)
659
+ 5500 val loss 6.9221
660
+ 5500 val perplexity 1014.4897
661
+ 5500 train 6.863081 (lr=9.8491e-06) (hash(x)=48720412)
662
+ 5400 val loss 6.6870
663
+ 5400 val perplexity 801.8978
664
+ 5400 train 6.628658 (lr=1.5204e-05) (hash(x)=49365400)
665
+ 5400 val loss 6.6302
666
+ 5400 val perplexity 757.6622
667
+ 5400 train 6.565665 (lr=2.5339e-05) (hash(x)=49365400)
668
+ 5600 val loss 6.4579
669
+ 5600 val perplexity 637.7405
670
+ 5600 train 6.811317 (lr=3.3473e-05) (hash(x)=55784800)
671
+ 5600 val loss 6.9129
672
+ 5600 val perplexity 1005.1623
673
+ 5600 train 7.253220 (lr=9.5636e-06) (hash(x)=55784800)
674
+ 5500 val loss 6.6801
675
+ 5500 val perplexity 796.3778
676
+ 5500 train 6.631643 (lr=1.4774e-05) (hash(x)=48720412)
677
+ 5500 val loss 6.6239
678
+ 5500 val perplexity 752.8723
679
+ 5500 train 6.577877 (lr=2.4623e-05) (hash(x)=48720412)
680
+ 5700 val loss 6.4601
681
+ 5700 val perplexity 639.1117
682
+ 5700 train 6.290679 (lr=3.2479e-05) (hash(x)=50073634)
683
+ 5700 val loss 6.9127
684
+ 5700 val perplexity 1004.9950
685
+ 5700 train 6.758874 (lr=9.2796e-06) (hash(x)=50073634)
686
+ 5600 val loss 6.6729
687
+ 5600 val perplexity 790.6545
688
+ 5600 train 7.028577 (lr=1.4345e-05) (hash(x)=55784800)
689
+ 5600 val loss 6.6146
690
+ 5600 val perplexity 745.8879
691
+ 5600 train 6.962052 (lr=2.3909e-05) (hash(x)=55784800)
692
+ 5800 val loss 6.4550
693
+ 5800 val perplexity 635.8431
694
+ 5800 train 6.277244 (lr=3.1491e-05) (hash(x)=50170324)
695
+ 5800 val loss 6.9058
696
+ 5800 val perplexity 998.0084
697
+ 5800 train 6.738111 (lr=8.9973e-06) (hash(x)=50170324)
698
+ 5700 val loss 6.6696
699
+ 5700 val perplexity 788.0435
700
+ 5700 train 6.520940 (lr=1.3919e-05) (hash(x)=50073634)
701
+ 5700 val loss 6.6135
702
+ 5700 val perplexity 745.0647
703
+ 5700 train 6.450591 (lr=2.3199e-05) (hash(x)=50073634)
704
+ 5900 val loss 6.4499
705
+ 5900 val perplexity 632.6409
706
+ 5900 train 6.166365 (lr=3.0510e-05) (hash(x)=48410268)
707
+ 5900 val loss 6.9088
708
+ 5900 val perplexity 1001.0864
709
+ 5900 train 6.630468 (lr=8.7171e-06) (hash(x)=48410268)
710
+ 5800 val loss 6.6111
711
+ 5800 val perplexity 743.3010
712
+ 5800 val loss 6.6599
713
+ 5800 val perplexity 780.4451
714
+ 5800 train 6.442124 (lr=2.2493e-05) (hash(x)=50170324)
715
+ 5800 train 6.497536 (lr=1.3496e-05) (hash(x)=50170324)
716
+ 6000 val loss 6.4450
717
+ 6000 val perplexity 629.5153
718
+ 6000 train 6.299426 (lr=2.9537e-05) (hash(x)=49527342)
719
+ 6000 val loss 6.9027
720
+ 6000 val perplexity 994.9408
721
+ 6000 train 6.801613 (lr=8.4393e-06) (hash(x)=49527342)
722
+ 5900 val loss 6.6010
723
+ 5900 val perplexity 735.8242
724
+ 5900 train 6.316404 (lr=2.1793e-05) (hash(x)=48410268)
725
+ 5900 val loss 6.6584
726
+ 5900 val perplexity 779.3054
727
+ 5900 train 6.376041 (lr=1.3076e-05) (hash(x)=48410268)
728
+ 6100 val loss 6.4465
729
+ 6100 val perplexity 630.4760
730
+ 6100 train 6.241553 (lr=2.8574e-05) (hash(x)=49550294)
731
+ 6100 val loss 6.8991
732
+ 6100 val perplexity 991.4178
733
+ 6100 train 6.739030 (lr=8.1640e-06) (hash(x)=49550294)
734
+ 6000 val loss 6.6054
735
+ 6000 val perplexity 739.0561
736
+ 6000 train 6.494583 (lr=2.1098e-05) (hash(x)=49527342)
737
+ 6000 val loss 6.6552
738
+ 6000 val perplexity 776.8270
739
+ 6000 train 6.549714 (lr=1.2659e-05) (hash(x)=49527342)
740
+ 6200 val loss 6.4413
741
+ 6200 val perplexity 627.2357
742
+ 6200 train 5.957240 (lr=2.7621e-05) (hash(x)=42126106)
743
+ 6200 val loss 6.8939
744
+ 6200 val perplexity 986.2703
745
+ 6200 train 6.446277 (lr=7.8917e-06) (hash(x)=42126106)
746
+ 6100 val loss 6.5968
747
+ 6100 val perplexity 732.7717
748
+ 6100 train 6.419609 (lr=2.0410e-05) (hash(x)=49550294)
749
+ 6100 val loss 6.6495
750
+ 6100 val perplexity 772.4101
751
+ 6100 train 6.473972 (lr=1.2246e-05) (hash(x)=49550294)
752
+ 6300 val loss 6.4439
753
+ 6300 val perplexity 628.8303
754
+ 6300 train 6.190170 (lr=2.6679e-05) (hash(x)=49608772)
755
+ 6300 val loss 6.8934
756
+ 6300 val perplexity 985.7131
757
+ 6300 train 6.680425 (lr=7.6226e-06) (hash(x)=49608772)
758
+ 6200 val loss 6.5978
759
+ 6200 val perplexity 733.4502
760
+ 6200 train 6.131247 (lr=1.9729e-05) (hash(x)=42126106)
761
+ 6200 val loss 6.6458
762
+ 6200 val perplexity 769.5638
763
+ 6200 train 6.174097 (lr=1.1838e-05) (hash(x)=42126106)
764
+ 6400 val loss 6.4316
765
+ 6400 val perplexity 621.1544
766
+ 6400 train 5.968055 (lr=2.5749e-05) (hash(x)=52324417)
767
+ 6400 val loss 6.8855
768
+ 6400 val perplexity 977.9658
769
+ 6400 train 6.481584 (lr=7.3569e-06) (hash(x)=52324417)
770
+ 6300 val loss 6.5943
771
+ 6300 val perplexity 730.8967
772
+ 6300 train 6.353149 (lr=1.9056e-05) (hash(x)=49608772)
773
+ 6300 val loss 6.6441
774
+ 6300 val perplexity 768.2593
775
+ 6300 train 6.394776 (lr=1.1434e-05) (hash(x)=49608772)
776
+ 6500 val loss 6.4077
777
+ 6500 val perplexity 606.4824
778
+ 6500 train 6.457166 (lr=2.4833e-05) (hash(x)=46207215)
779
+ 6500 val loss 6.8695
780
+ 6500 val perplexity 962.4999
781
+ 6500 train 6.899956 (lr=7.0950e-06) (hash(x)=46207215)
782
+ 6400 val loss 6.5863
783
+ 6400 val perplexity 725.1245
784
+ 6400 train 6.154537 (lr=1.8392e-05) (hash(x)=52324417)
785
+ 6400 val loss 6.6324
786
+ 6400 val perplexity 759.3197
787
+ 6400 train 6.207419 (lr=1.1035e-05) (hash(x)=52324417)
788
+ 6600 val loss 6.3960
789
+ 6600 val perplexity 599.4461
790
+ 6600 train 6.364021 (lr=2.3930e-05) (hash(x)=49027014)
791
+ 6600 val loss 6.8598
792
+ 6600 val perplexity 953.2013
793
+ 6600 train 6.788251 (lr=6.8372e-06) (hash(x)=49027014)
794
+ 6500 val loss 6.5651
795
+ 6500 val perplexity 709.9162
796
+ 6500 train 6.600822 (lr=1.7738e-05) (hash(x)=46207215)
797
+ 6500 val loss 6.6128
798
+ 6500 val perplexity 744.5444
799
+ 6500 train 6.661718 (lr=1.0643e-05) (hash(x)=46207215)
800
+ 6700 val loss 6.3989
801
+ 6700 val perplexity 601.1656
802
+ 6700 train 6.432356 (lr=2.3042e-05) (hash(x)=46232513)
803
+ 6700 val loss 6.8536
804
+ 6700 val perplexity 947.2444
805
+ 6700 train 6.824396 (lr=6.5835e-06) (hash(x)=46232513)
806
+ 6600 val loss 6.5582
807
+ 6600 val perplexity 704.9709
808
+ 6600 train 6.505264 (lr=1.7093e-05) (hash(x)=49027014)
809
+ 6600 val loss 6.6036
810
+ 6600 val perplexity 737.7280
811
+ 6600 train 6.547662 (lr=1.0256e-05) (hash(x)=49027014)
812
+ 6800 val loss 6.3890
813
+ 6800 val perplexity 595.2584
814
+ 6800 train 6.318826 (lr=2.2171e-05) (hash(x)=47348403)
815
+ 6800 val loss 6.8536
816
+ 6800 val perplexity 947.2534
817
+ 6800 train 6.787578 (lr=6.3345e-06) (hash(x)=47348403)
818
+ 6700 val loss 6.5480
819
+ 6700 val perplexity 697.8230
820
+ 6700 train 6.583887 (lr=1.6459e-05) (hash(x)=46232513)
821
+ 6700 val loss 6.5991
822
+ 6700 val perplexity 734.4021
823
+ 6700 train 6.615988 (lr=9.8753e-06) (hash(x)=46232513)
824
+ 6900 val loss 6.3967
825
+ 6900 val perplexity 599.8644
826
+ 6900 train 6.417536 (lr=2.1316e-05) (hash(x)=49806647)
827
+ 6900 val loss 6.8555
828
+ 6900 val perplexity 949.0623
829
+ 6900 train 6.840402 (lr=6.0902e-06) (hash(x)=49806647)
830
+ 6800 val loss 6.5493
831
+ 6800 val perplexity 698.7370
832
+ 6800 train 6.484587 (lr=1.5836e-05) (hash(x)=47348403)
833
+ 7000 val loss 6.3745
834
+ 7000 val perplexity 586.6633
835
+ 6800 val loss 6.5951
836
+ 6800 val perplexity 731.4786
837
+ 7000 train 6.413634 (lr=2.0478e-05) (hash(x)=50893018)
838
+ 6800 train 6.525431 (lr=9.5017e-06) (hash(x)=47348403)
839
+ 7000 val loss 6.8389
840
+ 7000 val perplexity 933.5077
841
+ 7000 train 6.886159 (lr=5.8510e-06) (hash(x)=50893018)
842
+ 6900 val loss 6.5492
843
+ 6900 val perplexity 698.6997
844
+ 6900 train 6.569705 (lr=1.5225e-05) (hash(x)=49806647)
845
+ 7100 val loss 6.3662
846
+ 7100 val perplexity 581.8609
847
+ 7100 train 6.399204 (lr=1.9660e-05) (hash(x)=49157639)
848
+ 6900 val loss 6.5899
849
+ 6900 val perplexity 727.7289
850
+ 6900 train 6.611711 (lr=9.1353e-06) (hash(x)=49806647)
851
+ 7100 val loss 6.8357
852
+ 7100 val perplexity 930.4750
853
+ 7100 train 6.874208 (lr=5.6170e-06) (hash(x)=49157639)
854
+ 7000 val loss 6.5346
855
+ 7000 val perplexity 688.5408
856
+ 7000 train 6.584486 (lr=1.4627e-05) (hash(x)=50893018)
857
+ 7200 val loss 6.3606
858
+ 7200 val perplexity 578.5686
859
+ 7200 train 6.369837 (lr=1.8860e-05) (hash(x)=47014759)
860
+ 7000 val loss 6.5786
861
+ 7000 val perplexity 719.5305
862
+ 7000 train 6.634720 (lr=8.7764e-06) (hash(x)=50893018)
863
+ 7200 val loss 6.8313
864
+ 7200 val perplexity 926.3760
865
+ 7200 train 6.809377 (lr=5.3886e-06) (hash(x)=47014759)
866
+ 7100 val loss 6.5322
867
+ 7100 val perplexity 686.9103
868
+ 7300 val loss 6.3576
869
+ 7300 val perplexity 576.8370
870
+ 7100 train 6.576263 (lr=1.4043e-05) (hash(x)=49157639)
871
+ 7300 train 6.341197 (lr=1.8081e-05) (hash(x)=47325591)
872
+ 7100 val loss 6.5737
873
+ 7100 val perplexity 716.0148
874
+ 7100 train 6.610196 (lr=8.4255e-06) (hash(x)=49157639)
875
+ 7300 val loss 6.8313
876
+ 7300 val perplexity 926.4335
877
+ 7300 train 6.779392 (lr=5.1659e-06) (hash(x)=47325591)
878
+ 7400 val loss 6.3512
879
+ 7400 val perplexity 573.1561
880
+ 7400 train 6.235737 (lr=1.7323e-05) (hash(x)=49184604)
881
+ 7200 val loss 6.5260
882
+ 7200 val perplexity 682.6921
883
+ 7200 train 6.518712 (lr=1.3471e-05) (hash(x)=47014759)
884
+ 7200 val loss 6.5675
885
+ 7200 val perplexity 711.5921
886
+ 7200 train 6.546741 (lr=8.0829e-06) (hash(x)=47014759)
887
+ 7400 val loss 6.8224
888
+ 7400 val perplexity 918.1414
889
+ 7400 train 6.680563 (lr=4.9493e-06) (hash(x)=49184604)
890
+ 7500 val loss 6.3495
891
+ 7500 val perplexity 572.2331
892
+ 7300 val loss 6.5227
893
+ 7300 val perplexity 680.4317
894
+ 7300 train 6.489656 (lr=1.2915e-05) (hash(x)=47325591)
895
+ 7500 train 6.565956 (lr=1.6586e-05) (hash(x)=55053584)
896
+ 7300 val loss 6.5669
897
+ 7300 val perplexity 711.1389
898
+ 7300 train 6.538014 (lr=7.7489e-06) (hash(x)=47325591)
899
+ 7500 val loss 6.8192
900
+ 7500 val perplexity 915.2512
901
+ 7500 train 7.127323 (lr=4.7389e-06) (hash(x)=55053584)
902
+ 7400 val loss 6.5154
903
+ 7400 val perplexity 675.4557
904
+ 7400 train 6.394844 (lr=1.2373e-05) (hash(x)=49184604)
905
+ 7600 val loss 6.3460
906
+ 7600 val perplexity 570.1979
907
+ 7600 train 6.293826 (lr=1.5872e-05) (hash(x)=48693923)
908
+ 7400 val loss 6.5571
909
+ 7400 val perplexity 704.2495
910
+ 7400 train 6.427595 (lr=7.4239e-06) (hash(x)=49184604)
911
+ 7600 val loss 6.8172
912
+ 7600 val perplexity 913.3918
913
+ 7600 train 6.765959 (lr=4.5349e-06) (hash(x)=48693923)
914
+ 7500 val loss 6.5101
915
+ 7500 val perplexity 671.9246
916
+ 7500 train 6.810478 (lr=1.1847e-05) (hash(x)=55053584)
917
+ 7700 val loss 6.3417
918
+ 7700 val perplexity 567.7472
919
+ 7700 train 5.880659 (lr=1.5181e-05) (hash(x)=40952882)
920
+ 7500 val loss 6.5521
921
+ 7500 val perplexity 700.7015
922
+ 7500 train 6.888753 (lr=7.1083e-06) (hash(x)=55053584)
923
+ 7700 val loss 6.8143
924
+ 7700 val perplexity 910.7701
925
+ 7700 train 6.320709 (lr=4.3375e-06) (hash(x)=40952882)
926
+ 7800 val loss 6.3429
927
+ 7800 val perplexity 568.4157
928
+ 7800 train 6.443004 (lr=1.4514e-05) (hash(x)=52487845)
929
+ 7600 val loss 6.5085
930
+ 7600 val perplexity 670.8518
931
+ 7600 train 6.467136 (lr=1.1337e-05) (hash(x)=48693923)
932
+ 7600 val loss 6.5490
933
+ 7600 val perplexity 698.5761
934
+ 7600 train 6.507745 (lr=6.8023e-06) (hash(x)=48693923)
935
+ 7800 val loss 6.8130
936
+ 7800 val perplexity 909.5957
937
+ 7800 train 6.895899 (lr=4.1470e-06) (hash(x)=52487845)
938
+ 7900 val loss 6.3350
939
+ 7900 val perplexity 563.9832
940
+ 7900 train 6.403783 (lr=1.3872e-05) (hash(x)=50221547)
941
+ 7700 val loss 6.5029
942
+ 7700 val perplexity 667.0413
943
+ 7700 train 6.025047 (lr=1.0844e-05) (hash(x)=40952882)
944
+ 7700 val loss 6.5466
945
+ 7700 val perplexity 696.8989
946
+ 7700 train 6.057941 (lr=6.5062e-06) (hash(x)=40952882)
947
+ 7900 val loss 6.8090
948
+ 7900 val perplexity 905.9458
949
+ 7900 train 6.892159 (lr=3.9635e-06) (hash(x)=50221547)
950
+ 8000 val loss 6.3324
951
+ 8000 val perplexity 562.5143
952
+ 8000 train 6.561235 (lr=1.3255e-05) (hash(x)=62294204)
953
+ 7800 val loss 6.4995
954
+ 7800 val perplexity 664.7764
955
+ 7800 train 6.599380 (lr=1.0367e-05) (hash(x)=52487845)
956
+ 7800 val loss 6.5443
957
+ 7800 val perplexity 695.2772
958
+ 7800 train 6.642978 (lr=6.2205e-06) (hash(x)=52487845)
959
+ 8000 val loss 6.8080
960
+ 8000 val perplexity 905.0434
961
+ 8000 train 7.153604 (lr=3.7873e-06) (hash(x)=62294204)
962
+ 8100 val loss 6.3286
963
+ 8100 val perplexity 560.3528
964
+ 8100 train 5.986205 (lr=1.2665e-05) (hash(x)=44401967)
965
+ 7900 val loss 6.4989
966
+ 7900 val perplexity 664.4284
967
+ 7900 train 6.581850 (lr=9.9088e-06) (hash(x)=50221547)
968
+ 7900 val loss 6.5393
969
+ 7900 val perplexity 691.8358
970
+ 7900 train 6.626468 (lr=5.9453e-06) (hash(x)=50221547)
971
+ 8100 val loss 6.8051
972
+ 8100 val perplexity 902.4113
973
+ 8100 train 6.457790 (lr=3.6184e-06) (hash(x)=44401967)
974
+ 8200 val loss 6.3292
975
+ 8200 val perplexity 560.6861
976
+ 8200 train 6.249769 (lr=1.2100e-05) (hash(x)=52769095)
977
+ 8000 val loss 6.4917
978
+ 8000 val perplexity 659.6194
979
+ 8000 train 6.824268 (lr=9.4682e-06) (hash(x)=62294204)
980
+ 8000 val loss 6.5362
981
+ 8000 val perplexity 689.6366
982
+ 8000 train 6.911385 (lr=5.6809e-06) (hash(x)=62294204)
983
+ 8200 val loss 6.8045
984
+ 8200 val perplexity 901.8989
985
+ 8200 train 6.761432 (lr=3.4572e-06) (hash(x)=52769095)
986
+ 8300 val loss 6.3299
987
+ 8300 val perplexity 561.1138
988
+ 8300 train 6.275354 (lr=1.1563e-05) (hash(x)=56829883)
989
+ 8100 val loss 6.4893
990
+ 8100 val perplexity 658.0931
991
+ 8100 train 6.156446 (lr=9.0461e-06) (hash(x)=44401967)
992
+ 8100 val loss 6.5327
993
+ 8100 val perplexity 687.2546
994
+ 8100 train 6.194956 (lr=5.4277e-06) (hash(x)=44401967)
995
+ 8300 val loss 6.8066
996
+ 8300 val perplexity 903.8031
997
+ 8300 train 6.767237 (lr=3.3037e-06) (hash(x)=56829883)
998
+ 8400 val loss 6.3270
999
+ 8400 val perplexity 559.4752
1000
+ 8400 train 6.257383 (lr=1.1053e-05) (hash(x)=52147375)
1001
+ 8200 val loss 6.4897
1002
+ 8200 val perplexity 658.3050
1003
+ 8200 train 6.420258 (lr=8.6430e-06) (hash(x)=52769095)
1004
+ 8200 val loss 6.5323
1005
+ 8200 val perplexity 686.9604
1006
+ 8200 train 6.474307 (lr=5.1858e-06) (hash(x)=52769095)
1007
+ 8400 val loss 6.8053
1008
+ 8400 val perplexity 902.6118
1009
+ 8400 train 6.752232 (lr=3.1581e-06) (hash(x)=52147375)
1010
+ 8500 val loss 6.3305
1011
+ 8500 val perplexity 561.4124
1012
+ 8500 train 6.627474 (lr=1.0572e-05) (hash(x)=60197820)
1013
+ 8300 val loss 6.4888
1014
+ 8300 val perplexity 657.7073
1015
+ 8300 train 6.448828 (lr=8.2593e-06) (hash(x)=56829883)
1016
+ 8300 val loss 6.5343
1017
+ 8300 val perplexity 688.3612
1018
+ 8300 train 6.490636 (lr=4.9556e-06) (hash(x)=56829883)
1019
+ 8500 val loss 6.8045
1020
+ 8500 val perplexity 901.8800
1021
+ 8500 train 7.037055 (lr=3.0206e-06) (hash(x)=60197820)
1022
+ 8600 val loss 6.3244
1023
+ 8600 val perplexity 558.0261
1024
+ 8600 train 6.074996 (lr=1.0119e-05) (hash(x)=49377068)
1025
+ 8400 val loss 6.4885
1026
+ 8400 val perplexity 657.5069
1027
+ 8400 train 6.421007 (lr=7.8953e-06) (hash(x)=52147375)
1028
+ 8400 val loss 6.5331
1029
+ 8400 val perplexity 687.5152
1030
+ 8400 train 6.478803 (lr=4.7372e-06) (hash(x)=52147375)
1031
+ 8600 val loss 6.8019
1032
+ 8600 val perplexity 899.5659
1033
+ 8600 train 6.592875 (lr=2.8913e-06) (hash(x)=49377068)
1034
+ 8700 val loss 6.3242
1035
+ 8700 val perplexity 557.9348
1036
+ 8700 train 6.308025 (lr=9.6960e-06) (hash(x)=51092724)
1037
+ 8500 val loss 6.4867
1038
+ 8500 val perplexity 656.3723
1039
+ 8500 train 6.771042 (lr=7.5515e-06) (hash(x)=60197820)
1040
+ 8500 val loss 6.5317
1041
+ 8500 val perplexity 686.5867
1042
+ 8500 train 6.808648 (lr=4.5309e-06) (hash(x)=60197820)
1043
+ 8700 val loss 6.8009
1044
+ 8700 val perplexity 898.6459
1045
+ 8700 train 6.764141 (lr=2.7703e-06) (hash(x)=51092724)
1046
+ 8800 val loss 6.3245
1047
+ 8800 val perplexity 558.0521
1048
+ 8800 train 6.319681 (lr=9.3021e-06) (hash(x)=48642928)
1049
+ 8600 val loss 6.4817
1050
+ 8600 val perplexity 653.0836
1051
+ 8600 train 6.252263 (lr=7.2282e-06) (hash(x)=49377068)
1052
+ 8600 val loss 6.5274
1053
+ 8600 val perplexity 683.5885
1054
+ 8600 train 6.298992 (lr=4.3369e-06) (hash(x)=49377068)
1055
+ 8800 val loss 6.7992
1056
+ 8800 val perplexity 897.1229
1057
+ 8800 train 6.821721 (lr=2.6577e-06) (hash(x)=48642928)
1058
+ 8900 val loss 6.3333
1059
+ 8900 val perplexity 563.0059
1060
+ 8900 train 6.536117 (lr=8.9382e-06) (hash(x)=55342246)
1061
+ 8700 val loss 6.4842
1062
+ 8700 val perplexity 654.7362
1063
+ 8700 train 6.468521 (lr=6.9257e-06) (hash(x)=51092724)
1064
+ 8700 val loss 6.5261
1065
+ 8700 val perplexity 682.7126
1066
+ 8700 train 6.495368 (lr=4.1554e-06) (hash(x)=51092724)
1067
+ 8900 val loss 6.8031
1068
+ 8900 val perplexity 900.6793
1069
+ 8900 train 7.076955 (lr=2.5538e-06) (hash(x)=55342246)
1070
+ 9000 val loss 6.3072
1071
+ 9000 val perplexity 548.5292
1072
+ 9000 train 6.311553 (lr=8.6047e-06) (hash(x)=48093368)
1073
+ 8800 val loss 6.4768
1074
+ 8800 val perplexity 649.8600
1075
+ 8800 train 6.481946 (lr=6.6444e-06) (hash(x)=48642928)
1076
+ 8800 val loss 6.5237
1077
+ 8800 val perplexity 681.0806
1078
+ 8800 train 6.529369 (lr=3.9866e-06) (hash(x)=48642928)
1079
+ 9000 val loss 6.7872
1080
+ 9000 val perplexity 886.3854
1081
+ 9000 train 6.737963 (lr=2.4585e-06) (hash(x)=48093368)
1082
+ 9100 val loss 6.3026
1083
+ 9100 val perplexity 545.9979
1084
+ 9100 train 6.379838 (lr=8.3020e-06) (hash(x)=48578183)
1085
+ 8900 val loss 6.4853
1086
+ 8900 val perplexity 655.4434
1087
+ 8900 train 6.709653 (lr=6.3845e-06) (hash(x)=55342246)
1088
+ 8900 val loss 6.5276
1089
+ 8900 val perplexity 683.7887
1090
+ 8900 train 6.777533 (lr=3.8307e-06) (hash(x)=55342246)
1091
+ 9100 val loss 6.7850
1092
+ 9100 val perplexity 884.5032
1093
+ 9100 train 6.877484 (lr=2.3720e-06) (hash(x)=48578183)
1094
+ 9200 val loss 6.2996
1095
+ 9200 val perplexity 544.3685
1096
+ 9200 train 6.467710 (lr=8.0302e-06) (hash(x)=50794720)
1097
+ 9000 val loss 6.4627
1098
+ 9000 val perplexity 640.7952
1099
+ 9000 train 6.448120 (lr=6.1462e-06) (hash(x)=48093368)
1100
+ 9000 val loss 6.5091
1101
+ 9000 val perplexity 671.1916
1102
+ 9000 train 6.476361 (lr=3.6877e-06) (hash(x)=48093368)
1103
+ 9200 val loss 6.7823
1104
+ 9200 val perplexity 882.1112
1105
+ 9200 train 6.995069 (lr=2.2943e-06) (hash(x)=50794720)
1106
+ 9300 val loss 6.2941
1107
+ 9300 val perplexity 541.3758
1108
+ 9300 train 6.044296 (lr=7.7898e-06) (hash(x)=46513190)
1109
+ 9100 val loss 6.4606
1110
+ 9100 val perplexity 639.4345
1111
+ 9100 train 6.546059 (lr=5.9300e-06) (hash(x)=48578183)
1112
+ 9100 val loss 6.5059
1113
+ 9100 val perplexity 669.0497
1114
+ 9100 train 6.587250 (lr=3.5580e-06) (hash(x)=48578183)
1115
+ 9300 val loss 6.7809
1116
+ 9300 val perplexity 880.8246
1117
+ 9300 train 6.540437 (lr=2.2256e-06) (hash(x)=46513190)
1118
+ 9400 val loss 6.2939
1119
+ 9400 val perplexity 541.2728
1120
+ 9400 train 5.963445 (lr=7.5809e-06) (hash(x)=43808238)
1121
+ 9200 val loss 6.4558
1122
+ 9200 val perplexity 636.3748
1123
+ 9200 train 6.658762 (lr=5.7359e-06) (hash(x)=50794720)
1124
+ 9200 val loss 6.5040
1125
+ 9200 val perplexity 667.7790
1126
+ 9200 train 6.728305 (lr=3.4415e-06) (hash(x)=50794720)
1127
+ 9400 val loss 6.7788
1128
+ 9400 val perplexity 879.0392
1129
+ 9400 train 6.409406 (lr=2.1660e-06) (hash(x)=43808238)
1130
+ 9500 val loss 6.2907
1131
+ 9500 val perplexity 539.5046
1132
+ 9500 train 6.018129 (lr=7.4038e-06) (hash(x)=45021888)
1133
+ 9300 val loss 6.4518
1134
+ 9300 val perplexity 633.8415
1135
+ 9300 train 6.202276 (lr=5.5641e-06) (hash(x)=46513190)
1136
+ 9300 val loss 6.5013
1137
+ 9300 val perplexity 666.0020
1138
+ 9300 train 6.249645 (lr=3.3385e-06) (hash(x)=46513190)
1139
+ 9500 val loss 6.7771
1140
+ 9500 val perplexity 877.5475
1141
+ 9500 train 6.546412 (lr=2.1154e-06) (hash(x)=45021888)
1142
+ 9600 val loss 6.2910
1143
+ 9600 val perplexity 539.6711
1144
+ 9600 train 6.393634 (lr=7.2586e-06) (hash(x)=56525570)
1145
+ 9400 val loss 6.4526
1146
+ 9400 val perplexity 634.3289
1147
+ 9400 train 6.111716 (lr=5.4149e-06) (hash(x)=43808238)
1148
+ 9400 val loss 6.5003
1149
+ 9400 val perplexity 665.3675
1150
+ 9400 train 6.157181 (lr=3.2490e-06) (hash(x)=43808238)
1151
+ 9600 val loss 6.7750
1152
+ 9600 val perplexity 875.6418
1153
+ 9600 train 6.851060 (lr=2.0739e-06) (hash(x)=56525570)
1154
+ 9700 val loss 6.2842
1155
+ 9700 val perplexity 536.0264
1156
+ 9700 train 6.466522 (lr=7.1456e-06) (hash(x)=52585913)
1157
+ 9500 val loss 6.4483
1158
+ 9500 val perplexity 631.6206
1159
+ 9500 train 6.191110 (lr=5.2884e-06) (hash(x)=45021888)
1160
+ 9500 val loss 6.4981
1161
+ 9500 val perplexity 663.8555
1162
+ 9500 train 6.242024 (lr=3.1730e-06) (hash(x)=45021888)
1163
+ 9700 val loss 6.7736
1164
+ 9700 val perplexity 874.4531
1165
+ 9700 train 6.914565 (lr=2.0416e-06) (hash(x)=52585913)
1166
+ 9800 val loss 6.2846
1167
+ 9800 val perplexity 536.2417
1168
+ 9800 train 6.464126 (lr=7.0647e-06) (hash(x)=52344698)
1169
+ 9600 val loss 6.4494
1170
+ 9600 val perplexity 632.3007
1171
+ 9600 train 6.547867 (lr=5.1847e-06) (hash(x)=56525570)
1172
+ 9600 val loss 6.4961
1173
+ 9600 val perplexity 662.5447
1174
+ 9600 train 6.603984 (lr=3.1108e-06) (hash(x)=56525570)
1175
+ 9800 val loss 6.7723
1176
+ 9800 val perplexity 873.3168
1177
+ 9800 train 6.975519 (lr=2.0185e-06) (hash(x)=52344698)
1178
+ 9900 val loss 6.2817
1179
+ 9900 val perplexity 534.6906
1180
+ 9900 train 6.284701 (lr=7.0162e-06) (hash(x)=51740945)
1181
+ 9700 val loss 6.4440
1182
+ 9700 val perplexity 628.9359
1183
+ 9700 val loss 6.4940
1184
+ 9700 val perplexity 661.1810
1185
+ 9700 train 6.661100 (lr=3.0624e-06) (hash(x)=52585913)
1186
+ 9700 train 6.610982 (lr=5.1040e-06) (hash(x)=52585913)
1187
+ 9800 val loss 6.4918
1188
+ 9800 val perplexity 659.7147
1189
+ 9800 train 6.709685 (lr=3.0277e-06) (hash(x)=52344698)
1190
+ 9900 val loss 6.7701
1191
+ 9900 val perplexity 871.4020
1192
+ 9900 train 6.791570 (lr=2.0046e-06) (hash(x)=51740945)
1193
+ 9999 val loss 6.2791
1194
+ 9999 val perplexity 533.3184
1195
+ 9800 val loss 6.4419
1196
+ 9800 val perplexity 627.5698
1197
+ 9800 train 6.656963 (lr=5.0462e-06) (hash(x)=52344698)
1198
+ 9900 val loss 6.4900
1199
+ 9900 val perplexity 658.5361
1200
+ 9900 train 6.512582 (lr=3.0069e-06) (hash(x)=51740945)
1201
+ 9999 val loss 6.4887
1202
+ 9999 val perplexity 657.6709
1203
+ 9999 val loss 6.7688
1204
+ 9999 val perplexity 870.2236
1205
+ 9900 val loss 6.4399
1206
+ 9900 val perplexity 626.3513
1207
+ 9900 train 6.445999 (lr=5.0116e-06) (hash(x)=51740945)
1208
+ 9999 val loss 6.4393
1209
+ 9999 val perplexity 625.9971
attention_kindselective_n_heads8_seed1338/model_02500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0f421aaeba65c1ac7d8f57367d664ff7980380b67f84b76cf91bd1c70585212
3
+ size 257976706
attention_kindselective_n_heads8_seed1338/model_05000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c4f9fc9b119cc1c40dd40402afd1d7d323f35ed99f346d00949f1f5bfc4ada0
3
+ size 257976706
attention_kindselective_n_heads8_seed1338/model_07500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a56b7521c2f642fe74f79d77ad0872fad2baca63cfa6818f3aa12e176fbc01ae
3
+ size 257976706
attention_kindselective_n_heads8_seed1338/model_09999.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2438144af715e7e5d8e4af477a0526246dcaafb32968d2669293d131fc55497
3
+ size 257976706
attention_kindselective_n_heads8_seed1338/optimizer_02500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2abe43e4b5e3c310d8acd2bab9e701733ffdc7fb74d3f2a93e525c2ff8db6f7e
3
+ size 509672838
attention_kindselective_n_heads8_seed1338/optimizer_05000.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48a423a0d6c8d63c72833cdeedcfbb50adb632c8e5335799734d33897f804c57
3
+ size 509672838
attention_kindselective_n_heads8_seed1338/optimizer_07500.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70054d22ec62aa54be8932a09eedbb337d629e7c18f1f1de827eaef7acb3375e
3
+ size 509672838
attention_kindselective_n_heads8_seed1338/optimizer_09999.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edfb58324c9646bbe15807f28e2c49e576fa2581cd7d8fb52ea20593dad7c531
3
+ size 509672838