andrew-healey commited on
Commit
3192efe
·
verified ·
1 Parent(s): 1d6279b

Upload folder using huggingface_hub

Browse files
n_heads12_lr10e-4_total_batch_size61440_seed1339/args.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"hellaswag": true, "attention_kind": "selective", "log_dir": "wider_is_better_11/n_heads12_lr10e-4_total_batch_size61440_seed1339", "resume_checkpoint": null, "resume_optimizer": false, "add_a_head": false, "add_head_to_start": true, "new_head_init": "normal", "n_heads": 12, "protect_bos_token": true, "prevent_from_masking_myself": true, "max_steps": 8750, "warmup_steps": 500, "group": "wider_is_better_11", "use_wandb": true, "kill_self_after_run": false, "random_seed": 1339, "memory_penalty_epsilon": 0.1, "selection_head_linear_combo": "none", "selection_head_linear_combo_scale": 1.0, "protection_kind": "none", "leaky_relu_alpha": null, "leaky_relu_bias": null, "use_compile": true, "use_mini_model": false, "upload_to_hf": true, "seq_len": 256, "batch_size": 60, "total_batch_size": 61440, "protection_head_scaling_factor": 1.0, "protection_head_bias": 0.0, "n_sliced_masks": null, "n_latent_masks": null, "mask_layernorm": false, "residual_attention_masks": false, "compute_base_shapes": false, "base_shapes_savefile": null, "mup": true, "disable_selection": false, "mup_enable_coord_check_logging": false, "max_lr": 0.001, "decay_lr": true, "readout_zero_init": false, "query_zero_init": false, "l1_loss": false, "debugpy": false, "key": "n_heads12_10e-4_61440", "n_embd": 768}
n_heads12_lr10e-4_total_batch_size61440_seed1339/dataloader_08749.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9bbf5b7eef643b850a13a0d3881f9b995af25e67f722e39e1853dc09a06ad1dc
3
+ size 964
n_heads12_lr10e-4_total_batch_size61440_seed1339/log2.txt ADDED
@@ -0,0 +1,1054 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ max_steps: 8750
2
+ 0 val loss 10.9491
3
+ 0 val perplexity 56902.2969
4
+ 0 train 10.947037 (lr=2.0000e-06) (hash(x)=84111305)
5
+ 10 train 9.805919 (lr=2.2000e-05) (hash(x)=76061209)
6
+ 20 train 9.400036 (lr=4.2000e-05) (hash(x)=72815142)
7
+ 30 train 9.092428 (lr=6.2000e-05) (hash(x)=74281158)
8
+ 40 train 8.515334 (lr=8.2000e-05) (hash(x)=70759787)
9
+ 50 train 8.091492 (lr=1.0200e-04) (hash(x)=67498044)
10
+ 60 train 7.754110 (lr=1.2200e-04) (hash(x)=74288138)
11
+ 70 train 7.565042 (lr=1.4200e-04) (hash(x)=72712110)
12
+ 80 train 7.567488 (lr=1.6200e-04) (hash(x)=76893978)
13
+ 90 train 7.350292 (lr=1.8200e-04) (hash(x)=69479499)
14
+ 100 val loss 7.2845
15
+ 100 val perplexity 1457.4838
16
+ 100 train 7.251658 (lr=2.0200e-04) (hash(x)=78805238)
17
+ 110 train 7.136649 (lr=2.2200e-04) (hash(x)=72379905)
18
+ 120 train 7.046024 (lr=2.4200e-04) (hash(x)=84941914)
19
+ 130 train 6.839346 (lr=2.6200e-04) (hash(x)=74520609)
20
+ 140 train 6.884595 (lr=2.8200e-04) (hash(x)=78600114)
21
+ 150 train 6.565871 (lr=3.0200e-04) (hash(x)=71364632)
22
+ 160 train 6.555339 (lr=3.2200e-04) (hash(x)=88692902)
23
+ 170 train 6.612408 (lr=3.4200e-04) (hash(x)=76813970)
24
+ 180 train 6.464255 (lr=3.6200e-04) (hash(x)=78277546)
25
+ 190 train 6.393544 (lr=3.8200e-04) (hash(x)=69274625)
26
+ 200 val loss 6.4242
27
+ 200 val perplexity 616.6071
28
+ 200 train 6.346609 (lr=4.0200e-04) (hash(x)=74752080)
29
+ 210 train 6.241256 (lr=4.2200e-04) (hash(x)=73366585)
30
+ 220 train 6.225573 (lr=4.4200e-04) (hash(x)=82441512)
31
+ 230 train 5.997927 (lr=4.6200e-04) (hash(x)=70370492)
32
+ 240 train 6.042721 (lr=4.8200e-04) (hash(x)=76271528)
33
+ 250 train 5.975184 (lr=5.0200e-04) (hash(x)=70391215)
34
+ 260 train 5.895554 (lr=5.2200e-04) (hash(x)=70932307)
35
+ 270 train 5.837888 (lr=5.4200e-04) (hash(x)=73903274)
36
+ 280 train 5.877750 (lr=5.6200e-04) (hash(x)=72833455)
37
+ 290 train 5.920525 (lr=5.8200e-04) (hash(x)=86446336)
38
+ 300 val loss 5.9734
39
+ 300 val perplexity 392.8230
40
+ 300 train 5.708982 (lr=6.0200e-04) (hash(x)=72231879)
41
+ 310 train 5.837598 (lr=6.2200e-04) (hash(x)=83289406)
42
+ 320 train 5.907551 (lr=6.4200e-04) (hash(x)=79292084)
43
+ 330 train 6.068362 (lr=6.6200e-04) (hash(x)=81963141)
44
+ 340 train 5.940867 (lr=6.8200e-04) (hash(x)=76688193)
45
+ 350 train 5.745305 (lr=7.0200e-04) (hash(x)=67974208)
46
+ 360 train 5.920948 (lr=7.2200e-04) (hash(x)=71759924)
47
+ 370 train 5.829780 (lr=7.4200e-04) (hash(x)=82538335)
48
+ 380 train 5.688867 (lr=7.6200e-04) (hash(x)=73719341)
49
+ 390 train 5.840536 (lr=7.8200e-04) (hash(x)=81731118)
50
+ 400 val loss 5.6736
51
+ 400 val perplexity 291.0920
52
+ 400 train 5.625833 (lr=8.0200e-04) (hash(x)=65538392)
53
+ 410 train 5.602108 (lr=8.2200e-04) (hash(x)=73037722)
54
+ 420 train 5.507573 (lr=8.4200e-04) (hash(x)=71698939)
55
+ 430 train 5.731693 (lr=8.6200e-04) (hash(x)=89649867)
56
+ 440 train 5.827013 (lr=8.8200e-04) (hash(x)=59262658)
57
+ 450 train 5.457302 (lr=9.0200e-04) (hash(x)=79057900)
58
+ 460 train 5.476742 (lr=9.2200e-04) (hash(x)=68890702)
59
+ 470 train 5.486958 (lr=9.4200e-04) (hash(x)=69474321)
60
+ 480 train 5.423604 (lr=9.6200e-04) (hash(x)=74715332)
61
+ 490 train 5.542636 (lr=9.8200e-04) (hash(x)=81929495)
62
+ 500 val loss 5.4476
63
+ 500 val perplexity 232.2077
64
+ 500 train 5.598450 (lr=1.0000e-03) (hash(x)=81405581)
65
+ 510 train 5.197465 (lr=1.0000e-03) (hash(x)=66596183)
66
+ 520 train 5.512326 (lr=9.9999e-04) (hash(x)=75298362)
67
+ 530 train 5.317410 (lr=9.9997e-04) (hash(x)=78449774)
68
+ 540 train 5.295600 (lr=9.9995e-04) (hash(x)=75749875)
69
+ 550 train 5.244550 (lr=9.9992e-04) (hash(x)=75371538)
70
+ 560 train 5.180340 (lr=9.9988e-04) (hash(x)=74708150)
71
+ 570 train 5.201348 (lr=9.9984e-04) (hash(x)=75984401)
72
+ 580 train 5.170445 (lr=9.9979e-04) (hash(x)=78930428)
73
+ 590 train 5.445223 (lr=9.9974e-04) (hash(x)=70858901)
74
+ 600 val loss 5.2832
75
+ 600 val perplexity 196.9989
76
+ 600 train 5.174394 (lr=9.9967e-04) (hash(x)=86811215)
77
+ 610 train 5.040601 (lr=9.9961e-04) (hash(x)=74537543)
78
+ 620 train 5.022754 (lr=9.9953e-04) (hash(x)=64425159)
79
+ 630 train 4.906867 (lr=9.9945e-04) (hash(x)=72575441)
80
+ 640 train 4.915975 (lr=9.9936e-04) (hash(x)=77886921)
81
+ 650 train 5.029866 (lr=9.9927e-04) (hash(x)=73649304)
82
+ 660 train 4.797788 (lr=9.9917e-04) (hash(x)=64469482)
83
+ 670 train 5.179537 (lr=9.9906e-04) (hash(x)=76580403)
84
+ 680 train 4.828611 (lr=9.9894e-04) (hash(x)=60374603)
85
+ 690 train 4.978015 (lr=9.9882e-04) (hash(x)=69827901)
86
+ 700 val loss 5.1498
87
+ 700 val perplexity 172.4033
88
+ 700 train 5.172194 (lr=9.9870e-04) (hash(x)=77235911)
89
+ 710 train 5.141053 (lr=9.9856e-04) (hash(x)=76598106)
90
+ 720 train 5.205838 (lr=9.9842e-04) (hash(x)=75315013)
91
+ 730 train 5.049296 (lr=9.9828e-04) (hash(x)=76940932)
92
+ 740 train 5.056776 (lr=9.9812e-04) (hash(x)=75540179)
93
+ 750 train 5.034538 (lr=9.9796e-04) (hash(x)=73283995)
94
+ 760 train 5.066340 (lr=9.9780e-04) (hash(x)=84096065)
95
+ 770 train 4.845884 (lr=9.9762e-04) (hash(x)=75832016)
96
+ 780 train 5.015874 (lr=9.9744e-04) (hash(x)=65611269)
97
+ 790 train 5.068170 (lr=9.9726e-04) (hash(x)=72352412)
98
+ 800 val loss 4.9891
99
+ 800 val perplexity 146.8019
100
+ 800 train 4.976930 (lr=9.9707e-04) (hash(x)=78191559)
101
+ 810 train 4.909904 (lr=9.9687e-04) (hash(x)=73255448)
102
+ 820 train 5.004680 (lr=9.9666e-04) (hash(x)=75336936)
103
+ 830 train 4.806792 (lr=9.9645e-04) (hash(x)=71181581)
104
+ 840 train 5.452914 (lr=9.9623e-04) (hash(x)=76217477)
105
+ 850 train 4.890632 (lr=9.9601e-04) (hash(x)=75090324)
106
+ 860 train 4.833434 (lr=9.9578e-04) (hash(x)=71125083)
107
+ 870 train 4.843399 (lr=9.9554e-04) (hash(x)=74444188)
108
+ 880 train 4.875022 (lr=9.9530e-04) (hash(x)=72143748)
109
+ 890 train 4.736992 (lr=9.9505e-04) (hash(x)=75638384)
110
+ 900 val loss 4.8579
111
+ 900 val perplexity 128.7546
112
+ 900 train 4.889030 (lr=9.9479e-04) (hash(x)=76967190)
113
+ 910 train 4.831389 (lr=9.9453e-04) (hash(x)=84611316)
114
+ 920 train 4.618568 (lr=9.9426e-04) (hash(x)=61365145)
115
+ 930 train 4.599793 (lr=9.9398e-04) (hash(x)=54695617)
116
+ 940 train 4.798724 (lr=9.9370e-04) (hash(x)=76251415)
117
+ 950 train 4.576065 (lr=9.9341e-04) (hash(x)=71503060)
118
+ 960 train 4.776672 (lr=9.9311e-04) (hash(x)=70925251)
119
+ 970 train 4.712270 (lr=9.9281e-04) (hash(x)=69495443)
120
+ 980 train 4.673240 (lr=9.9250e-04) (hash(x)=74994053)
121
+ 990 train 4.513422 (lr=9.9219e-04) (hash(x)=81245808)
122
+ 1000 val loss 4.7651
123
+ 1000 val perplexity 117.3448
124
+ 1000 train 4.737851 (lr=9.9187e-04) (hash(x)=73801537)
125
+ 1010 train 4.671319 (lr=9.9154e-04) (hash(x)=78784929)
126
+ 1020 train 4.420188 (lr=9.9121e-04) (hash(x)=84649514)
127
+ 1030 train 4.553526 (lr=9.9087e-04) (hash(x)=71200051)
128
+ 1040 train 4.421772 (lr=9.9052e-04) (hash(x)=68279861)
129
+ 1050 train 4.477155 (lr=9.9017e-04) (hash(x)=86748930)
130
+ 1060 train 4.336933 (lr=9.8981e-04) (hash(x)=71319577)
131
+ 1070 train 4.471417 (lr=9.8944e-04) (hash(x)=72682588)
132
+ 1080 train 4.437065 (lr=9.8907e-04) (hash(x)=71408336)
133
+ 1090 train 4.631516 (lr=9.8869e-04) (hash(x)=78863558)
134
+ 1100 val loss 4.6624
135
+ 1100 val perplexity 105.8934
136
+ 1100 train 4.911688 (lr=9.8831e-04) (hash(x)=75285186)
137
+ 1110 train 4.705091 (lr=9.8791e-04) (hash(x)=78202715)
138
+ 1120 train 4.503892 (lr=9.8752e-04) (hash(x)=75187925)
139
+ 1130 train 4.696259 (lr=9.8711e-04) (hash(x)=70507475)
140
+ 1140 train 4.596849 (lr=9.8670e-04) (hash(x)=77911876)
141
+ 1150 train 4.524327 (lr=9.8629e-04) (hash(x)=77470391)
142
+ 1160 train 4.520265 (lr=9.8586e-04) (hash(x)=76266017)
143
+ 1170 train 4.806881 (lr=9.8543e-04) (hash(x)=77505346)
144
+ 1180 train 4.600049 (lr=9.8500e-04) (hash(x)=78978095)
145
+ 1190 train 4.656416 (lr=9.8456e-04) (hash(x)=78602083)
146
+ 1200 val loss 4.5877
147
+ 1200 val perplexity 98.2684
148
+ 1200 train 4.681677 (lr=9.8411e-04) (hash(x)=81700834)
149
+ 1210 train 4.565983 (lr=9.8365e-04) (hash(x)=75776079)
150
+ 1220 train 4.549867 (lr=9.8319e-04) (hash(x)=95908213)
151
+ 1230 train 4.471623 (lr=9.8272e-04) (hash(x)=84269607)
152
+ 1240 train 4.567751 (lr=9.8225e-04) (hash(x)=69460140)
153
+ 1250 train 4.573816 (lr=9.8177e-04) (hash(x)=71522595)
154
+ 1260 train 4.603982 (lr=9.8129e-04) (hash(x)=81155508)
155
+ 1270 train 4.737506 (lr=9.8079e-04) (hash(x)=72842244)
156
+ 1280 train 4.575050 (lr=9.8030e-04) (hash(x)=80034345)
157
+ 1290 train 4.428643 (lr=9.7979e-04) (hash(x)=70948511)
158
+ 1300 val loss 4.5141
159
+ 1300 val perplexity 91.2984
160
+ 1300 train 4.476729 (lr=9.7928e-04) (hash(x)=69667983)
161
+ 1310 train 4.517889 (lr=9.7876e-04) (hash(x)=84077906)
162
+ 1320 train 4.234534 (lr=9.7824e-04) (hash(x)=67481757)
163
+ 1330 train 4.403744 (lr=9.7771e-04) (hash(x)=83761654)
164
+ 1340 train 4.505623 (lr=9.7717e-04) (hash(x)=76630114)
165
+ 1350 train 4.533320 (lr=9.7663e-04) (hash(x)=81080698)
166
+ 1360 train 4.420274 (lr=9.7608e-04) (hash(x)=71019125)
167
+ 1370 train 4.410006 (lr=9.7553e-04) (hash(x)=78550770)
168
+ 1380 train 4.418458 (lr=9.7497e-04) (hash(x)=67017249)
169
+ 1390 train 4.331968 (lr=9.7440e-04) (hash(x)=75430624)
170
+ 1400 val loss 4.4904
171
+ 1400 val perplexity 89.1552
172
+ 1400 train 4.436020 (lr=9.7383e-04) (hash(x)=72720863)
173
+ 1410 train 4.178013 (lr=9.7325e-04) (hash(x)=73824272)
174
+ 1420 train 4.305560 (lr=9.7267e-04) (hash(x)=69424571)
175
+ 1430 train 4.268951 (lr=9.7207e-04) (hash(x)=71030131)
176
+ 1440 train 4.217274 (lr=9.7148e-04) (hash(x)=80088993)
177
+ 1450 train 4.168185 (lr=9.7087e-04) (hash(x)=74332002)
178
+ 1460 train 4.093929 (lr=9.7026e-04) (hash(x)=75132184)
179
+ 1470 train 4.149832 (lr=9.6965e-04) (hash(x)=72644858)
180
+ 1480 train 4.152047 (lr=9.6903e-04) (hash(x)=70677454)
181
+ 1490 train 4.173537 (lr=9.6840e-04) (hash(x)=74408093)
182
+ 1500 val loss 4.4340
183
+ 1500 val perplexity 84.2718
184
+ 1500 train 4.407459 (lr=9.6777e-04) (hash(x)=82777829)
185
+ 1510 train 4.494979 (lr=9.6713e-04) (hash(x)=82481742)
186
+ 1520 train 4.431778 (lr=9.6648e-04) (hash(x)=81580164)
187
+ 1530 train 4.399837 (lr=9.6583e-04) (hash(x)=75892741)
188
+ 1540 train 4.336584 (lr=9.6517e-04) (hash(x)=70574468)
189
+ 1550 train 4.377357 (lr=9.6451e-04) (hash(x)=77451919)
190
+ 1560 train 4.380383 (lr=9.6384e-04) (hash(x)=73148801)
191
+ 1570 train 4.290299 (lr=9.6316e-04) (hash(x)=74661859)
192
+ 1580 train 4.432652 (lr=9.6248e-04) (hash(x)=85484059)
193
+ 1590 train 4.553872 (lr=9.6179e-04) (hash(x)=72848788)
194
+ 1600 val loss 4.3532
195
+ 1600 val perplexity 77.7236
196
+ 1600 train 4.358281 (lr=9.6110e-04) (hash(x)=79468408)
197
+ 1610 train 4.365561 (lr=9.6040e-04) (hash(x)=78654805)
198
+ 1620 train 4.405781 (lr=9.5969e-04) (hash(x)=72348939)
199
+ 1630 train 4.464393 (lr=9.5898e-04) (hash(x)=74107545)
200
+ 1640 train 4.529867 (lr=9.5826e-04) (hash(x)=75470148)
201
+ 1650 train 4.414195 (lr=9.5754e-04) (hash(x)=74478093)
202
+ 1660 train 4.320990 (lr=9.5681e-04) (hash(x)=74077733)
203
+ 1670 train 4.218833 (lr=9.5607e-04) (hash(x)=60922881)
204
+ 1680 train 4.476503 (lr=9.5533e-04) (hash(x)=79401992)
205
+ 1690 train 4.411097 (lr=9.5458e-04) (hash(x)=75405453)
206
+ 1700 val loss 4.3546
207
+ 1700 val perplexity 77.8328
208
+ 1700 train 4.361378 (lr=9.5383e-04) (hash(x)=67378635)
209
+ 1710 train 4.272304 (lr=9.5307e-04) (hash(x)=74327780)
210
+ 1720 train 4.447429 (lr=9.5231e-04) (hash(x)=83288144)
211
+ 1730 train 4.192892 (lr=9.5153e-04) (hash(x)=68094391)
212
+ 1740 train 4.245912 (lr=9.5076e-04) (hash(x)=74943031)
213
+ 1750 train 4.289745 (lr=9.4998e-04) (hash(x)=77148592)
214
+ 1760 train 4.265447 (lr=9.4919e-04) (hash(x)=71352508)
215
+ 1770 train 4.281930 (lr=9.4839e-04) (hash(x)=73819276)
216
+ 1780 train 4.179723 (lr=9.4759e-04) (hash(x)=75338495)
217
+ 1790 train 4.386617 (lr=9.4679e-04) (hash(x)=100973274)
218
+ 1800 val loss 4.3184
219
+ 1800 val perplexity 75.0707
220
+ 1800 train 4.403097 (lr=9.4598e-04) (hash(x)=82399749)
221
+ 1810 train 4.419158 (lr=9.4516e-04) (hash(x)=80859117)
222
+ 1820 train 4.453051 (lr=9.4434e-04) (hash(x)=74215317)
223
+ 1830 train 4.312271 (lr=9.4351e-04) (hash(x)=73527133)
224
+ 1840 train 4.476528 (lr=9.4268e-04) (hash(x)=70049534)
225
+ 1850 train 4.195002 (lr=9.4184e-04) (hash(x)=71891328)
226
+ 1860 train 4.443893 (lr=9.4099e-04) (hash(x)=76433977)
227
+ 1870 train 4.414962 (lr=9.4014e-04) (hash(x)=69103302)
228
+ 1880 train 4.432602 (lr=9.3928e-04) (hash(x)=73328951)
229
+ 1890 train 4.188231 (lr=9.3842e-04) (hash(x)=77718295)
230
+ 1900 val loss 4.2715
231
+ 1900 val perplexity 71.6254
232
+ 1900 train 4.375031 (lr=9.3755e-04) (hash(x)=74708596)
233
+ 1910 train 4.330207 (lr=9.3668e-04) (hash(x)=80918764)
234
+ 1920 train 4.263230 (lr=9.3580e-04) (hash(x)=77248658)
235
+ 1930 train 4.364061 (lr=9.3491e-04) (hash(x)=73945895)
236
+ 1940 train 4.235381 (lr=9.3402e-04) (hash(x)=72252382)
237
+ 1950 train 4.291055 (lr=9.3313e-04) (hash(x)=87410072)
238
+ 1960 train 4.343136 (lr=9.3223e-04) (hash(x)=79408896)
239
+ 1970 train 4.129812 (lr=9.3132e-04) (hash(x)=78848496)
240
+ 1980 train 4.207821 (lr=9.3041e-04) (hash(x)=77601406)
241
+ 1990 train 4.165915 (lr=9.2949e-04) (hash(x)=71775143)
242
+ 2000 val loss 4.2275
243
+ 2000 val perplexity 68.5449
244
+ 2000 train 4.252492 (lr=9.2856e-04) (hash(x)=73721824)
245
+ 2010 train 4.213508 (lr=9.2763e-04) (hash(x)=74757670)
246
+ 2020 train 4.272916 (lr=9.2670e-04) (hash(x)=86234926)
247
+ 2030 train 4.457973 (lr=9.2576e-04) (hash(x)=69882050)
248
+ 2040 train 4.181327 (lr=9.2481e-04) (hash(x)=81004021)
249
+ 2050 train 4.073867 (lr=9.2386e-04) (hash(x)=75592124)
250
+ 2060 train 4.002054 (lr=9.2291e-04) (hash(x)=74811474)
251
+ 2070 train 4.194228 (lr=9.2195e-04) (hash(x)=76209571)
252
+ 2080 train 4.145471 (lr=9.2098e-04) (hash(x)=74973104)
253
+ 2090 train 4.198992 (lr=9.2001e-04) (hash(x)=74977086)
254
+ 2100 val loss 4.2534
255
+ 2100 val perplexity 70.3429
256
+ 2100 train 4.148748 (lr=9.1903e-04) (hash(x)=79579234)
257
+ 2110 train 4.098315 (lr=9.1804e-04) (hash(x)=73442407)
258
+ 2120 train 4.274135 (lr=9.1706e-04) (hash(x)=80281222)
259
+ 2130 train 4.217751 (lr=9.1606e-04) (hash(x)=83139613)
260
+ 2140 train 4.273276 (lr=9.1506e-04) (hash(x)=69936205)
261
+ 2150 train 4.316131 (lr=9.1406e-04) (hash(x)=74768225)
262
+ 2160 train 4.338952 (lr=9.1305e-04) (hash(x)=67473891)
263
+ 2170 train 4.290520 (lr=9.1203e-04) (hash(x)=80351641)
264
+ 2180 train 4.299978 (lr=9.1101e-04) (hash(x)=76303033)
265
+ 2190 train 4.231803 (lr=9.0999e-04) (hash(x)=73585511)
266
+ 2200 val loss 4.1864
267
+ 2200 val perplexity 65.7852
268
+ 2200 train 4.158675 (lr=9.0896e-04) (hash(x)=73195726)
269
+ 2210 train 4.311562 (lr=9.0792e-04) (hash(x)=80614640)
270
+ 2220 train 4.077424 (lr=9.0688e-04) (hash(x)=73107823)
271
+ 2230 train 4.153723 (lr=9.0583e-04) (hash(x)=78921188)
272
+ 2240 train 4.173172 (lr=9.0478e-04) (hash(x)=82527812)
273
+ 2250 train 4.088119 (lr=9.0372e-04) (hash(x)=75909041)
274
+ 2260 train 4.266368 (lr=9.0266e-04) (hash(x)=75266897)
275
+ 2270 train 4.214224 (lr=9.0160e-04) (hash(x)=74444372)
276
+ 2280 train 4.073836 (lr=9.0052e-04) (hash(x)=71643352)
277
+ 2290 train 4.251841 (lr=8.9945e-04) (hash(x)=73081122)
278
+ 2300 val loss 4.1653
279
+ 2300 val perplexity 64.4138
280
+ 2300 train 4.039475 (lr=8.9836e-04) (hash(x)=73133506)
281
+ 2310 train 4.099180 (lr=8.9728e-04) (hash(x)=73796905)
282
+ 2320 train 4.075785 (lr=8.9618e-04) (hash(x)=79133958)
283
+ 2330 train 4.096443 (lr=8.9509e-04) (hash(x)=79177833)
284
+ 2340 train 4.130397 (lr=8.9398e-04) (hash(x)=72790081)
285
+ 2350 train 4.110607 (lr=8.9288e-04) (hash(x)=65070947)
286
+ 2360 train 4.264915 (lr=8.9177e-04) (hash(x)=73178485)
287
+ 2370 train 4.219382 (lr=8.9065e-04) (hash(x)=71904548)
288
+ 2380 train 4.043080 (lr=8.8953e-04) (hash(x)=73416656)
289
+ 2390 train 4.184124 (lr=8.8840e-04) (hash(x)=71563647)
290
+ 2400 val loss 4.1734
291
+ 2400 val perplexity 64.9370
292
+ 2400 train 4.261228 (lr=8.8727e-04) (hash(x)=82236758)
293
+ 2410 train 4.063186 (lr=8.8613e-04) (hash(x)=76229570)
294
+ 2420 train 4.272271 (lr=8.8499e-04) (hash(x)=71056939)
295
+ 2430 train 4.122241 (lr=8.8384e-04) (hash(x)=76280968)
296
+ 2440 train 3.875177 (lr=8.8269e-04) (hash(x)=84199966)
297
+ 2450 train 4.299001 (lr=8.8153e-04) (hash(x)=84985998)
298
+ 2460 train 4.048522 (lr=8.8037e-04) (hash(x)=94074119)
299
+ 2470 train 4.247514 (lr=8.7921e-04) (hash(x)=81246200)
300
+ 2480 train 4.191588 (lr=8.7804e-04) (hash(x)=76225235)
301
+ 2490 train 4.119070 (lr=8.7686e-04) (hash(x)=76541445)
302
+ 2500 val loss 4.1162
303
+ 2500 val perplexity 61.3264
304
+ 2500 train 4.196284 (lr=8.7568e-04) (hash(x)=77708978)
305
+ 2510 train 4.066960 (lr=8.7450e-04) (hash(x)=75319253)
306
+ 2520 train 4.092705 (lr=8.7331e-04) (hash(x)=67190186)
307
+ 2530 train 4.271063 (lr=8.7211e-04) (hash(x)=82496813)
308
+ 2540 train 4.037810 (lr=8.7091e-04) (hash(x)=78100449)
309
+ 2550 train 4.103054 (lr=8.6971e-04) (hash(x)=80190873)
310
+ 2560 train 4.018284 (lr=8.6850e-04) (hash(x)=77711923)
311
+ 2570 train 4.132126 (lr=8.6729e-04) (hash(x)=79825674)
312
+ 2580 train 3.981089 (lr=8.6607e-04) (hash(x)=77024199)
313
+ 2590 train 4.059968 (lr=8.6485e-04) (hash(x)=72139372)
314
+ 2600 val loss 4.1028
315
+ 2600 val perplexity 60.5065
316
+ 2600 train 4.015438 (lr=8.6362e-04) (hash(x)=71362290)
317
+ 2610 train 4.152131 (lr=8.6239e-04) (hash(x)=80679759)
318
+ 2620 train 3.976839 (lr=8.6116e-04) (hash(x)=75079422)
319
+ 2630 train 4.085567 (lr=8.5991e-04) (hash(x)=70323867)
320
+ 2640 train 3.985546 (lr=8.5867e-04) (hash(x)=81444590)
321
+ 2650 train 4.009567 (lr=8.5742e-04) (hash(x)=71086644)
322
+ 2660 train 3.953842 (lr=8.5617e-04) (hash(x)=73339459)
323
+ 2670 train 4.129621 (lr=8.5491e-04) (hash(x)=74666974)
324
+ 2680 train 4.159144 (lr=8.5365e-04) (hash(x)=79117538)
325
+ 2690 train 4.154873 (lr=8.5238e-04) (hash(x)=65928715)
326
+ 2700 val loss 4.0972
327
+ 2700 val perplexity 60.1725
328
+ 2700 train 4.159644 (lr=8.5111e-04) (hash(x)=79416673)
329
+ 2710 train 4.004745 (lr=8.4983e-04) (hash(x)=65110265)
330
+ 2720 train 4.219120 (lr=8.4855e-04) (hash(x)=60942352)
331
+ 2730 train 4.182342 (lr=8.4727e-04) (hash(x)=68786492)
332
+ 2740 train 4.071064 (lr=8.4598e-04) (hash(x)=71480596)
333
+ 2750 train 4.041572 (lr=8.4469e-04) (hash(x)=84054551)
334
+ 2760 train 4.004358 (lr=8.4339e-04) (hash(x)=94608646)
335
+ 2770 train 4.083294 (lr=8.4209e-04) (hash(x)=73708388)
336
+ 2780 train 4.005244 (lr=8.4078e-04) (hash(x)=88054328)
337
+ 2790 train 4.172704 (lr=8.3947e-04) (hash(x)=88920089)
338
+ 2800 val loss 4.0637
339
+ 2800 val perplexity 58.1873
340
+ 2800 train 4.058393 (lr=8.3816e-04) (hash(x)=80878641)
341
+ 2810 train 4.101612 (lr=8.3684e-04) (hash(x)=71310778)
342
+ 2820 train 4.096463 (lr=8.3552e-04) (hash(x)=69986577)
343
+ 2830 train 4.073813 (lr=8.3419e-04) (hash(x)=78866280)
344
+ 2840 train 4.161679 (lr=8.3286e-04) (hash(x)=84135925)
345
+ 2850 train 4.157485 (lr=8.3153e-04) (hash(x)=74153539)
346
+ 2860 train 4.047184 (lr=8.3019e-04) (hash(x)=74749949)
347
+ 2870 train 4.132497 (lr=8.2884e-04) (hash(x)=72237368)
348
+ 2880 train 3.981327 (lr=8.2750e-04) (hash(x)=77118077)
349
+ 2890 train 4.082878 (lr=8.2615e-04) (hash(x)=73010418)
350
+ 2900 val loss 4.0431
351
+ 2900 val perplexity 57.0048
352
+ 2900 train 3.968037 (lr=8.2479e-04) (hash(x)=70624604)
353
+ 2910 train 3.979253 (lr=8.2343e-04) (hash(x)=72862034)
354
+ 2920 train 3.907590 (lr=8.2207e-04) (hash(x)=74542687)
355
+ 2930 train 4.108903 (lr=8.2070e-04) (hash(x)=71424119)
356
+ 2940 train 4.007574 (lr=8.1933e-04) (hash(x)=68581990)
357
+ 2950 train 4.033174 (lr=8.1796e-04) (hash(x)=75184497)
358
+ 2960 train 4.161138 (lr=8.1658e-04) (hash(x)=79104818)
359
+ 2970 train 3.949968 (lr=8.1520e-04) (hash(x)=75872591)
360
+ 2980 train 4.025938 (lr=8.1381e-04) (hash(x)=78565892)
361
+ 2990 train 3.973404 (lr=8.1242e-04) (hash(x)=77347407)
362
+ 3000 val loss 4.0543
363
+ 3000 val perplexity 57.6469
364
+ 3000 train 3.968811 (lr=8.1103e-04) (hash(x)=73568709)
365
+ 3010 train 3.954822 (lr=8.0963e-04) (hash(x)=83954777)
366
+ 3020 train 3.950898 (lr=8.0823e-04) (hash(x)=91325052)
367
+ 3030 train 4.182109 (lr=8.0682e-04) (hash(x)=73073624)
368
+ 3040 train 4.013264 (lr=8.0541e-04) (hash(x)=78263491)
369
+ 3050 train 4.059776 (lr=8.0400e-04) (hash(x)=75006046)
370
+ 3060 train 3.809413 (lr=8.0258e-04) (hash(x)=77283473)
371
+ 3070 train 4.167386 (lr=8.0116e-04) (hash(x)=70498330)
372
+ 3080 train 4.015658 (lr=7.9974e-04) (hash(x)=74632023)
373
+ 3090 train 4.016079 (lr=7.9831e-04) (hash(x)=71582163)
374
+ 3100 val loss 4.0139
375
+ 3100 val perplexity 55.3628
376
+ 3100 train 4.008207 (lr=7.9688e-04) (hash(x)=76032144)
377
+ 3110 train 4.085329 (lr=7.9545e-04) (hash(x)=79935326)
378
+ 3120 train 4.104363 (lr=7.9401e-04) (hash(x)=81668940)
379
+ 3130 train 4.029022 (lr=7.9257e-04) (hash(x)=81530706)
380
+ 3140 train 4.021668 (lr=7.9112e-04) (hash(x)=75958994)
381
+ 3150 train 4.108933 (lr=7.8967e-04) (hash(x)=70375358)
382
+ 3160 train 4.026141 (lr=7.8822e-04) (hash(x)=81742382)
383
+ 3170 train 4.008487 (lr=7.8677e-04) (hash(x)=73892129)
384
+ 3180 train 4.024062 (lr=7.8531e-04) (hash(x)=69584996)
385
+ 3190 train 4.003625 (lr=7.8384e-04) (hash(x)=74711675)
386
+ 3200 val loss 4.0076
387
+ 3200 val perplexity 55.0135
388
+ 3200 train 4.113000 (lr=7.8238e-04) (hash(x)=72943579)
389
+ 3210 train 4.072075 (lr=7.8091e-04) (hash(x)=76110656)
390
+ 3220 train 3.854102 (lr=7.7944e-04) (hash(x)=70203094)
391
+ 3230 train 4.005814 (lr=7.7796e-04) (hash(x)=60287090)
392
+ 3240 train 3.941430 (lr=7.7648e-04) (hash(x)=74302555)
393
+ 3250 train 4.046152 (lr=7.7500e-04) (hash(x)=72366702)
394
+ 3260 train 3.946336 (lr=7.7351e-04) (hash(x)=73014004)
395
+ 3270 train 3.832401 (lr=7.7203e-04) (hash(x)=75491051)
396
+ 3280 train 4.064177 (lr=7.7053e-04) (hash(x)=82821096)
397
+ 3290 train 3.889500 (lr=7.6904e-04) (hash(x)=77913090)
398
+ 3300 val loss 4.0103
399
+ 3300 val perplexity 55.1629
400
+ 3300 train 4.036943 (lr=7.6754e-04) (hash(x)=72779327)
401
+ 3310 train 3.965623 (lr=7.6604e-04) (hash(x)=76032908)
402
+ 3320 train 3.836689 (lr=7.6453e-04) (hash(x)=66875476)
403
+ 3330 train 3.896916 (lr=7.6303e-04) (hash(x)=67498762)
404
+ 3340 train 4.034355 (lr=7.6151e-04) (hash(x)=77920169)
405
+ 3350 train 4.027437 (lr=7.6000e-04) (hash(x)=87527839)
406
+ 3360 train 3.859725 (lr=7.5848e-04) (hash(x)=78010101)
407
+ 3370 train 4.070470 (lr=7.5696e-04) (hash(x)=76688335)
408
+ 3380 train 4.059451 (lr=7.5544e-04) (hash(x)=80832602)
409
+ 3390 train 3.905339 (lr=7.5391e-04) (hash(x)=75944307)
410
+ 3400 val loss 3.9887
411
+ 3400 val perplexity 53.9863
412
+ 3400 train 4.184787 (lr=7.5238e-04) (hash(x)=67516962)
413
+ 3410 train 4.042036 (lr=7.5085e-04) (hash(x)=72960118)
414
+ 3420 train 4.062222 (lr=7.4932e-04) (hash(x)=77405757)
415
+ 3430 train 4.091372 (lr=7.4778e-04) (hash(x)=76094785)
416
+ 3440 train 3.901554 (lr=7.4624e-04) (hash(x)=68336851)
417
+ 3450 train 3.893520 (lr=7.4470e-04) (hash(x)=71910579)
418
+ 3460 train 3.936821 (lr=7.4315e-04) (hash(x)=74170641)
419
+ 3470 train 4.033819 (lr=7.4160e-04) (hash(x)=75423851)
420
+ 3480 train 3.986952 (lr=7.4005e-04) (hash(x)=73367119)
421
+ 3490 train 3.853799 (lr=7.3849e-04) (hash(x)=75558326)
422
+ 3500 val loss 3.9811
423
+ 3500 val perplexity 53.5768
424
+ 3500 train 3.876295 (lr=7.3694e-04) (hash(x)=69187503)
425
+ 3510 train 3.986142 (lr=7.3538e-04) (hash(x)=80052583)
426
+ 3520 train 3.978205 (lr=7.3381e-04) (hash(x)=73819848)
427
+ 3530 train 3.798000 (lr=7.3225e-04) (hash(x)=71358103)
428
+ 3540 train 3.952610 (lr=7.3068e-04) (hash(x)=95551988)
429
+ 3550 train 4.004719 (lr=7.2911e-04) (hash(x)=75088121)
430
+ 3560 train 4.011243 (lr=7.2754e-04) (hash(x)=80473866)
431
+ 3570 train 4.026552 (lr=7.2596e-04) (hash(x)=73228665)
432
+ 3580 train 3.925741 (lr=7.2438e-04) (hash(x)=76321895)
433
+ 3590 train 3.849825 (lr=7.2280e-04) (hash(x)=74530382)
434
+ 3600 val loss 3.9675
435
+ 3600 val perplexity 52.8528
436
+ 3600 train 3.929077 (lr=7.2122e-04) (hash(x)=69686188)
437
+ 3610 train 3.960974 (lr=7.1963e-04) (hash(x)=66842873)
438
+ 3620 train 3.952906 (lr=7.1804e-04) (hash(x)=74921650)
439
+ 3630 train 3.850386 (lr=7.1645e-04) (hash(x)=79834656)
440
+ 3640 train 3.970330 (lr=7.1486e-04) (hash(x)=75122572)
441
+ 3650 train 3.903623 (lr=7.1326e-04) (hash(x)=77134480)
442
+ 3660 train 4.063825 (lr=7.1167e-04) (hash(x)=73027661)
443
+ 3670 train 4.118287 (lr=7.1007e-04) (hash(x)=85366911)
444
+ 3680 train 4.000724 (lr=7.0846e-04) (hash(x)=75873712)
445
+ 3690 train 4.092496 (lr=7.0686e-04) (hash(x)=81324159)
446
+ 3700 val loss 3.9540
447
+ 3700 val perplexity 52.1413
448
+ 3700 train 3.987675 (lr=7.0525e-04) (hash(x)=75354374)
449
+ 3710 train 4.039375 (lr=7.0364e-04) (hash(x)=73820036)
450
+ 3720 train 4.044855 (lr=7.0203e-04) (hash(x)=69279349)
451
+ 3730 train 4.011619 (lr=7.0041e-04) (hash(x)=76885056)
452
+ 3740 train 3.887223 (lr=6.9880e-04) (hash(x)=66083153)
453
+ 3750 train 3.886369 (lr=6.9718e-04) (hash(x)=82585096)
454
+ 3760 train 3.855583 (lr=6.9556e-04) (hash(x)=74734456)
455
+ 3770 train 3.844720 (lr=6.9394e-04) (hash(x)=74077986)
456
+ 3780 train 3.848018 (lr=6.9231e-04) (hash(x)=73794394)
457
+ 3790 train 3.875733 (lr=6.9069e-04) (hash(x)=75877136)
458
+ 3800 val loss 3.9414
459
+ 3800 val perplexity 51.4908
460
+ 3800 train 3.972227 (lr=6.8906e-04) (hash(x)=72150028)
461
+ 3810 train 4.061382 (lr=6.8743e-04) (hash(x)=77316418)
462
+ 3820 train 4.255692 (lr=6.8579e-04) (hash(x)=79951144)
463
+ 3830 train 3.941373 (lr=6.8416e-04) (hash(x)=72971714)
464
+ 3840 train 3.965837 (lr=6.8252e-04) (hash(x)=78168722)
465
+ 3850 train 3.798503 (lr=6.8088e-04) (hash(x)=72103764)
466
+ 3860 train 3.848412 (lr=6.7924e-04) (hash(x)=70399281)
467
+ 3870 train 4.032396 (lr=6.7760e-04) (hash(x)=71911595)
468
+ 3880 train 3.836545 (lr=6.7596e-04) (hash(x)=72034433)
469
+ 3890 train 3.831523 (lr=6.7431e-04) (hash(x)=73270524)
470
+ 3900 val loss 3.9355
471
+ 3900 val perplexity 51.1879
472
+ 3900 train 3.823839 (lr=6.7266e-04) (hash(x)=71934546)
473
+ 3910 train 3.987900 (lr=6.7101e-04) (hash(x)=78906659)
474
+ 3920 train 3.882963 (lr=6.6936e-04) (hash(x)=67852098)
475
+ 3930 train 3.845071 (lr=6.6771e-04) (hash(x)=75387628)
476
+ 3940 train 3.956375 (lr=6.6605e-04) (hash(x)=72407767)
477
+ 3950 train 3.814538 (lr=6.6440e-04) (hash(x)=66515238)
478
+ 3960 train 3.997642 (lr=6.6274e-04) (hash(x)=73337537)
479
+ 3970 train 3.994295 (lr=6.6108e-04) (hash(x)=74639467)
480
+ 3980 train 3.989024 (lr=6.5942e-04) (hash(x)=76044746)
481
+ 3990 train 4.003694 (lr=6.5776e-04) (hash(x)=71769001)
482
+ 4000 val loss 3.9121
483
+ 4000 val perplexity 50.0062
484
+ 4000 train 3.877454 (lr=6.5609e-04) (hash(x)=70257299)
485
+ 4010 train 3.928067 (lr=6.5443e-04) (hash(x)=71314416)
486
+ 4020 train 4.046541 (lr=6.5276e-04) (hash(x)=82315713)
487
+ 4030 train 3.980618 (lr=6.5109e-04) (hash(x)=65813944)
488
+ 4040 train 3.869294 (lr=6.4942e-04) (hash(x)=78277446)
489
+ 4050 train 3.924748 (lr=6.4775e-04) (hash(x)=77850258)
490
+ 4060 train 4.026044 (lr=6.4607e-04) (hash(x)=69511868)
491
+ 4070 train 3.850436 (lr=6.4440e-04) (hash(x)=77262761)
492
+ 4080 train 3.899817 (lr=6.4272e-04) (hash(x)=79691185)
493
+ 4090 train 3.718890 (lr=6.4104e-04) (hash(x)=74848196)
494
+ 4100 val loss 3.9099
495
+ 4100 val perplexity 49.8921
496
+ 4100 train 3.877767 (lr=6.3937e-04) (hash(x)=72761342)
497
+ 4110 train 3.972394 (lr=6.3769e-04) (hash(x)=58710996)
498
+ 4120 train 3.854548 (lr=6.3600e-04) (hash(x)=75794649)
499
+ 4130 train 3.973106 (lr=6.3432e-04) (hash(x)=72213336)
500
+ 4140 train 3.886581 (lr=6.3264e-04) (hash(x)=75842443)
501
+ 4150 train 3.799202 (lr=6.3095e-04) (hash(x)=70329458)
502
+ 4160 train 3.786905 (lr=6.2927e-04) (hash(x)=88536084)
503
+ 4170 train 3.945808 (lr=6.2758e-04) (hash(x)=76887790)
504
+ 4180 train 3.830776 (lr=6.2589e-04) (hash(x)=76755599)
505
+ 4190 train 3.837867 (lr=6.2420e-04) (hash(x)=71379783)
506
+ 4200 val loss 3.9094
507
+ 4200 val perplexity 49.8693
508
+ 4200 train 3.822284 (lr=6.2251e-04) (hash(x)=112576746)
509
+ 4210 train 3.733948 (lr=6.2082e-04) (hash(x)=77101027)
510
+ 4220 train 3.912746 (lr=6.1913e-04) (hash(x)=71607880)
511
+ 4230 train 3.823435 (lr=6.1743e-04) (hash(x)=72892070)
512
+ 4240 train 3.765296 (lr=6.1574e-04) (hash(x)=70967675)
513
+ 4250 train 3.862766 (lr=6.1404e-04) (hash(x)=75050831)
514
+ 4260 train 3.921880 (lr=6.1235e-04) (hash(x)=75249845)
515
+ 4270 train 3.913503 (lr=6.1065e-04) (hash(x)=82744946)
516
+ 4280 train 3.895804 (lr=6.0895e-04) (hash(x)=68857061)
517
+ 4290 train 3.884595 (lr=6.0725e-04) (hash(x)=80434125)
518
+ 4300 val loss 3.8900
519
+ 4300 val perplexity 48.9130
520
+ 4300 train 3.851652 (lr=6.0555e-04) (hash(x)=84624166)
521
+ 4310 train 3.781058 (lr=6.0385e-04) (hash(x)=60925156)
522
+ 4320 train 3.923411 (lr=6.0215e-04) (hash(x)=75956316)
523
+ 4330 train 3.841059 (lr=6.0044e-04) (hash(x)=74068056)
524
+ 4340 train 4.084041 (lr=5.9874e-04) (hash(x)=94977597)
525
+ 4350 train 3.969670 (lr=5.9704e-04) (hash(x)=71781944)
526
+ 4360 train 4.021977 (lr=5.9533e-04) (hash(x)=74308923)
527
+ 4370 train 3.867145 (lr=5.9363e-04) (hash(x)=76645666)
528
+ 4380 train 3.847078 (lr=5.9192e-04) (hash(x)=73326958)
529
+ 4390 train 3.835787 (lr=5.9022e-04) (hash(x)=74743749)
530
+ 4400 val loss 3.8791
531
+ 4400 val perplexity 48.3830
532
+ 4400 train 3.962578 (lr=5.8851e-04) (hash(x)=73116611)
533
+ 4410 train 3.736036 (lr=5.8680e-04) (hash(x)=68578568)
534
+ 4420 train 3.968826 (lr=5.8509e-04) (hash(x)=78085026)
535
+ 4430 train 3.853173 (lr=5.8338e-04) (hash(x)=71504765)
536
+ 4440 train 4.046264 (lr=5.8168e-04) (hash(x)=73163972)
537
+ 4450 train 3.864668 (lr=5.7997e-04) (hash(x)=81967324)
538
+ 4460 train 3.835377 (lr=5.7826e-04) (hash(x)=67796275)
539
+ 4470 train 3.865762 (lr=5.7655e-04) (hash(x)=75136849)
540
+ 4480 train 3.824352 (lr=5.7483e-04) (hash(x)=80109861)
541
+ 4490 train 3.750341 (lr=5.7312e-04) (hash(x)=72094345)
542
+ 4500 val loss 3.8688
543
+ 4500 val perplexity 47.8852
544
+ 4500 train 3.763227 (lr=5.7141e-04) (hash(x)=77660356)
545
+ 4510 train 3.744982 (lr=5.6970e-04) (hash(x)=71303771)
546
+ 4520 train 3.958481 (lr=5.6799e-04) (hash(x)=67353213)
547
+ 4530 train 3.589906 (lr=5.6628e-04) (hash(x)=91984280)
548
+ 4540 train 3.740797 (lr=5.6456e-04) (hash(x)=74918518)
549
+ 4550 train 3.839872 (lr=5.6285e-04) (hash(x)=80129084)
550
+ 4560 train 3.768359 (lr=5.6114e-04) (hash(x)=74762343)
551
+ 4570 train 3.735575 (lr=5.5942e-04) (hash(x)=83264256)
552
+ 4580 train 3.844501 (lr=5.5771e-04) (hash(x)=77701601)
553
+ 4590 train 3.778143 (lr=5.5600e-04) (hash(x)=72730921)
554
+ 4600 val loss 3.8617
555
+ 4600 val perplexity 47.5478
556
+ 4600 train 3.828333 (lr=5.5428e-04) (hash(x)=77801190)
557
+ 4610 train 3.899195 (lr=5.5257e-04) (hash(x)=64563241)
558
+ 4620 train 3.833240 (lr=5.5086e-04) (hash(x)=75462951)
559
+ 4630 train 3.881160 (lr=5.4914e-04) (hash(x)=74708327)
560
+ 4640 train 3.785066 (lr=5.4743e-04) (hash(x)=83297632)
561
+ 4650 train 3.894285 (lr=5.4572e-04) (hash(x)=81914758)
562
+ 4660 train 3.915007 (lr=5.4400e-04) (hash(x)=71367268)
563
+ 4670 train 3.795990 (lr=5.4229e-04) (hash(x)=73587555)
564
+ 4680 train 3.910823 (lr=5.4058e-04) (hash(x)=74734602)
565
+ 4690 train 4.010897 (lr=5.3886e-04) (hash(x)=77508704)
566
+ 4700 val loss 3.8466
567
+ 4700 val perplexity 46.8340
568
+ 4700 train 3.811380 (lr=5.3715e-04) (hash(x)=73228682)
569
+ 4710 train 3.835506 (lr=5.3544e-04) (hash(x)=73851956)
570
+ 4720 train 3.883692 (lr=5.3372e-04) (hash(x)=71244190)
571
+ 4730 train 3.646925 (lr=5.3201e-04) (hash(x)=74416580)
572
+ 4740 train 3.879030 (lr=5.3030e-04) (hash(x)=69259353)
573
+ 4750 train 3.850823 (lr=5.2859e-04) (hash(x)=77662887)
574
+ 4760 train 3.792641 (lr=5.2688e-04) (hash(x)=73001257)
575
+ 4770 train 3.744576 (lr=5.2517e-04) (hash(x)=81466432)
576
+ 4780 train 3.687579 (lr=5.2345e-04) (hash(x)=76710780)
577
+ 4790 train 3.838884 (lr=5.2174e-04) (hash(x)=72440698)
578
+ 4800 val loss 3.8486
579
+ 4800 val perplexity 46.9251
580
+ 4800 train 3.720203 (lr=5.2003e-04) (hash(x)=75002116)
581
+ 4810 train 3.776699 (lr=5.1832e-04) (hash(x)=84873693)
582
+ 4820 train 3.623518 (lr=5.1662e-04) (hash(x)=68563264)
583
+ 4830 train 3.741002 (lr=5.1491e-04) (hash(x)=76811982)
584
+ 4840 train 3.655573 (lr=5.1320e-04) (hash(x)=73763746)
585
+ 4850 train 3.868413 (lr=5.1149e-04) (hash(x)=72834169)
586
+ 4860 train 3.808540 (lr=5.0978e-04) (hash(x)=71420794)
587
+ 4870 train 3.927446 (lr=5.0808e-04) (hash(x)=82083784)
588
+ 4880 train 3.902657 (lr=5.0637e-04) (hash(x)=71399278)
589
+ 4890 train 3.962420 (lr=5.0467e-04) (hash(x)=80791916)
590
+ 4900 val loss 3.8318
591
+ 4900 val perplexity 46.1453
592
+ 4900 train 3.848641 (lr=5.0296e-04) (hash(x)=76275522)
593
+ 4910 train 4.007618 (lr=5.0126e-04) (hash(x)=73846113)
594
+ 4920 train 3.935615 (lr=4.9956e-04) (hash(x)=65906497)
595
+ 4930 train 3.739518 (lr=4.9785e-04) (hash(x)=74564994)
596
+ 4940 train 4.004200 (lr=4.9615e-04) (hash(x)=78282849)
597
+ 4950 train 3.815697 (lr=4.9445e-04) (hash(x)=75618079)
598
+ 4960 train 3.959253 (lr=4.9275e-04) (hash(x)=78308363)
599
+ 4970 train 3.820936 (lr=4.9105e-04) (hash(x)=81122810)
600
+ 4980 train 3.962206 (lr=4.8935e-04) (hash(x)=72474525)
601
+ 4990 train 3.820616 (lr=4.8765e-04) (hash(x)=71682033)
602
+ 5000 val loss 3.8143
603
+ 5000 val perplexity 45.3471
604
+ 5000 train 3.771644 (lr=4.8596e-04) (hash(x)=74751266)
605
+ 5010 train 3.841506 (lr=4.8426e-04) (hash(x)=71876255)
606
+ 5020 train 4.012558 (lr=4.8257e-04) (hash(x)=67006293)
607
+ 5030 train 3.821250 (lr=4.8087e-04) (hash(x)=55446837)
608
+ 5040 train 3.931035 (lr=4.7918e-04) (hash(x)=72139304)
609
+ 5050 train 3.912959 (lr=4.7749e-04) (hash(x)=72848819)
610
+ 5060 train 3.840793 (lr=4.7580e-04) (hash(x)=74756891)
611
+ 5070 train 3.826256 (lr=4.7411e-04) (hash(x)=71425192)
612
+ 5080 train 3.704717 (lr=4.7242e-04) (hash(x)=91717336)
613
+ 5090 train 3.711437 (lr=4.7073e-04) (hash(x)=78489482)
614
+ 5100 val loss 3.8154
615
+ 5100 val perplexity 45.3969
616
+ 5100 train 3.703903 (lr=4.6905e-04) (hash(x)=102417092)
617
+ 5110 train 3.773282 (lr=4.6736e-04) (hash(x)=67631270)
618
+ 5120 train 3.762649 (lr=4.6568e-04) (hash(x)=70575439)
619
+ 5130 train 3.797820 (lr=4.6400e-04) (hash(x)=73933044)
620
+ 5140 train 3.664560 (lr=4.6231e-04) (hash(x)=75105935)
621
+ 5150 train 3.784578 (lr=4.6063e-04) (hash(x)=75662761)
622
+ 5160 train 3.784907 (lr=4.5896e-04) (hash(x)=77254705)
623
+ 5170 train 3.750488 (lr=4.5728e-04) (hash(x)=67051450)
624
+ 5180 train 3.575444 (lr=4.5560e-04) (hash(x)=69736129)
625
+ 5190 train 3.590010 (lr=4.5393e-04) (hash(x)=68256740)
626
+ 5200 val loss 3.8137
627
+ 5200 val perplexity 45.3177
628
+ 5200 train 3.551154 (lr=4.5225e-04) (hash(x)=69483022)
629
+ 5210 train 3.607293 (lr=4.5058e-04) (hash(x)=74805034)
630
+ 5220 train 3.584081 (lr=4.4891e-04) (hash(x)=75297212)
631
+ 5230 train 3.508024 (lr=4.4724e-04) (hash(x)=69772814)
632
+ 5240 train 3.622215 (lr=4.4557e-04) (hash(x)=72164549)
633
+ 5250 train 3.484519 (lr=4.4391e-04) (hash(x)=73896592)
634
+ 5260 train 3.629845 (lr=4.4224e-04) (hash(x)=75901314)
635
+ 5270 train 3.504444 (lr=4.4058e-04) (hash(x)=76054156)
636
+ 5280 train 3.843983 (lr=4.3892e-04) (hash(x)=85568643)
637
+ 5290 train 3.884977 (lr=4.3726e-04) (hash(x)=84053592)
638
+ 5300 val loss 3.7978
639
+ 5300 val perplexity 44.6012
640
+ 5300 train 3.767685 (lr=4.3560e-04) (hash(x)=73135718)
641
+ 5310 train 3.913254 (lr=4.3395e-04) (hash(x)=73149116)
642
+ 5320 train 3.891977 (lr=4.3229e-04) (hash(x)=77101540)
643
+ 5330 train 3.725110 (lr=4.3064e-04) (hash(x)=84274756)
644
+ 5340 train 3.748969 (lr=4.2899e-04) (hash(x)=74181619)
645
+ 5350 train 3.633959 (lr=4.2734e-04) (hash(x)=69664265)
646
+ 5360 train 3.860588 (lr=4.2569e-04) (hash(x)=68837592)
647
+ 5370 train 3.795830 (lr=4.2404e-04) (hash(x)=75148213)
648
+ 5380 train 3.886878 (lr=4.2240e-04) (hash(x)=68908917)
649
+ 5390 train 3.852165 (lr=4.2076e-04) (hash(x)=72753463)
650
+ 5400 val loss 3.7794
651
+ 5400 val perplexity 43.7889
652
+ 5400 train 3.749935 (lr=4.1912e-04) (hash(x)=69948122)
653
+ 5410 train 3.655521 (lr=4.1748e-04) (hash(x)=74530436)
654
+ 5420 train 3.821488 (lr=4.1584e-04) (hash(x)=67532641)
655
+ 5430 train 3.792599 (lr=4.1421e-04) (hash(x)=77416323)
656
+ 5440 train 3.771484 (lr=4.1257e-04) (hash(x)=77450277)
657
+ 5450 train 3.699586 (lr=4.1094e-04) (hash(x)=69637629)
658
+ 5460 train 3.750093 (lr=4.0931e-04) (hash(x)=77134731)
659
+ 5470 train 3.882771 (lr=4.0769e-04) (hash(x)=72637484)
660
+ 5480 train 3.819701 (lr=4.0606e-04) (hash(x)=76342282)
661
+ 5490 train 3.741092 (lr=4.0444e-04) (hash(x)=74794888)
662
+ 5500 val loss 3.7783
663
+ 5500 val perplexity 43.7397
664
+ 5500 train 3.739415 (lr=4.0282e-04) (hash(x)=70491787)
665
+ 5510 train 3.765331 (lr=4.0120e-04) (hash(x)=73673009)
666
+ 5520 train 3.698666 (lr=3.9959e-04) (hash(x)=73134241)
667
+ 5530 train 3.702623 (lr=3.9797e-04) (hash(x)=80260202)
668
+ 5540 train 3.737857 (lr=3.9636e-04) (hash(x)=75403578)
669
+ 5550 train 3.668324 (lr=3.9475e-04) (hash(x)=65589999)
670
+ 5560 train 3.749234 (lr=3.9314e-04) (hash(x)=71523917)
671
+ 5570 train 3.648072 (lr=3.9154e-04) (hash(x)=71871396)
672
+ 5580 train 3.463234 (lr=3.8993e-04) (hash(x)=77446225)
673
+ 5590 train 3.477172 (lr=3.8833e-04) (hash(x)=79892116)
674
+ 5600 val loss 3.7838
675
+ 5600 val perplexity 43.9822
676
+ 5600 train 3.696796 (lr=3.8674e-04) (hash(x)=72893174)
677
+ 5610 train 3.527065 (lr=3.8514e-04) (hash(x)=74362751)
678
+ 5620 train 3.661416 (lr=3.8355e-04) (hash(x)=75125741)
679
+ 5630 train 3.541162 (lr=3.8196e-04) (hash(x)=74303205)
680
+ 5640 train 3.547629 (lr=3.8037e-04) (hash(x)=65727805)
681
+ 5650 train 3.631219 (lr=3.7878e-04) (hash(x)=79694849)
682
+ 5660 train 3.649570 (lr=3.7720e-04) (hash(x)=74705862)
683
+ 5670 train 3.623400 (lr=3.7562e-04) (hash(x)=74962821)
684
+ 5680 train 3.995806 (lr=3.7404e-04) (hash(x)=76748843)
685
+ 5690 train 3.842822 (lr=3.7246e-04) (hash(x)=72562868)
686
+ 5700 val loss 3.7744
687
+ 5700 val perplexity 43.5693
688
+ 5700 train 3.787758 (lr=3.7089e-04) (hash(x)=83245544)
689
+ 5710 train 3.710258 (lr=3.6932e-04) (hash(x)=76485320)
690
+ 5720 train 3.824130 (lr=3.6775e-04) (hash(x)=75340650)
691
+ 5730 train 3.806679 (lr=3.6619e-04) (hash(x)=73202272)
692
+ 5740 train 3.782930 (lr=3.6462e-04) (hash(x)=75623558)
693
+ 5750 train 3.802449 (lr=3.6306e-04) (hash(x)=73893861)
694
+ 5760 train 3.771797 (lr=3.6151e-04) (hash(x)=78314196)
695
+ 5770 train 3.790672 (lr=3.5995e-04) (hash(x)=72732120)
696
+ 5780 train 3.768619 (lr=3.5840e-04) (hash(x)=79236217)
697
+ 5790 train 3.752296 (lr=3.5685e-04) (hash(x)=82223910)
698
+ 5800 val loss 3.7545
699
+ 5800 val perplexity 42.7123
700
+ 5800 train 3.766829 (lr=3.5530e-04) (hash(x)=71077079)
701
+ 5810 train 3.770199 (lr=3.5376e-04) (hash(x)=82793170)
702
+ 5820 train 3.768299 (lr=3.5222e-04) (hash(x)=100985291)
703
+ 5830 train 3.769893 (lr=3.5068e-04) (hash(x)=74375677)
704
+ 5840 train 3.903067 (lr=3.4915e-04) (hash(x)=71221972)
705
+ 5850 train 3.682836 (lr=3.4762e-04) (hash(x)=72588583)
706
+ 5860 train 3.768000 (lr=3.4609e-04) (hash(x)=88719929)
707
+ 5870 train 3.739300 (lr=3.4456e-04) (hash(x)=73369246)
708
+ 5880 train 3.800587 (lr=3.4304e-04) (hash(x)=73331315)
709
+ 5890 train 3.766808 (lr=3.4152e-04) (hash(x)=53449512)
710
+ 5900 val loss 3.7572
711
+ 5900 val perplexity 42.8272
712
+ 5900 train 3.655138 (lr=3.4000e-04) (hash(x)=75334660)
713
+ 5910 train 3.723900 (lr=3.3849e-04) (hash(x)=75037764)
714
+ 5920 train 3.792995 (lr=3.3697e-04) (hash(x)=72040912)
715
+ 5930 train 3.721459 (lr=3.3547e-04) (hash(x)=68296721)
716
+ 5940 train 3.685977 (lr=3.3396e-04) (hash(x)=72575829)
717
+ 5950 train 3.568504 (lr=3.3246e-04) (hash(x)=71274235)
718
+ 5960 train 3.753669 (lr=3.3096e-04) (hash(x)=70377631)
719
+ 5970 train 3.586278 (lr=3.2947e-04) (hash(x)=80434203)
720
+ 5980 train 3.455035 (lr=3.2797e-04) (hash(x)=71681516)
721
+ 5990 train 3.576262 (lr=3.2649e-04) (hash(x)=74917317)
722
+ 6000 val loss 3.7560
723
+ 6000 val perplexity 42.7777
724
+ 6000 train 3.502906 (lr=3.2500e-04) (hash(x)=69246326)
725
+ 6010 train 3.594422 (lr=3.2352e-04) (hash(x)=76209797)
726
+ 6020 train 3.700918 (lr=3.2204e-04) (hash(x)=77523561)
727
+ 6030 train 3.671709 (lr=3.2056e-04) (hash(x)=77066537)
728
+ 6040 train 3.640259 (lr=3.1909e-04) (hash(x)=81795462)
729
+ 6050 train 3.545828 (lr=3.1762e-04) (hash(x)=74328005)
730
+ 6060 train 3.560942 (lr=3.1616e-04) (hash(x)=76260120)
731
+ 6070 train 3.688489 (lr=3.1469e-04) (hash(x)=73864904)
732
+ 6080 train 3.660241 (lr=3.1323e-04) (hash(x)=83311739)
733
+ 6090 train 3.964492 (lr=3.1178e-04) (hash(x)=85943188)
734
+ 6100 val loss 3.7370
735
+ 6100 val perplexity 41.9723
736
+ 6100 train 3.626211 (lr=3.1033e-04) (hash(x)=70412657)
737
+ 6110 train 3.785763 (lr=3.0888e-04) (hash(x)=78569898)
738
+ 6120 train 3.839788 (lr=3.0743e-04) (hash(x)=53010374)
739
+ 6130 train 3.834085 (lr=3.0599e-04) (hash(x)=77387166)
740
+ 6140 train 3.817585 (lr=3.0455e-04) (hash(x)=68495380)
741
+ 6150 train 3.702678 (lr=3.0312e-04) (hash(x)=77674661)
742
+ 6160 train 3.782394 (lr=3.0169e-04) (hash(x)=73099224)
743
+ 6170 train 3.783549 (lr=3.0026e-04) (hash(x)=75550331)
744
+ 6180 train 3.852736 (lr=2.9884e-04) (hash(x)=70208558)
745
+ 6190 train 3.695030 (lr=2.9742e-04) (hash(x)=81474783)
746
+ 6200 val loss 3.7208
747
+ 6200 val perplexity 41.2980
748
+ 6200 train 3.688138 (lr=2.9600e-04) (hash(x)=72863420)
749
+ 6210 train 3.823990 (lr=2.9459e-04) (hash(x)=67338127)
750
+ 6220 train 3.749213 (lr=2.9318e-04) (hash(x)=72064676)
751
+ 6230 train 3.628724 (lr=2.9177e-04) (hash(x)=63929148)
752
+ 6240 train 3.661108 (lr=2.9037e-04) (hash(x)=75244700)
753
+ 6250 train 3.760086 (lr=2.8897e-04) (hash(x)=72131685)
754
+ 6260 train 3.820707 (lr=2.8758e-04) (hash(x)=77239186)
755
+ 6270 train 3.729449 (lr=2.8619e-04) (hash(x)=80827392)
756
+ 6280 train 3.692749 (lr=2.8480e-04) (hash(x)=68612370)
757
+ 6290 train 3.747125 (lr=2.8342e-04) (hash(x)=70664463)
758
+ 6300 val loss 3.7183
759
+ 6300 val perplexity 41.1954
760
+ 6300 train 3.686190 (lr=2.8204e-04) (hash(x)=73020224)
761
+ 6310 train 3.703270 (lr=2.8067e-04) (hash(x)=70614977)
762
+ 6320 train 3.693339 (lr=2.7930e-04) (hash(x)=76744877)
763
+ 6330 train 3.741690 (lr=2.7793e-04) (hash(x)=83384143)
764
+ 6340 train 3.599163 (lr=2.7657e-04) (hash(x)=72241681)
765
+ 6350 train 3.624433 (lr=2.7521e-04) (hash(x)=74037360)
766
+ 6360 train 3.471763 (lr=2.7385e-04) (hash(x)=70095163)
767
+ 6370 train 3.394009 (lr=2.7250e-04) (hash(x)=69510647)
768
+ 6380 train 3.567859 (lr=2.7116e-04) (hash(x)=74297517)
769
+ 6390 train 3.702486 (lr=2.6981e-04) (hash(x)=78058972)
770
+ 6400 val loss 3.7210
771
+ 6400 val perplexity 41.3055
772
+ 6400 train 3.615313 (lr=2.6847e-04) (hash(x)=82953514)
773
+ 6410 train 3.388636 (lr=2.6714e-04) (hash(x)=72289413)
774
+ 6420 train 3.525731 (lr=2.6581e-04) (hash(x)=66436985)
775
+ 6430 train 3.633868 (lr=2.6448e-04) (hash(x)=76842704)
776
+ 6440 train 3.581012 (lr=2.6316e-04) (hash(x)=77576509)
777
+ 6450 train 3.593360 (lr=2.6184e-04) (hash(x)=63757462)
778
+ 6460 train 3.353406 (lr=2.6053e-04) (hash(x)=74725544)
779
+ 6470 train 3.806491 (lr=2.5922e-04) (hash(x)=74976987)
780
+ 6480 train 3.939163 (lr=2.5791e-04) (hash(x)=68415455)
781
+ 6490 train 3.739840 (lr=2.5661e-04) (hash(x)=75902883)
782
+ 6500 val loss 3.7085
783
+ 6500 val perplexity 40.7919
784
+ 6500 train 3.678328 (lr=2.5531e-04) (hash(x)=76888442)
785
+ 6510 train 3.780519 (lr=2.5402e-04) (hash(x)=73767111)
786
+ 6520 train 3.782838 (lr=2.5273e-04) (hash(x)=76218061)
787
+ 6530 train 3.741245 (lr=2.5145e-04) (hash(x)=76573241)
788
+ 6540 train 3.724410 (lr=2.5017e-04) (hash(x)=71980504)
789
+ 6550 train 3.782207 (lr=2.4889e-04) (hash(x)=77715451)
790
+ 6560 train 3.660448 (lr=2.4762e-04) (hash(x)=73772644)
791
+ 6570 train 3.727784 (lr=2.4635e-04) (hash(x)=70907139)
792
+ 6580 train 3.736925 (lr=2.4509e-04) (hash(x)=68045832)
793
+ 6590 train 3.753581 (lr=2.4383e-04) (hash(x)=75212157)
794
+ 6600 val loss 3.7036
795
+ 6600 val perplexity 40.5946
796
+ 6600 train 3.692216 (lr=2.4258e-04) (hash(x)=69232309)
797
+ 6610 train 3.675817 (lr=2.4133e-04) (hash(x)=67915539)
798
+ 6620 train 3.538253 (lr=2.4009e-04) (hash(x)=73942199)
799
+ 6630 train 3.711311 (lr=2.3884e-04) (hash(x)=71108653)
800
+ 6640 train 3.702012 (lr=2.3761e-04) (hash(x)=70766590)
801
+ 6650 train 3.586818 (lr=2.3638e-04) (hash(x)=69205319)
802
+ 6660 train 3.693664 (lr=2.3515e-04) (hash(x)=81026148)
803
+ 6670 train 3.621096 (lr=2.3393e-04) (hash(x)=77657830)
804
+ 6680 train 3.530802 (lr=2.3271e-04) (hash(x)=76483678)
805
+ 6690 train 3.585719 (lr=2.3150e-04) (hash(x)=71151306)
806
+ 6700 val loss 3.7037
807
+ 6700 val perplexity 40.5980
808
+ 6700 train 3.682486 (lr=2.3029e-04) (hash(x)=74577530)
809
+ 6710 train 3.573588 (lr=2.2909e-04) (hash(x)=69171643)
810
+ 6720 train 3.615970 (lr=2.2789e-04) (hash(x)=67250925)
811
+ 6730 train 3.598993 (lr=2.2669e-04) (hash(x)=68971618)
812
+ 6740 train 3.557070 (lr=2.2550e-04) (hash(x)=75990195)
813
+ 6750 train 3.654644 (lr=2.2432e-04) (hash(x)=84198679)
814
+ 6760 train 3.735959 (lr=2.2314e-04) (hash(x)=100211939)
815
+ 6770 train 3.758752 (lr=2.2196e-04) (hash(x)=70028335)
816
+ 6780 train 3.821722 (lr=2.2079e-04) (hash(x)=76176743)
817
+ 6790 train 3.902557 (lr=2.1963e-04) (hash(x)=75642962)
818
+ 6800 val loss 3.6879
819
+ 6800 val perplexity 39.9615
820
+ 6800 train 3.631905 (lr=2.1847e-04) (hash(x)=73064800)
821
+ 6810 train 3.742045 (lr=2.1731e-04) (hash(x)=71307839)
822
+ 6820 train 3.662904 (lr=2.1616e-04) (hash(x)=80361526)
823
+ 6830 train 3.678871 (lr=2.1501e-04) (hash(x)=74155886)
824
+ 6840 train 3.762501 (lr=2.1387e-04) (hash(x)=83498789)
825
+ 6850 train 3.872828 (lr=2.1273e-04) (hash(x)=72686972)
826
+ 6860 train 3.821345 (lr=2.1160e-04) (hash(x)=74835805)
827
+ 6870 train 3.755506 (lr=2.1047e-04) (hash(x)=74809435)
828
+ 6880 train 3.814886 (lr=2.0935e-04) (hash(x)=71911947)
829
+ 6890 train 3.689610 (lr=2.0823e-04) (hash(x)=70201693)
830
+ 6900 val loss 3.6828
831
+ 6900 val perplexity 39.7581
832
+ 6900 train 3.712472 (lr=2.0712e-04) (hash(x)=77162446)
833
+ 6910 train 3.835165 (lr=2.0602e-04) (hash(x)=60395458)
834
+ 6920 train 3.740775 (lr=2.0491e-04) (hash(x)=78466934)
835
+ 6930 train 3.723298 (lr=2.0382e-04) (hash(x)=80996788)
836
+ 6940 train 3.615596 (lr=2.0272e-04) (hash(x)=74634433)
837
+ 6950 train 3.737562 (lr=2.0164e-04) (hash(x)=69618707)
838
+ 6960 train 3.601078 (lr=2.0055e-04) (hash(x)=77936797)
839
+ 6970 train 3.612643 (lr=1.9948e-04) (hash(x)=78698749)
840
+ 6980 train 3.609472 (lr=1.9840e-04) (hash(x)=71152904)
841
+ 6990 train 3.720703 (lr=1.9734e-04) (hash(x)=78876720)
842
+ 7000 val loss 3.6782
843
+ 7000 val perplexity 39.5765
844
+ 7000 train 3.698404 (lr=1.9628e-04) (hash(x)=70500963)
845
+ 7010 train 3.661447 (lr=1.9522e-04) (hash(x)=75732988)
846
+ 7020 train 3.834758 (lr=1.9417e-04) (hash(x)=72655885)
847
+ 7030 train 3.541782 (lr=1.9312e-04) (hash(x)=75132388)
848
+ 7040 train 3.611836 (lr=1.9208e-04) (hash(x)=88674279)
849
+ 7050 train 3.591294 (lr=1.9104e-04) (hash(x)=73050510)
850
+ 7060 train 3.629106 (lr=1.9001e-04) (hash(x)=74850157)
851
+ 7070 train 3.623950 (lr=1.8899e-04) (hash(x)=80733840)
852
+ 7080 train 3.614174 (lr=1.8797e-04) (hash(x)=68592323)
853
+ 7090 train 3.550707 (lr=1.8695e-04) (hash(x)=74142615)
854
+ 7100 val loss 3.6780
855
+ 7100 val perplexity 39.5670
856
+ 7100 train 3.566196 (lr=1.8594e-04) (hash(x)=78220018)
857
+ 7110 train 3.559711 (lr=1.8494e-04) (hash(x)=68388633)
858
+ 7120 train 3.380956 (lr=1.8394e-04) (hash(x)=71391877)
859
+ 7130 train 3.695735 (lr=1.8294e-04) (hash(x)=80144529)
860
+ 7140 train 3.789128 (lr=1.8196e-04) (hash(x)=79507532)
861
+ 7150 train 3.692015 (lr=1.8097e-04) (hash(x)=76982262)
862
+ 7160 train 3.855980 (lr=1.7999e-04) (hash(x)=82516157)
863
+ 7170 train 3.609463 (lr=1.7902e-04) (hash(x)=72371860)
864
+ 7180 train 3.862401 (lr=1.7805e-04) (hash(x)=72376921)
865
+ 7190 train 3.708340 (lr=1.7709e-04) (hash(x)=81008418)
866
+ 7200 val loss 3.6642
867
+ 7200 val perplexity 39.0257
868
+ 7200 train 3.771976 (lr=1.7614e-04) (hash(x)=73818590)
869
+ 7210 train 3.738631 (lr=1.7519e-04) (hash(x)=71683916)
870
+ 7220 train 3.691939 (lr=1.7424e-04) (hash(x)=77659620)
871
+ 7230 train 3.732988 (lr=1.7330e-04) (hash(x)=79997614)
872
+ 7240 train 3.748168 (lr=1.7237e-04) (hash(x)=74054734)
873
+ 7250 train 3.839936 (lr=1.7144e-04) (hash(x)=84630195)
874
+ 7260 train 3.742384 (lr=1.7051e-04) (hash(x)=83833114)
875
+ 7270 train 3.659254 (lr=1.6959e-04) (hash(x)=78625868)
876
+ 7280 train 3.868357 (lr=1.6868e-04) (hash(x)=76859769)
877
+ 7290 train 3.667498 (lr=1.6777e-04) (hash(x)=77079387)
878
+ 7300 val loss 3.6592
879
+ 7300 val perplexity 38.8299
880
+ 7300 train 3.661269 (lr=1.6687e-04) (hash(x)=74181120)
881
+ 7310 train 3.730875 (lr=1.6598e-04) (hash(x)=72220366)
882
+ 7320 train 3.669443 (lr=1.6509e-04) (hash(x)=77955423)
883
+ 7330 train 3.709283 (lr=1.6420e-04) (hash(x)=75397467)
884
+ 7340 train 3.703813 (lr=1.6332e-04) (hash(x)=73697940)
885
+ 7350 train 3.603831 (lr=1.6245e-04) (hash(x)=83531555)
886
+ 7360 train 3.717419 (lr=1.6158e-04) (hash(x)=72009097)
887
+ 7370 train 3.666575 (lr=1.6072e-04) (hash(x)=76354441)
888
+ 7380 train 3.706801 (lr=1.5986e-04) (hash(x)=77219348)
889
+ 7390 train 3.502160 (lr=1.5901e-04) (hash(x)=79816131)
890
+ 7400 val loss 3.6572
891
+ 7400 val perplexity 38.7531
892
+ 7400 train 3.686786 (lr=1.5816e-04) (hash(x)=67942598)
893
+ 7410 train 3.706089 (lr=1.5732e-04) (hash(x)=67638178)
894
+ 7420 train 3.642549 (lr=1.5649e-04) (hash(x)=71682874)
895
+ 7430 train 3.509881 (lr=1.5566e-04) (hash(x)=71297263)
896
+ 7440 train 3.518374 (lr=1.5484e-04) (hash(x)=73903999)
897
+ 7450 train 3.621816 (lr=1.5402e-04) (hash(x)=74071685)
898
+ 7460 train 3.416296 (lr=1.5321e-04) (hash(x)=71088160)
899
+ 7470 train 3.401508 (lr=1.5241e-04) (hash(x)=74178017)
900
+ 7480 train 3.435699 (lr=1.5161e-04) (hash(x)=84610585)
901
+ 7490 train 3.601688 (lr=1.5081e-04) (hash(x)=67695236)
902
+ 7500 val loss 3.6585
903
+ 7500 val perplexity 38.8026
904
+ 7500 train 3.542574 (lr=1.5002e-04) (hash(x)=72022604)
905
+ 7510 train 3.580838 (lr=1.4924e-04) (hash(x)=78266706)
906
+ 7520 train 3.673334 (lr=1.4847e-04) (hash(x)=72576417)
907
+ 7530 train 3.890779 (lr=1.4769e-04) (hash(x)=71494052)
908
+ 7540 train 3.816607 (lr=1.4693e-04) (hash(x)=76071936)
909
+ 7550 train 3.690997 (lr=1.4617e-04) (hash(x)=78379534)
910
+ 7560 train 3.653231 (lr=1.4542e-04) (hash(x)=75712040)
911
+ 7570 train 3.641760 (lr=1.4467e-04) (hash(x)=83109134)
912
+ 7580 train 3.700063 (lr=1.4393e-04) (hash(x)=78846949)
913
+ 7590 train 3.781287 (lr=1.4319e-04) (hash(x)=63919123)
914
+ 7600 val loss 3.6473
915
+ 7600 val perplexity 38.3707
916
+ 7600 train 3.658151 (lr=1.4246e-04) (hash(x)=73563619)
917
+ 7610 train 3.721436 (lr=1.4174e-04) (hash(x)=79745028)
918
+ 7620 train 3.701515 (lr=1.4102e-04) (hash(x)=68113846)
919
+ 7630 train 3.682285 (lr=1.4031e-04) (hash(x)=70369067)
920
+ 7640 train 3.685153 (lr=1.3960e-04) (hash(x)=79635086)
921
+ 7650 train 3.603169 (lr=1.3890e-04) (hash(x)=77348303)
922
+ 7660 train 3.607666 (lr=1.3821e-04) (hash(x)=74939810)
923
+ 7670 train 3.741526 (lr=1.3752e-04) (hash(x)=77151692)
924
+ 7680 train 3.555557 (lr=1.3684e-04) (hash(x)=69543853)
925
+ 7690 train 3.797997 (lr=1.3616e-04) (hash(x)=75604567)
926
+ 7700 val loss 3.6423
927
+ 7700 val perplexity 38.1779
928
+ 7700 train 3.772254 (lr=1.3549e-04) (hash(x)=84362557)
929
+ 7710 train 3.708710 (lr=1.3483e-04) (hash(x)=71030514)
930
+ 7720 train 3.663011 (lr=1.3417e-04) (hash(x)=76132027)
931
+ 7730 train 3.701341 (lr=1.3352e-04) (hash(x)=77799943)
932
+ 7740 train 3.681250 (lr=1.3287e-04) (hash(x)=74951280)
933
+ 7750 train 3.705631 (lr=1.3223e-04) (hash(x)=73706939)
934
+ 7760 train 3.619767 (lr=1.3160e-04) (hash(x)=78667790)
935
+ 7770 train 3.728979 (lr=1.3097e-04) (hash(x)=72949775)
936
+ 7780 train 3.659230 (lr=1.3035e-04) (hash(x)=74537611)
937
+ 7790 train 3.614482 (lr=1.2974e-04) (hash(x)=72876060)
938
+ 7800 val loss 3.6417
939
+ 7800 val perplexity 38.1550
940
+ 7800 train 3.715736 (lr=1.2913e-04) (hash(x)=73302610)
941
+ 7810 train 3.468505 (lr=1.2852e-04) (hash(x)=66480452)
942
+ 7820 train 3.486559 (lr=1.2793e-04) (hash(x)=74833008)
943
+ 7830 train 3.504539 (lr=1.2733e-04) (hash(x)=66729619)
944
+ 7840 train 3.607941 (lr=1.2675e-04) (hash(x)=87861820)
945
+ 7850 train 3.416695 (lr=1.2617e-04) (hash(x)=73130610)
946
+ 7860 train 3.629555 (lr=1.2560e-04) (hash(x)=67874016)
947
+ 7870 train 3.415697 (lr=1.2503e-04) (hash(x)=70600131)
948
+ 7880 train 3.554274 (lr=1.2447e-04) (hash(x)=68990980)
949
+ 7890 train 3.508148 (lr=1.2392e-04) (hash(x)=79666741)
950
+ 7900 val loss 3.6428
951
+ 7900 val perplexity 38.1974
952
+ 7900 train 3.526054 (lr=1.2337e-04) (hash(x)=74072340)
953
+ 7910 train 3.595364 (lr=1.2283e-04) (hash(x)=80452381)
954
+ 7920 train 3.728763 (lr=1.2229e-04) (hash(x)=83292891)
955
+ 7930 train 3.610550 (lr=1.2176e-04) (hash(x)=80375639)
956
+ 7940 train 3.523642 (lr=1.2124e-04) (hash(x)=82765856)
957
+ 7950 train 3.616413 (lr=1.2072e-04) (hash(x)=72962177)
958
+ 7960 train 3.688146 (lr=1.2021e-04) (hash(x)=74551129)
959
+ 7970 train 3.771068 (lr=1.1970e-04) (hash(x)=89185352)
960
+ 7980 train 3.750160 (lr=1.1921e-04) (hash(x)=81338443)
961
+ 7990 train 3.537557 (lr=1.1871e-04) (hash(x)=83820810)
962
+ 8000 val loss 3.6334
963
+ 8000 val perplexity 37.8429
964
+ 8000 train 3.709970 (lr=1.1823e-04) (hash(x)=72329617)
965
+ 8010 train 3.694041 (lr=1.1775e-04) (hash(x)=82188069)
966
+ 8020 train 3.716743 (lr=1.1728e-04) (hash(x)=78460596)
967
+ 8030 train 3.638503 (lr=1.1681e-04) (hash(x)=76416474)
968
+ 8040 train 3.805143 (lr=1.1635e-04) (hash(x)=70629903)
969
+ 8050 train 3.605256 (lr=1.1589e-04) (hash(x)=76561819)
970
+ 8060 train 3.640028 (lr=1.1544e-04) (hash(x)=71993036)
971
+ 8070 train 3.810941 (lr=1.1500e-04) (hash(x)=73270928)
972
+ 8080 train 3.708490 (lr=1.1457e-04) (hash(x)=82402459)
973
+ 8090 train 3.740438 (lr=1.1414e-04) (hash(x)=67208853)
974
+ 8100 val loss 3.6294
975
+ 8100 val perplexity 37.6890
976
+ 8100 train 3.817009 (lr=1.1371e-04) (hash(x)=85981056)
977
+ 8110 train 3.765708 (lr=1.1330e-04) (hash(x)=77469728)
978
+ 8120 train 3.669600 (lr=1.1289e-04) (hash(x)=84702034)
979
+ 8130 train 3.555752 (lr=1.1248e-04) (hash(x)=96254798)
980
+ 8140 train 3.626621 (lr=1.1209e-04) (hash(x)=71865027)
981
+ 8150 train 3.602555 (lr=1.1169e-04) (hash(x)=79902920)
982
+ 8160 train 3.749343 (lr=1.1131e-04) (hash(x)=70307892)
983
+ 8170 train 3.629095 (lr=1.1093e-04) (hash(x)=73853567)
984
+ 8180 train 3.631683 (lr=1.1056e-04) (hash(x)=72321642)
985
+ 8190 train 3.653842 (lr=1.1019e-04) (hash(x)=82101781)
986
+ 8200 val loss 3.6281
987
+ 8200 val perplexity 37.6394
988
+ 8200 train 3.666280 (lr=1.0983e-04) (hash(x)=66290405)
989
+ 8210 train 3.543933 (lr=1.0948e-04) (hash(x)=69697952)
990
+ 8220 train 3.793462 (lr=1.0913e-04) (hash(x)=72618212)
991
+ 8230 train 3.543591 (lr=1.0879e-04) (hash(x)=76210758)
992
+ 8240 train 3.550376 (lr=1.0846e-04) (hash(x)=70820601)
993
+ 8250 train 3.620776 (lr=1.0813e-04) (hash(x)=75548154)
994
+ 8260 train 3.554576 (lr=1.0781e-04) (hash(x)=73211139)
995
+ 8270 train 3.614822 (lr=1.0750e-04) (hash(x)=82252960)
996
+ 8280 train 3.567270 (lr=1.0719e-04) (hash(x)=69408116)
997
+ 8290 train 3.733830 (lr=1.0689e-04) (hash(x)=74462098)
998
+ 8300 val loss 3.6272
999
+ 8300 val perplexity 37.6068
1000
+ 8300 train 3.739029 (lr=1.0659e-04) (hash(x)=79945329)
1001
+ 8310 train 3.492974 (lr=1.0630e-04) (hash(x)=82459712)
1002
+ 8320 train 3.675274 (lr=1.0602e-04) (hash(x)=81135573)
1003
+ 8330 train 3.650322 (lr=1.0574e-04) (hash(x)=69939721)
1004
+ 8340 train 3.594853 (lr=1.0547e-04) (hash(x)=84687517)
1005
+ 8350 train 3.563389 (lr=1.0521e-04) (hash(x)=77515837)
1006
+ 8360 train 3.567242 (lr=1.0495e-04) (hash(x)=72146536)
1007
+ 8370 train 3.604628 (lr=1.0470e-04) (hash(x)=73198063)
1008
+ 8380 train 3.607463 (lr=1.0446e-04) (hash(x)=68952542)
1009
+ 8390 train 3.646948 (lr=1.0422e-04) (hash(x)=74385341)
1010
+ 8400 val loss 3.6199
1011
+ 8400 val perplexity 37.3320
1012
+ 8400 train 3.467125 (lr=1.0399e-04) (hash(x)=65193622)
1013
+ 8410 train 3.709282 (lr=1.0377e-04) (hash(x)=66981351)
1014
+ 8420 train 3.604985 (lr=1.0355e-04) (hash(x)=97752378)
1015
+ 8430 train 3.800304 (lr=1.0334e-04) (hash(x)=70936998)
1016
+ 8440 train 3.715131 (lr=1.0313e-04) (hash(x)=71557836)
1017
+ 8450 train 3.647805 (lr=1.0293e-04) (hash(x)=71340008)
1018
+ 8460 train 3.696526 (lr=1.0274e-04) (hash(x)=69634745)
1019
+ 8470 train 3.726096 (lr=1.0256e-04) (hash(x)=74224484)
1020
+ 8480 train 3.669441 (lr=1.0238e-04) (hash(x)=75490845)
1021
+ 8490 train 3.546716 (lr=1.0220e-04) (hash(x)=78619557)
1022
+ 8500 val loss 3.6194
1023
+ 8500 val perplexity 37.3161
1024
+ 8500 train 3.561023 (lr=1.0204e-04) (hash(x)=70959421)
1025
+ 8510 train 3.646967 (lr=1.0188e-04) (hash(x)=75975416)
1026
+ 8520 train 3.360667 (lr=1.0172e-04) (hash(x)=76336317)
1027
+ 8530 train 3.623477 (lr=1.0158e-04) (hash(x)=67836668)
1028
+ 8540 train 3.584897 (lr=1.0144e-04) (hash(x)=71139919)
1029
+ 8550 train 3.829602 (lr=1.0130e-04) (hash(x)=69476711)
1030
+ 8560 train 3.578383 (lr=1.0118e-04) (hash(x)=75144101)
1031
+ 8570 train 3.526005 (lr=1.0106e-04) (hash(x)=76246282)
1032
+ 8580 train 3.624581 (lr=1.0094e-04) (hash(x)=69466257)
1033
+ 8590 train 3.598452 (lr=1.0083e-04) (hash(x)=71601133)
1034
+ 8600 val loss 3.6198
1035
+ 8600 val perplexity 37.3288
1036
+ 8600 train 3.568667 (lr=1.0073e-04) (hash(x)=72319507)
1037
+ 8610 train 3.704520 (lr=1.0064e-04) (hash(x)=74618018)
1038
+ 8620 train 3.557616 (lr=1.0055e-04) (hash(x)=76115665)
1039
+ 8630 train 3.714321 (lr=1.0047e-04) (hash(x)=82063941)
1040
+ 8640 train 3.677985 (lr=1.0039e-04) (hash(x)=79867543)
1041
+ 8650 train 3.484867 (lr=1.0033e-04) (hash(x)=73802504)
1042
+ 8660 train 3.593276 (lr=1.0026e-04) (hash(x)=74251531)
1043
+ 8670 train 3.574250 (lr=1.0021e-04) (hash(x)=74637655)
1044
+ 8680 train 3.671430 (lr=1.0016e-04) (hash(x)=82021614)
1045
+ 8690 train 3.685838 (lr=1.0012e-04) (hash(x)=77938695)
1046
+ 8700 val loss 3.6113
1047
+ 8700 val perplexity 37.0127
1048
+ 8700 train 3.720045 (lr=1.0008e-04) (hash(x)=72431581)
1049
+ 8710 train 3.736783 (lr=1.0005e-04) (hash(x)=74875522)
1050
+ 8720 train 3.672248 (lr=1.0003e-04) (hash(x)=69574551)
1051
+ 8730 train 3.510911 (lr=1.0001e-04) (hash(x)=71568849)
1052
+ 8740 train 3.752005 (lr=1.0000e-04) (hash(x)=77714444)
1053
+ 8749 val loss 3.6109
1054
+ 8749 val perplexity 36.9978
n_heads12_lr10e-4_total_batch_size61440_seed1339/model_08749.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cce7b2de6a2c120b4e1ea4805cddb5a8690e9b7bfd03cf50c6237639ca85854c
3
+ size 498607490
n_heads12_lr10e-4_total_batch_size61440_seed1339/optimizer_08749.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1db9005365cfd4cb1062f1a9f51fe1bc4fa2e6a8507f6e78c4a32f9bd1a156d
3
+ size 990934406