| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 6.337135614702155, | |
| "eval_steps": 500, | |
| "global_step": 10000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0063371356147021544, | |
| "grad_norm": 9.841416358947754, | |
| "learning_rate": 2.0000000000000003e-06, | |
| "loss": 1.3229, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.012674271229404309, | |
| "grad_norm": 10.776036262512207, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 1.3359, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.019011406844106463, | |
| "grad_norm": 8.647089004516602, | |
| "learning_rate": 6e-06, | |
| "loss": 1.2916, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.025348542458808618, | |
| "grad_norm": 3.833814859390259, | |
| "learning_rate": 8.000000000000001e-06, | |
| "loss": 0.8317, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.031685678073510776, | |
| "grad_norm": 4.45695686340332, | |
| "learning_rate": 1e-05, | |
| "loss": 0.3917, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.03802281368821293, | |
| "grad_norm": 4.198420524597168, | |
| "learning_rate": 1.2e-05, | |
| "loss": 0.3646, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.044359949302915085, | |
| "grad_norm": 1.619905948638916, | |
| "learning_rate": 1.4000000000000001e-05, | |
| "loss": 0.2154, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.050697084917617236, | |
| "grad_norm": 1.2290397882461548, | |
| "learning_rate": 1.6000000000000003e-05, | |
| "loss": 0.1981, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.057034220532319393, | |
| "grad_norm": 1.4010839462280273, | |
| "learning_rate": 1.8e-05, | |
| "loss": 0.1628, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.06337135614702155, | |
| "grad_norm": 1.5164788961410522, | |
| "learning_rate": 2e-05, | |
| "loss": 0.1349, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.0697084917617237, | |
| "grad_norm": 1.504683017730713, | |
| "learning_rate": 2.2000000000000003e-05, | |
| "loss": 0.1351, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.07604562737642585, | |
| "grad_norm": 1.1718053817749023, | |
| "learning_rate": 2.4e-05, | |
| "loss": 0.1315, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.08238276299112801, | |
| "grad_norm": 1.2100762128829956, | |
| "learning_rate": 2.6000000000000002e-05, | |
| "loss": 0.1088, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.08871989860583017, | |
| "grad_norm": 1.060626745223999, | |
| "learning_rate": 2.8000000000000003e-05, | |
| "loss": 0.1019, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.09505703422053231, | |
| "grad_norm": 1.1521592140197754, | |
| "learning_rate": 3e-05, | |
| "loss": 0.0981, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.10139416983523447, | |
| "grad_norm": 0.9289679527282715, | |
| "learning_rate": 3.2000000000000005e-05, | |
| "loss": 0.0904, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.10773130544993663, | |
| "grad_norm": 1.2749850749969482, | |
| "learning_rate": 3.4000000000000007e-05, | |
| "loss": 0.0903, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.11406844106463879, | |
| "grad_norm": 0.6532027721405029, | |
| "learning_rate": 3.6e-05, | |
| "loss": 0.0844, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.12040557667934093, | |
| "grad_norm": 1.0130654573440552, | |
| "learning_rate": 3.8e-05, | |
| "loss": 0.0952, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.1267427122940431, | |
| "grad_norm": 1.218198299407959, | |
| "learning_rate": 4e-05, | |
| "loss": 0.0875, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.13307984790874525, | |
| "grad_norm": 0.7836218476295471, | |
| "learning_rate": 4.2e-05, | |
| "loss": 0.0813, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.1394169835234474, | |
| "grad_norm": 0.7320715188980103, | |
| "learning_rate": 4.4000000000000006e-05, | |
| "loss": 0.0865, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.14575411913814956, | |
| "grad_norm": 0.6925137639045715, | |
| "learning_rate": 4.600000000000001e-05, | |
| "loss": 0.0651, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.1520912547528517, | |
| "grad_norm": 1.0975725650787354, | |
| "learning_rate": 4.8e-05, | |
| "loss": 0.0735, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.15842839036755388, | |
| "grad_norm": 0.7768546342849731, | |
| "learning_rate": 5e-05, | |
| "loss": 0.0647, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.16476552598225602, | |
| "grad_norm": 0.5613494515419006, | |
| "learning_rate": 5.2000000000000004e-05, | |
| "loss": 0.0608, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.17110266159695817, | |
| "grad_norm": 0.6478670239448547, | |
| "learning_rate": 5.4000000000000005e-05, | |
| "loss": 0.0526, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.17743979721166034, | |
| "grad_norm": 0.9302045106887817, | |
| "learning_rate": 5.6000000000000006e-05, | |
| "loss": 0.0657, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.18377693282636248, | |
| "grad_norm": 0.5966052412986755, | |
| "learning_rate": 5.8e-05, | |
| "loss": 0.0518, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.19011406844106463, | |
| "grad_norm": 0.6293818354606628, | |
| "learning_rate": 6e-05, | |
| "loss": 0.0617, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.1964512040557668, | |
| "grad_norm": 0.7958998084068298, | |
| "learning_rate": 6.2e-05, | |
| "loss": 0.0559, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.20278833967046894, | |
| "grad_norm": 0.7211484909057617, | |
| "learning_rate": 6.400000000000001e-05, | |
| "loss": 0.0534, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.20912547528517111, | |
| "grad_norm": 1.108765959739685, | |
| "learning_rate": 6.6e-05, | |
| "loss": 0.0572, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.21546261089987326, | |
| "grad_norm": 0.9579587578773499, | |
| "learning_rate": 6.800000000000001e-05, | |
| "loss": 0.0513, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.2217997465145754, | |
| "grad_norm": 1.0826679468154907, | |
| "learning_rate": 7e-05, | |
| "loss": 0.0591, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.22813688212927757, | |
| "grad_norm": 0.9620176553726196, | |
| "learning_rate": 7.2e-05, | |
| "loss": 0.0614, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.23447401774397972, | |
| "grad_norm": 0.3815687298774719, | |
| "learning_rate": 7.4e-05, | |
| "loss": 0.0452, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.24081115335868186, | |
| "grad_norm": 1.0042985677719116, | |
| "learning_rate": 7.6e-05, | |
| "loss": 0.0561, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.24714828897338403, | |
| "grad_norm": 0.6386390328407288, | |
| "learning_rate": 7.800000000000001e-05, | |
| "loss": 0.0619, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.2534854245880862, | |
| "grad_norm": 0.7774208784103394, | |
| "learning_rate": 8e-05, | |
| "loss": 0.0478, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.2598225602027883, | |
| "grad_norm": 0.7507523894309998, | |
| "learning_rate": 8.2e-05, | |
| "loss": 0.053, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.2661596958174905, | |
| "grad_norm": 0.8982664346694946, | |
| "learning_rate": 8.4e-05, | |
| "loss": 0.0585, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.27249683143219267, | |
| "grad_norm": 0.83918297290802, | |
| "learning_rate": 8.6e-05, | |
| "loss": 0.0527, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.2788339670468948, | |
| "grad_norm": 0.5620638132095337, | |
| "learning_rate": 8.800000000000001e-05, | |
| "loss": 0.0477, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.28517110266159695, | |
| "grad_norm": 0.7724658250808716, | |
| "learning_rate": 9e-05, | |
| "loss": 0.0503, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.2915082382762991, | |
| "grad_norm": 0.5999326109886169, | |
| "learning_rate": 9.200000000000001e-05, | |
| "loss": 0.0426, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.29784537389100124, | |
| "grad_norm": 0.6010283827781677, | |
| "learning_rate": 9.4e-05, | |
| "loss": 0.0464, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.3041825095057034, | |
| "grad_norm": 0.8040053248405457, | |
| "learning_rate": 9.6e-05, | |
| "loss": 0.0455, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.3105196451204056, | |
| "grad_norm": 1.1320879459381104, | |
| "learning_rate": 9.8e-05, | |
| "loss": 0.0513, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.31685678073510776, | |
| "grad_norm": 0.7393860816955566, | |
| "learning_rate": 0.0001, | |
| "loss": 0.0506, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.3231939163498099, | |
| "grad_norm": 0.9977770447731018, | |
| "learning_rate": 9.999972660400536e-05, | |
| "loss": 0.0511, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.32953105196451205, | |
| "grad_norm": 0.9153000116348267, | |
| "learning_rate": 9.999890641901125e-05, | |
| "loss": 0.0446, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.3358681875792142, | |
| "grad_norm": 0.7802489399909973, | |
| "learning_rate": 9.999753945398704e-05, | |
| "loss": 0.045, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.34220532319391633, | |
| "grad_norm": 0.7009689211845398, | |
| "learning_rate": 9.99956257238817e-05, | |
| "loss": 0.0537, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.3485424588086185, | |
| "grad_norm": 0.5555402636528015, | |
| "learning_rate": 9.999316524962345e-05, | |
| "loss": 0.0416, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.3548795944233207, | |
| "grad_norm": 0.613745391368866, | |
| "learning_rate": 9.999015805811965e-05, | |
| "loss": 0.0463, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.3612167300380228, | |
| "grad_norm": 0.9725404381752014, | |
| "learning_rate": 9.998660418225645e-05, | |
| "loss": 0.0534, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.36755386565272496, | |
| "grad_norm": 0.6908450126647949, | |
| "learning_rate": 9.998250366089848e-05, | |
| "loss": 0.0486, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.37389100126742714, | |
| "grad_norm": 0.8201190829277039, | |
| "learning_rate": 9.997785653888835e-05, | |
| "loss": 0.0381, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.38022813688212925, | |
| "grad_norm": 0.7427898645401001, | |
| "learning_rate": 9.997266286704631e-05, | |
| "loss": 0.0504, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.3865652724968314, | |
| "grad_norm": 0.5355698466300964, | |
| "learning_rate": 9.996692270216947e-05, | |
| "loss": 0.0454, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.3929024081115336, | |
| "grad_norm": 0.6287090182304382, | |
| "learning_rate": 9.996063610703137e-05, | |
| "loss": 0.0481, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.39923954372623577, | |
| "grad_norm": 0.5858126878738403, | |
| "learning_rate": 9.995380315038119e-05, | |
| "loss": 0.0379, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.4055766793409379, | |
| "grad_norm": 0.3876695930957794, | |
| "learning_rate": 9.994642390694308e-05, | |
| "loss": 0.0387, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.41191381495564006, | |
| "grad_norm": 0.5101262331008911, | |
| "learning_rate": 9.993849845741524e-05, | |
| "loss": 0.0499, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.41825095057034223, | |
| "grad_norm": 0.7811976075172424, | |
| "learning_rate": 9.993002688846913e-05, | |
| "loss": 0.0458, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.42458808618504434, | |
| "grad_norm": 0.6726566553115845, | |
| "learning_rate": 9.992100929274846e-05, | |
| "loss": 0.0359, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.4309252217997465, | |
| "grad_norm": 0.43720176815986633, | |
| "learning_rate": 9.991144576886823e-05, | |
| "loss": 0.0525, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.4372623574144487, | |
| "grad_norm": 1.0024137496948242, | |
| "learning_rate": 9.990133642141359e-05, | |
| "loss": 0.04, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.4435994930291508, | |
| "grad_norm": 0.6060113310813904, | |
| "learning_rate": 9.989068136093873e-05, | |
| "loss": 0.0354, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.449936628643853, | |
| "grad_norm": 0.8775745630264282, | |
| "learning_rate": 9.987948070396571e-05, | |
| "loss": 0.0432, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.45627376425855515, | |
| "grad_norm": 0.7044681310653687, | |
| "learning_rate": 9.986773457298311e-05, | |
| "loss": 0.0347, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.46261089987325726, | |
| "grad_norm": 0.5258815884590149, | |
| "learning_rate": 9.985544309644475e-05, | |
| "loss": 0.0428, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.46894803548795944, | |
| "grad_norm": 0.5640819072723389, | |
| "learning_rate": 9.984260640876821e-05, | |
| "loss": 0.0477, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.4752851711026616, | |
| "grad_norm": 0.9232666492462158, | |
| "learning_rate": 9.98292246503335e-05, | |
| "loss": 0.0398, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.4816223067173637, | |
| "grad_norm": 0.7809548377990723, | |
| "learning_rate": 9.981529796748134e-05, | |
| "loss": 0.042, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.4879594423320659, | |
| "grad_norm": 0.5453559756278992, | |
| "learning_rate": 9.980082651251175e-05, | |
| "loss": 0.0413, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.49429657794676807, | |
| "grad_norm": 0.43146708607673645, | |
| "learning_rate": 9.97858104436822e-05, | |
| "loss": 0.0395, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.5006337135614702, | |
| "grad_norm": 0.6322312951087952, | |
| "learning_rate": 9.977024992520602e-05, | |
| "loss": 0.0303, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.5069708491761724, | |
| "grad_norm": 0.3923812508583069, | |
| "learning_rate": 9.975414512725057e-05, | |
| "loss": 0.0417, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.5133079847908745, | |
| "grad_norm": 0.43715372681617737, | |
| "learning_rate": 9.973749622593534e-05, | |
| "loss": 0.0357, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.5196451204055766, | |
| "grad_norm": 0.4882504343986511, | |
| "learning_rate": 9.972030340333001e-05, | |
| "loss": 0.0483, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.5259822560202788, | |
| "grad_norm": 0.47166863083839417, | |
| "learning_rate": 9.970256684745258e-05, | |
| "loss": 0.033, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.532319391634981, | |
| "grad_norm": 0.6284196972846985, | |
| "learning_rate": 9.968428675226714e-05, | |
| "loss": 0.034, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.5386565272496832, | |
| "grad_norm": 0.34937915205955505, | |
| "learning_rate": 9.966546331768191e-05, | |
| "loss": 0.0288, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.5449936628643853, | |
| "grad_norm": 0.6910156607627869, | |
| "learning_rate": 9.964609674954696e-05, | |
| "loss": 0.0413, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.5513307984790875, | |
| "grad_norm": 0.3398362994194031, | |
| "learning_rate": 9.962618725965196e-05, | |
| "loss": 0.0295, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.5576679340937896, | |
| "grad_norm": 0.5590093731880188, | |
| "learning_rate": 9.96057350657239e-05, | |
| "loss": 0.0332, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.5640050697084917, | |
| "grad_norm": 0.444553405046463, | |
| "learning_rate": 9.95847403914247e-05, | |
| "loss": 0.0331, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.5703422053231939, | |
| "grad_norm": 0.4849596619606018, | |
| "learning_rate": 9.956320346634876e-05, | |
| "loss": 0.0388, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.5766793409378961, | |
| "grad_norm": 0.33844664692878723, | |
| "learning_rate": 9.954112452602045e-05, | |
| "loss": 0.0324, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.5830164765525983, | |
| "grad_norm": 0.2623671889305115, | |
| "learning_rate": 9.95185038118915e-05, | |
| "loss": 0.0403, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.5893536121673004, | |
| "grad_norm": 0.5219829678535461, | |
| "learning_rate": 9.949534157133844e-05, | |
| "loss": 0.0466, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.5956907477820025, | |
| "grad_norm": 0.40922799706459045, | |
| "learning_rate": 9.94716380576598e-05, | |
| "loss": 0.0295, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.6020278833967047, | |
| "grad_norm": 0.37367522716522217, | |
| "learning_rate": 9.944739353007344e-05, | |
| "loss": 0.0352, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.6083650190114068, | |
| "grad_norm": 0.8563676476478577, | |
| "learning_rate": 9.942260825371358e-05, | |
| "loss": 0.0424, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.614702154626109, | |
| "grad_norm": 0.396813303232193, | |
| "learning_rate": 9.939728249962807e-05, | |
| "loss": 0.0386, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.6210392902408112, | |
| "grad_norm": 0.5535157918930054, | |
| "learning_rate": 9.937141654477528e-05, | |
| "loss": 0.0396, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.6273764258555133, | |
| "grad_norm": 0.4769563376903534, | |
| "learning_rate": 9.934501067202117e-05, | |
| "loss": 0.0372, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.6337135614702155, | |
| "grad_norm": 0.5703756809234619, | |
| "learning_rate": 9.931806517013612e-05, | |
| "loss": 0.0325, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.6400506970849176, | |
| "grad_norm": 0.31805068254470825, | |
| "learning_rate": 9.929058033379181e-05, | |
| "loss": 0.0365, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.6463878326996197, | |
| "grad_norm": 0.5349836945533752, | |
| "learning_rate": 9.926255646355804e-05, | |
| "loss": 0.0265, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.6527249683143219, | |
| "grad_norm": 0.5509829521179199, | |
| "learning_rate": 9.923399386589933e-05, | |
| "loss": 0.0273, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.6590621039290241, | |
| "grad_norm": 0.5818154215812683, | |
| "learning_rate": 9.92048928531717e-05, | |
| "loss": 0.0334, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.6653992395437263, | |
| "grad_norm": 0.633552074432373, | |
| "learning_rate": 9.917525374361912e-05, | |
| "loss": 0.0297, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.6717363751584284, | |
| "grad_norm": 0.505277156829834, | |
| "learning_rate": 9.914507686137019e-05, | |
| "loss": 0.0386, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.6780735107731305, | |
| "grad_norm": 0.30526870489120483, | |
| "learning_rate": 9.911436253643445e-05, | |
| "loss": 0.0315, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.6844106463878327, | |
| "grad_norm": 0.47638586163520813, | |
| "learning_rate": 9.90831111046988e-05, | |
| "loss": 0.036, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.6907477820025348, | |
| "grad_norm": 0.38751357793807983, | |
| "learning_rate": 9.905132290792394e-05, | |
| "loss": 0.0324, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.697084917617237, | |
| "grad_norm": 0.20010784268379211, | |
| "learning_rate": 9.901899829374047e-05, | |
| "loss": 0.0231, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.7034220532319392, | |
| "grad_norm": 0.4107750654220581, | |
| "learning_rate": 9.89861376156452e-05, | |
| "loss": 0.0251, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.7097591888466414, | |
| "grad_norm": 0.24435921013355255, | |
| "learning_rate": 9.895274123299723e-05, | |
| "loss": 0.027, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.7160963244613435, | |
| "grad_norm": 0.4334017336368561, | |
| "learning_rate": 9.891880951101407e-05, | |
| "loss": 0.0345, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.7224334600760456, | |
| "grad_norm": 0.4201258420944214, | |
| "learning_rate": 9.888434282076758e-05, | |
| "loss": 0.0283, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.7287705956907478, | |
| "grad_norm": 0.41078707575798035, | |
| "learning_rate": 9.884934153917997e-05, | |
| "loss": 0.0335, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.7351077313054499, | |
| "grad_norm": 0.37457314133644104, | |
| "learning_rate": 9.881380604901964e-05, | |
| "loss": 0.0298, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.7414448669201521, | |
| "grad_norm": 0.7165825366973877, | |
| "learning_rate": 9.877773673889701e-05, | |
| "loss": 0.0338, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.7477820025348543, | |
| "grad_norm": 0.3498821556568146, | |
| "learning_rate": 9.87411340032603e-05, | |
| "loss": 0.0335, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.7541191381495564, | |
| "grad_norm": 0.3448495864868164, | |
| "learning_rate": 9.870399824239117e-05, | |
| "loss": 0.0273, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.7604562737642585, | |
| "grad_norm": 0.5568084716796875, | |
| "learning_rate": 9.86663298624003e-05, | |
| "loss": 0.0396, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.7667934093789607, | |
| "grad_norm": 0.4418658912181854, | |
| "learning_rate": 9.862812927522309e-05, | |
| "loss": 0.0376, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.7731305449936628, | |
| "grad_norm": 0.5665590167045593, | |
| "learning_rate": 9.858939689861506e-05, | |
| "loss": 0.0307, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.779467680608365, | |
| "grad_norm": 0.45505577325820923, | |
| "learning_rate": 9.855013315614725e-05, | |
| "loss": 0.0263, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.7858048162230672, | |
| "grad_norm": 0.4842069745063782, | |
| "learning_rate": 9.851033847720166e-05, | |
| "loss": 0.0308, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.7921419518377694, | |
| "grad_norm": 0.4048508405685425, | |
| "learning_rate": 9.847001329696653e-05, | |
| "loss": 0.0243, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.7984790874524715, | |
| "grad_norm": 0.24489720165729523, | |
| "learning_rate": 9.842915805643155e-05, | |
| "loss": 0.0281, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.8048162230671736, | |
| "grad_norm": 0.533295750617981, | |
| "learning_rate": 9.838777320238312e-05, | |
| "loss": 0.0263, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.8111533586818758, | |
| "grad_norm": 0.36894357204437256, | |
| "learning_rate": 9.834585918739936e-05, | |
| "loss": 0.0316, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.8174904942965779, | |
| "grad_norm": 0.2535454332828522, | |
| "learning_rate": 9.830341646984521e-05, | |
| "loss": 0.0294, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.8238276299112801, | |
| "grad_norm": 0.263375461101532, | |
| "learning_rate": 9.826044551386744e-05, | |
| "loss": 0.031, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.8301647655259823, | |
| "grad_norm": 0.4550211429595947, | |
| "learning_rate": 9.821694678938953e-05, | |
| "loss": 0.0278, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.8365019011406845, | |
| "grad_norm": 0.5714216828346252, | |
| "learning_rate": 9.817292077210659e-05, | |
| "loss": 0.0295, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.8428390367553865, | |
| "grad_norm": 0.6447302103042603, | |
| "learning_rate": 9.812836794348004e-05, | |
| "loss": 0.0331, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.8491761723700887, | |
| "grad_norm": 0.4367750585079193, | |
| "learning_rate": 9.808328879073251e-05, | |
| "loss": 0.0363, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.8555133079847909, | |
| "grad_norm": 0.34794682264328003, | |
| "learning_rate": 9.803768380684242e-05, | |
| "loss": 0.0241, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.861850443599493, | |
| "grad_norm": 0.6189941167831421, | |
| "learning_rate": 9.799155349053851e-05, | |
| "loss": 0.0325, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.8681875792141952, | |
| "grad_norm": 0.4560254216194153, | |
| "learning_rate": 9.794489834629455e-05, | |
| "loss": 0.0242, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.8745247148288974, | |
| "grad_norm": 0.5019411444664001, | |
| "learning_rate": 9.789771888432375e-05, | |
| "loss": 0.0321, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.8808618504435995, | |
| "grad_norm": 0.7201408743858337, | |
| "learning_rate": 9.785001562057309e-05, | |
| "loss": 0.0334, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.8871989860583016, | |
| "grad_norm": 0.3263881206512451, | |
| "learning_rate": 9.780178907671789e-05, | |
| "loss": 0.0292, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.8935361216730038, | |
| "grad_norm": 0.552166759967804, | |
| "learning_rate": 9.775303978015585e-05, | |
| "loss": 0.04, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.899873257287706, | |
| "grad_norm": 0.520951509475708, | |
| "learning_rate": 9.77037682640015e-05, | |
| "loss": 0.0321, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.9062103929024081, | |
| "grad_norm": 0.6059055924415588, | |
| "learning_rate": 9.765397506708023e-05, | |
| "loss": 0.0344, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.9125475285171103, | |
| "grad_norm": 0.6472659707069397, | |
| "learning_rate": 9.760366073392246e-05, | |
| "loss": 0.0272, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.9188846641318125, | |
| "grad_norm": 0.2647618055343628, | |
| "learning_rate": 9.755282581475769e-05, | |
| "loss": 0.0316, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.9252217997465145, | |
| "grad_norm": 0.4789205491542816, | |
| "learning_rate": 9.750147086550844e-05, | |
| "loss": 0.0269, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.9315589353612167, | |
| "grad_norm": 0.3620862364768982, | |
| "learning_rate": 9.744959644778422e-05, | |
| "loss": 0.0198, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.9378960709759189, | |
| "grad_norm": 0.33386313915252686, | |
| "learning_rate": 9.739720312887535e-05, | |
| "loss": 0.0248, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.944233206590621, | |
| "grad_norm": 0.3961511254310608, | |
| "learning_rate": 9.734429148174675e-05, | |
| "loss": 0.0244, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.9505703422053232, | |
| "grad_norm": 0.29284903407096863, | |
| "learning_rate": 9.729086208503174e-05, | |
| "loss": 0.0278, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.9569074778200254, | |
| "grad_norm": 0.2947191596031189, | |
| "learning_rate": 9.723691552302562e-05, | |
| "loss": 0.0202, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.9632446134347274, | |
| "grad_norm": 0.5471847653388977, | |
| "learning_rate": 9.718245238567939e-05, | |
| "loss": 0.0235, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.9695817490494296, | |
| "grad_norm": 0.6192600727081299, | |
| "learning_rate": 9.712747326859315e-05, | |
| "loss": 0.0265, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.9759188846641318, | |
| "grad_norm": 0.49848631024360657, | |
| "learning_rate": 9.707197877300974e-05, | |
| "loss": 0.0211, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.982256020278834, | |
| "grad_norm": 0.5650535225868225, | |
| "learning_rate": 9.701596950580806e-05, | |
| "loss": 0.0323, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.9885931558935361, | |
| "grad_norm": 0.4596526324748993, | |
| "learning_rate": 9.695944607949649e-05, | |
| "loss": 0.0214, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.9949302915082383, | |
| "grad_norm": 0.43541160225868225, | |
| "learning_rate": 9.690240911220618e-05, | |
| "loss": 0.0249, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.0012674271229405, | |
| "grad_norm": 0.3652093708515167, | |
| "learning_rate": 9.684485922768422e-05, | |
| "loss": 0.0217, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.0076045627376427, | |
| "grad_norm": 0.6113967895507812, | |
| "learning_rate": 9.6786797055287e-05, | |
| "loss": 0.0288, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.0139416983523448, | |
| "grad_norm": 0.35289207100868225, | |
| "learning_rate": 9.672822322997305e-05, | |
| "loss": 0.0233, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.020278833967047, | |
| "grad_norm": 0.4643147587776184, | |
| "learning_rate": 9.66691383922964e-05, | |
| "loss": 0.0255, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.026615969581749, | |
| "grad_norm": 0.186619833111763, | |
| "learning_rate": 9.660954318839933e-05, | |
| "loss": 0.0239, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.0329531051964511, | |
| "grad_norm": 0.46518614888191223, | |
| "learning_rate": 9.654943827000548e-05, | |
| "loss": 0.0275, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.0392902408111533, | |
| "grad_norm": 0.32950979471206665, | |
| "learning_rate": 9.648882429441257e-05, | |
| "loss": 0.0289, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.0456273764258555, | |
| "grad_norm": 0.35503265261650085, | |
| "learning_rate": 9.642770192448536e-05, | |
| "loss": 0.023, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.0519645120405576, | |
| "grad_norm": 0.29523852467536926, | |
| "learning_rate": 9.636607182864827e-05, | |
| "loss": 0.029, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.0583016476552598, | |
| "grad_norm": 0.5499984622001648, | |
| "learning_rate": 9.630393468087818e-05, | |
| "loss": 0.0232, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.064638783269962, | |
| "grad_norm": 0.3574666678905487, | |
| "learning_rate": 9.624129116069694e-05, | |
| "loss": 0.0232, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.0709759188846641, | |
| "grad_norm": 0.4880392551422119, | |
| "learning_rate": 9.617814195316411e-05, | |
| "loss": 0.0348, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.0773130544993663, | |
| "grad_norm": 0.6365160942077637, | |
| "learning_rate": 9.611448774886924e-05, | |
| "loss": 0.0299, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.0836501901140685, | |
| "grad_norm": 0.27736377716064453, | |
| "learning_rate": 9.605032924392457e-05, | |
| "loss": 0.0219, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.0899873257287707, | |
| "grad_norm": 0.34657785296440125, | |
| "learning_rate": 9.598566713995718e-05, | |
| "loss": 0.0291, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.0963244613434728, | |
| "grad_norm": 0.3848840296268463, | |
| "learning_rate": 9.59205021441015e-05, | |
| "loss": 0.0244, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.102661596958175, | |
| "grad_norm": 0.29880332946777344, | |
| "learning_rate": 9.58548349689915e-05, | |
| "loss": 0.0276, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.1089987325728772, | |
| "grad_norm": 0.32027584314346313, | |
| "learning_rate": 9.578866633275288e-05, | |
| "loss": 0.0178, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.1153358681875791, | |
| "grad_norm": 0.32828688621520996, | |
| "learning_rate": 9.572199695899522e-05, | |
| "loss": 0.0221, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.1216730038022813, | |
| "grad_norm": 0.3575758934020996, | |
| "learning_rate": 9.565482757680415e-05, | |
| "loss": 0.0226, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.1280101394169835, | |
| "grad_norm": 0.39491409063339233, | |
| "learning_rate": 9.558715892073323e-05, | |
| "loss": 0.0213, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.1343472750316856, | |
| "grad_norm": 0.3913297951221466, | |
| "learning_rate": 9.551899173079607e-05, | |
| "loss": 0.0242, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.1406844106463878, | |
| "grad_norm": 0.3732500374317169, | |
| "learning_rate": 9.545032675245813e-05, | |
| "loss": 0.0262, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.14702154626109, | |
| "grad_norm": 0.3922960162162781, | |
| "learning_rate": 9.538116473662861e-05, | |
| "loss": 0.0244, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.1533586818757922, | |
| "grad_norm": 0.30090585350990295, | |
| "learning_rate": 9.531150643965223e-05, | |
| "loss": 0.019, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.1596958174904943, | |
| "grad_norm": 0.20526911318302155, | |
| "learning_rate": 9.524135262330098e-05, | |
| "loss": 0.021, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.1660329531051965, | |
| "grad_norm": 0.281562864780426, | |
| "learning_rate": 9.517070405476575e-05, | |
| "loss": 0.0237, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.1723700887198987, | |
| "grad_norm": 0.3962242603302002, | |
| "learning_rate": 9.509956150664796e-05, | |
| "loss": 0.023, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.1787072243346008, | |
| "grad_norm": 0.4091711938381195, | |
| "learning_rate": 9.502792575695112e-05, | |
| "loss": 0.0227, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.1850443599493028, | |
| "grad_norm": 0.5421469807624817, | |
| "learning_rate": 9.49557975890723e-05, | |
| "loss": 0.0222, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.1913814955640052, | |
| "grad_norm": 0.3753933906555176, | |
| "learning_rate": 9.488317779179361e-05, | |
| "loss": 0.027, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.1977186311787071, | |
| "grad_norm": 0.2167666256427765, | |
| "learning_rate": 9.481006715927351e-05, | |
| "loss": 0.0263, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.2040557667934093, | |
| "grad_norm": 0.4261206090450287, | |
| "learning_rate": 9.473646649103818e-05, | |
| "loss": 0.0309, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.2103929024081115, | |
| "grad_norm": 0.5334855318069458, | |
| "learning_rate": 9.46623765919727e-05, | |
| "loss": 0.0242, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.2167300380228137, | |
| "grad_norm": 0.2908576726913452, | |
| "learning_rate": 9.458779827231237e-05, | |
| "loss": 0.0199, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.2230671736375158, | |
| "grad_norm": 0.3591972589492798, | |
| "learning_rate": 9.451273234763371e-05, | |
| "loss": 0.0247, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.229404309252218, | |
| "grad_norm": 0.3506092131137848, | |
| "learning_rate": 9.443717963884569e-05, | |
| "loss": 0.0213, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.2357414448669202, | |
| "grad_norm": 0.2472591996192932, | |
| "learning_rate": 9.43611409721806e-05, | |
| "loss": 0.0237, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.2420785804816223, | |
| "grad_norm": 0.33952316641807556, | |
| "learning_rate": 9.428461717918511e-05, | |
| "loss": 0.0219, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.2484157160963245, | |
| "grad_norm": 0.17157837748527527, | |
| "learning_rate": 9.420760909671118e-05, | |
| "loss": 0.0268, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.2547528517110267, | |
| "grad_norm": 0.2349107414484024, | |
| "learning_rate": 9.413011756690685e-05, | |
| "loss": 0.0323, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.2610899873257289, | |
| "grad_norm": 0.3310667872428894, | |
| "learning_rate": 9.405214343720707e-05, | |
| "loss": 0.022, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 1.2674271229404308, | |
| "grad_norm": 0.45049792528152466, | |
| "learning_rate": 9.397368756032445e-05, | |
| "loss": 0.0236, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 1.2737642585551332, | |
| "grad_norm": 0.5270654559135437, | |
| "learning_rate": 9.389475079423988e-05, | |
| "loss": 0.0233, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 1.2801013941698351, | |
| "grad_norm": 0.4595666229724884, | |
| "learning_rate": 9.381533400219318e-05, | |
| "loss": 0.0224, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 1.2864385297845373, | |
| "grad_norm": 0.510229766368866, | |
| "learning_rate": 9.373543805267368e-05, | |
| "loss": 0.025, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 1.2927756653992395, | |
| "grad_norm": 0.38161230087280273, | |
| "learning_rate": 9.365506381941066e-05, | |
| "loss": 0.0206, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 1.2991128010139417, | |
| "grad_norm": 0.3577732443809509, | |
| "learning_rate": 9.357421218136386e-05, | |
| "loss": 0.0242, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 1.3054499366286438, | |
| "grad_norm": 0.2776769995689392, | |
| "learning_rate": 9.349288402271388e-05, | |
| "loss": 0.0244, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 1.311787072243346, | |
| "grad_norm": 0.2369217425584793, | |
| "learning_rate": 9.341108023285238e-05, | |
| "loss": 0.0274, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 1.3181242078580482, | |
| "grad_norm": 0.3153173327445984, | |
| "learning_rate": 9.332880170637252e-05, | |
| "loss": 0.0223, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.3244613434727504, | |
| "grad_norm": 0.3351927399635315, | |
| "learning_rate": 9.32460493430591e-05, | |
| "loss": 0.0191, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 1.3307984790874525, | |
| "grad_norm": 0.3907555639743805, | |
| "learning_rate": 9.316282404787871e-05, | |
| "loss": 0.0199, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 1.3371356147021547, | |
| "grad_norm": 0.3048558235168457, | |
| "learning_rate": 9.30791267309698e-05, | |
| "loss": 0.0257, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 1.3434727503168569, | |
| "grad_norm": 0.4380377233028412, | |
| "learning_rate": 9.299495830763286e-05, | |
| "loss": 0.0317, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 1.3498098859315588, | |
| "grad_norm": 0.47419941425323486, | |
| "learning_rate": 9.291031969832026e-05, | |
| "loss": 0.0294, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 1.3561470215462612, | |
| "grad_norm": 0.30324944853782654, | |
| "learning_rate": 9.282521182862629e-05, | |
| "loss": 0.0233, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 1.3624841571609632, | |
| "grad_norm": 0.3189409375190735, | |
| "learning_rate": 9.273963562927695e-05, | |
| "loss": 0.025, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 1.3688212927756653, | |
| "grad_norm": 0.34753671288490295, | |
| "learning_rate": 9.265359203611987e-05, | |
| "loss": 0.0302, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 1.3751584283903675, | |
| "grad_norm": 0.2331438660621643, | |
| "learning_rate": 9.256708199011401e-05, | |
| "loss": 0.0165, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 1.3814955640050697, | |
| "grad_norm": 0.27667322754859924, | |
| "learning_rate": 9.248010643731935e-05, | |
| "loss": 0.0175, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 1.3878326996197718, | |
| "grad_norm": 0.5424685478210449, | |
| "learning_rate": 9.239266632888659e-05, | |
| "loss": 0.024, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 1.394169835234474, | |
| "grad_norm": 0.2840613126754761, | |
| "learning_rate": 9.230476262104677e-05, | |
| "loss": 0.0187, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 1.4005069708491762, | |
| "grad_norm": 0.43624839186668396, | |
| "learning_rate": 9.221639627510076e-05, | |
| "loss": 0.0253, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 1.4068441064638784, | |
| "grad_norm": 0.27831771969795227, | |
| "learning_rate": 9.212756825740873e-05, | |
| "loss": 0.0206, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 1.4131812420785805, | |
| "grad_norm": 0.2653898298740387, | |
| "learning_rate": 9.20382795393797e-05, | |
| "loss": 0.0228, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 1.4195183776932827, | |
| "grad_norm": 0.4191180467605591, | |
| "learning_rate": 9.194853109746074e-05, | |
| "loss": 0.0204, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 1.4258555133079849, | |
| "grad_norm": 0.3698088526725769, | |
| "learning_rate": 9.185832391312644e-05, | |
| "loss": 0.0201, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 1.4321926489226868, | |
| "grad_norm": 0.32586073875427246, | |
| "learning_rate": 9.176765897286813e-05, | |
| "loss": 0.0311, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 1.4385297845373892, | |
| "grad_norm": 0.26891350746154785, | |
| "learning_rate": 9.167653726818305e-05, | |
| "loss": 0.0228, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 1.4448669201520912, | |
| "grad_norm": 0.4289737939834595, | |
| "learning_rate": 9.158495979556358e-05, | |
| "loss": 0.0221, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 1.4512040557667933, | |
| "grad_norm": 0.3612203598022461, | |
| "learning_rate": 9.14929275564863e-05, | |
| "loss": 0.0215, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 1.4575411913814955, | |
| "grad_norm": 0.511667013168335, | |
| "learning_rate": 9.140044155740101e-05, | |
| "loss": 0.0214, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 1.4638783269961977, | |
| "grad_norm": 0.31154608726501465, | |
| "learning_rate": 9.130750280971978e-05, | |
| "loss": 0.0228, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 1.4702154626108999, | |
| "grad_norm": 0.33954188227653503, | |
| "learning_rate": 9.121411232980588e-05, | |
| "loss": 0.0267, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 1.476552598225602, | |
| "grad_norm": 0.4101477563381195, | |
| "learning_rate": 9.112027113896262e-05, | |
| "loss": 0.0204, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 1.4828897338403042, | |
| "grad_norm": 0.3624727427959442, | |
| "learning_rate": 9.102598026342222e-05, | |
| "loss": 0.0194, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 1.4892268694550064, | |
| "grad_norm": 0.29435303807258606, | |
| "learning_rate": 9.093124073433463e-05, | |
| "loss": 0.0227, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 1.4955640050697085, | |
| "grad_norm": 0.3966459035873413, | |
| "learning_rate": 9.083605358775612e-05, | |
| "loss": 0.02, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 1.5019011406844105, | |
| "grad_norm": 0.459172785282135, | |
| "learning_rate": 9.074041986463808e-05, | |
| "loss": 0.025, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 1.508238276299113, | |
| "grad_norm": 0.32978248596191406, | |
| "learning_rate": 9.064434061081562e-05, | |
| "loss": 0.0186, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 1.5145754119138148, | |
| "grad_norm": 0.38167238235473633, | |
| "learning_rate": 9.0547816876996e-05, | |
| "loss": 0.0245, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 1.5209125475285172, | |
| "grad_norm": 0.32322126626968384, | |
| "learning_rate": 9.045084971874738e-05, | |
| "loss": 0.022, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 1.5272496831432192, | |
| "grad_norm": 0.2745392322540283, | |
| "learning_rate": 9.035344019648702e-05, | |
| "loss": 0.0209, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 1.5335868187579216, | |
| "grad_norm": 0.3605414032936096, | |
| "learning_rate": 9.025558937546988e-05, | |
| "loss": 0.0213, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 1.5399239543726235, | |
| "grad_norm": 0.32733994722366333, | |
| "learning_rate": 9.015729832577681e-05, | |
| "loss": 0.0172, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 1.5462610899873257, | |
| "grad_norm": 0.6043369174003601, | |
| "learning_rate": 9.005856812230304e-05, | |
| "loss": 0.0239, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 1.5525982256020279, | |
| "grad_norm": 0.22895248234272003, | |
| "learning_rate": 8.995939984474624e-05, | |
| "loss": 0.019, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 1.55893536121673, | |
| "grad_norm": 0.4059270918369293, | |
| "learning_rate": 8.98597945775948e-05, | |
| "loss": 0.0224, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 1.5652724968314322, | |
| "grad_norm": 0.3693525791168213, | |
| "learning_rate": 8.975975341011596e-05, | |
| "loss": 0.0187, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 1.5716096324461344, | |
| "grad_norm": 0.47394832968711853, | |
| "learning_rate": 8.965927743634391e-05, | |
| "loss": 0.021, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 1.5779467680608366, | |
| "grad_norm": 0.3060173988342285, | |
| "learning_rate": 8.955836775506776e-05, | |
| "loss": 0.0216, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 1.5842839036755385, | |
| "grad_norm": 0.2988053560256958, | |
| "learning_rate": 8.945702546981969e-05, | |
| "loss": 0.024, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 1.590621039290241, | |
| "grad_norm": 0.5819903612136841, | |
| "learning_rate": 8.935525168886262e-05, | |
| "loss": 0.0225, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 1.5969581749049429, | |
| "grad_norm": 0.34001755714416504, | |
| "learning_rate": 8.92530475251784e-05, | |
| "loss": 0.02, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 1.6032953105196452, | |
| "grad_norm": 0.45877692103385925, | |
| "learning_rate": 8.91504140964553e-05, | |
| "loss": 0.0215, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 1.6096324461343472, | |
| "grad_norm": 0.4635033905506134, | |
| "learning_rate": 8.90473525250761e-05, | |
| "loss": 0.0325, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 1.6159695817490496, | |
| "grad_norm": 0.42221376299858093, | |
| "learning_rate": 8.894386393810563e-05, | |
| "loss": 0.0196, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 1.6223067173637515, | |
| "grad_norm": 0.34406083822250366, | |
| "learning_rate": 8.883994946727849e-05, | |
| "loss": 0.0245, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 1.6286438529784537, | |
| "grad_norm": 0.4621538519859314, | |
| "learning_rate": 8.873561024898668e-05, | |
| "loss": 0.0236, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 1.6349809885931559, | |
| "grad_norm": 0.3699153959751129, | |
| "learning_rate": 8.863084742426719e-05, | |
| "loss": 0.024, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 1.641318124207858, | |
| "grad_norm": 0.4299221336841583, | |
| "learning_rate": 8.852566213878947e-05, | |
| "loss": 0.0173, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 1.6476552598225602, | |
| "grad_norm": 0.25442177057266235, | |
| "learning_rate": 8.842005554284296e-05, | |
| "loss": 0.021, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 1.6539923954372624, | |
| "grad_norm": 0.31282278895378113, | |
| "learning_rate": 8.831402879132446e-05, | |
| "loss": 0.0238, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 1.6603295310519646, | |
| "grad_norm": 0.400104820728302, | |
| "learning_rate": 8.820758304372557e-05, | |
| "loss": 0.0194, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 1.6666666666666665, | |
| "grad_norm": 0.5096003413200378, | |
| "learning_rate": 8.810071946411989e-05, | |
| "loss": 0.0289, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 1.673003802281369, | |
| "grad_norm": 0.36454638838768005, | |
| "learning_rate": 8.799343922115044e-05, | |
| "loss": 0.0245, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 1.6793409378960709, | |
| "grad_norm": 0.3085906207561493, | |
| "learning_rate": 8.788574348801675e-05, | |
| "loss": 0.0225, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 1.6856780735107733, | |
| "grad_norm": 0.25602665543556213, | |
| "learning_rate": 8.77776334424621e-05, | |
| "loss": 0.0202, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 1.6920152091254752, | |
| "grad_norm": 0.3288155496120453, | |
| "learning_rate": 8.766911026676064e-05, | |
| "loss": 0.0307, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 1.6983523447401776, | |
| "grad_norm": 0.2693244218826294, | |
| "learning_rate": 8.756017514770443e-05, | |
| "loss": 0.0169, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 1.7046894803548795, | |
| "grad_norm": 0.20038799941539764, | |
| "learning_rate": 8.745082927659047e-05, | |
| "loss": 0.0177, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 1.7110266159695817, | |
| "grad_norm": 0.2368573248386383, | |
| "learning_rate": 8.73410738492077e-05, | |
| "loss": 0.02, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 1.717363751584284, | |
| "grad_norm": 0.3750855326652527, | |
| "learning_rate": 8.723091006582389e-05, | |
| "loss": 0.0157, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 1.723700887198986, | |
| "grad_norm": 0.468447744846344, | |
| "learning_rate": 8.71203391311725e-05, | |
| "loss": 0.0212, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 1.7300380228136882, | |
| "grad_norm": 0.32794639468193054, | |
| "learning_rate": 8.700936225443959e-05, | |
| "loss": 0.015, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 1.7363751584283904, | |
| "grad_norm": 0.5568444132804871, | |
| "learning_rate": 8.689798064925049e-05, | |
| "loss": 0.0182, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 1.7427122940430926, | |
| "grad_norm": 0.36910462379455566, | |
| "learning_rate": 8.678619553365659e-05, | |
| "loss": 0.0213, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 1.7490494296577945, | |
| "grad_norm": 0.21325695514678955, | |
| "learning_rate": 8.6674008130122e-05, | |
| "loss": 0.0195, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 1.755386565272497, | |
| "grad_norm": 0.22268667817115784, | |
| "learning_rate": 8.656141966551019e-05, | |
| "loss": 0.0192, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 1.7617237008871989, | |
| "grad_norm": 0.5589450597763062, | |
| "learning_rate": 8.644843137107059e-05, | |
| "loss": 0.0227, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 1.7680608365019013, | |
| "grad_norm": 0.22853608429431915, | |
| "learning_rate": 8.633504448242505e-05, | |
| "loss": 0.0234, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 1.7743979721166032, | |
| "grad_norm": 0.3304268419742584, | |
| "learning_rate": 8.622126023955446e-05, | |
| "loss": 0.0202, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 1.7807351077313056, | |
| "grad_norm": 0.30013397336006165, | |
| "learning_rate": 8.610707988678503e-05, | |
| "loss": 0.0189, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 1.7870722433460076, | |
| "grad_norm": 0.3276195228099823, | |
| "learning_rate": 8.599250467277483e-05, | |
| "loss": 0.0174, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 1.7934093789607097, | |
| "grad_norm": 0.37568747997283936, | |
| "learning_rate": 8.587753585050004e-05, | |
| "loss": 0.0217, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 1.799746514575412, | |
| "grad_norm": 0.6022403836250305, | |
| "learning_rate": 8.576217467724128e-05, | |
| "loss": 0.0253, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 1.806083650190114, | |
| "grad_norm": 0.3828267455101013, | |
| "learning_rate": 8.564642241456986e-05, | |
| "loss": 0.0219, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 1.8124207858048162, | |
| "grad_norm": 0.37504634261131287, | |
| "learning_rate": 8.553028032833397e-05, | |
| "loss": 0.021, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 1.8187579214195184, | |
| "grad_norm": 0.6533935070037842, | |
| "learning_rate": 8.541374968864487e-05, | |
| "loss": 0.027, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 1.8250950570342206, | |
| "grad_norm": 0.4715104103088379, | |
| "learning_rate": 8.529683176986295e-05, | |
| "loss": 0.0165, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 1.8314321926489225, | |
| "grad_norm": 0.5481289029121399, | |
| "learning_rate": 8.517952785058385e-05, | |
| "loss": 0.0253, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 1.837769328263625, | |
| "grad_norm": 0.3077695667743683, | |
| "learning_rate": 8.506183921362443e-05, | |
| "loss": 0.0192, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 1.8441064638783269, | |
| "grad_norm": 0.5000676512718201, | |
| "learning_rate": 8.494376714600878e-05, | |
| "loss": 0.0234, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 1.8504435994930293, | |
| "grad_norm": 0.253926545381546, | |
| "learning_rate": 8.482531293895412e-05, | |
| "loss": 0.0225, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 1.8567807351077312, | |
| "grad_norm": 0.3240850269794464, | |
| "learning_rate": 8.470647788785665e-05, | |
| "loss": 0.0185, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 1.8631178707224336, | |
| "grad_norm": 0.4573562741279602, | |
| "learning_rate": 8.458726329227747e-05, | |
| "loss": 0.0218, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 1.8694550063371356, | |
| "grad_norm": 0.443369597196579, | |
| "learning_rate": 8.44676704559283e-05, | |
| "loss": 0.0244, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 1.8757921419518377, | |
| "grad_norm": 0.3431839942932129, | |
| "learning_rate": 8.434770068665723e-05, | |
| "loss": 0.0253, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 1.88212927756654, | |
| "grad_norm": 0.37696412205696106, | |
| "learning_rate": 8.422735529643444e-05, | |
| "loss": 0.0202, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 1.888466413181242, | |
| "grad_norm": 0.6237473487854004, | |
| "learning_rate": 8.410663560133784e-05, | |
| "loss": 0.0197, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 1.8948035487959443, | |
| "grad_norm": 0.39705732464790344, | |
| "learning_rate": 8.398554292153866e-05, | |
| "loss": 0.0239, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 1.9011406844106464, | |
| "grad_norm": 0.37278616428375244, | |
| "learning_rate": 8.386407858128706e-05, | |
| "loss": 0.0289, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 1.9074778200253486, | |
| "grad_norm": 0.3443812131881714, | |
| "learning_rate": 8.37422439088976e-05, | |
| "loss": 0.0217, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 1.9138149556400506, | |
| "grad_norm": 0.28766191005706787, | |
| "learning_rate": 8.362004023673474e-05, | |
| "loss": 0.024, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 1.920152091254753, | |
| "grad_norm": 0.17464351654052734, | |
| "learning_rate": 8.349746890119826e-05, | |
| "loss": 0.0201, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 1.926489226869455, | |
| "grad_norm": 0.37782022356987, | |
| "learning_rate": 8.337453124270863e-05, | |
| "loss": 0.0225, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 1.9328263624841573, | |
| "grad_norm": 0.28459540009498596, | |
| "learning_rate": 8.32512286056924e-05, | |
| "loss": 0.0247, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 1.9391634980988592, | |
| "grad_norm": 0.3449212610721588, | |
| "learning_rate": 8.31275623385675e-05, | |
| "loss": 0.0185, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 1.9455006337135616, | |
| "grad_norm": 0.451109379529953, | |
| "learning_rate": 8.300353379372834e-05, | |
| "loss": 0.0144, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 1.9518377693282636, | |
| "grad_norm": 0.44775041937828064, | |
| "learning_rate": 8.287914432753123e-05, | |
| "loss": 0.0239, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 1.9581749049429658, | |
| "grad_norm": 0.3843693435192108, | |
| "learning_rate": 8.275439530027948e-05, | |
| "loss": 0.0205, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 1.964512040557668, | |
| "grad_norm": 0.32762160897254944, | |
| "learning_rate": 8.262928807620843e-05, | |
| "loss": 0.0222, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 1.97084917617237, | |
| "grad_norm": 0.323930025100708, | |
| "learning_rate": 8.250382402347065e-05, | |
| "loss": 0.0191, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 1.9771863117870723, | |
| "grad_norm": 0.26865267753601074, | |
| "learning_rate": 8.237800451412095e-05, | |
| "loss": 0.0205, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 1.9835234474017744, | |
| "grad_norm": 0.27057474851608276, | |
| "learning_rate": 8.225183092410128e-05, | |
| "loss": 0.0181, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 1.9898605830164766, | |
| "grad_norm": 0.38607391715049744, | |
| "learning_rate": 8.212530463322583e-05, | |
| "loss": 0.0206, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 1.9961977186311786, | |
| "grad_norm": 0.3551093637943268, | |
| "learning_rate": 8.199842702516583e-05, | |
| "loss": 0.0226, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 2.002534854245881, | |
| "grad_norm": 0.3289443552494049, | |
| "learning_rate": 8.18711994874345e-05, | |
| "loss": 0.0159, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 2.008871989860583, | |
| "grad_norm": 0.29715219140052795, | |
| "learning_rate": 8.174362341137177e-05, | |
| "loss": 0.0188, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 2.0152091254752853, | |
| "grad_norm": 0.25433430075645447, | |
| "learning_rate": 8.161570019212921e-05, | |
| "loss": 0.0185, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 2.0215462610899873, | |
| "grad_norm": 0.2330443561077118, | |
| "learning_rate": 8.148743122865463e-05, | |
| "loss": 0.016, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 2.0278833967046896, | |
| "grad_norm": 0.3372531235218048, | |
| "learning_rate": 8.135881792367686e-05, | |
| "loss": 0.0216, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 2.0342205323193916, | |
| "grad_norm": 0.24772562086582184, | |
| "learning_rate": 8.12298616836904e-05, | |
| "loss": 0.0156, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 2.040557667934094, | |
| "grad_norm": 0.3528910279273987, | |
| "learning_rate": 8.110056391894005e-05, | |
| "loss": 0.0149, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 2.046894803548796, | |
| "grad_norm": 0.2773556113243103, | |
| "learning_rate": 8.097092604340542e-05, | |
| "loss": 0.0172, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 2.053231939163498, | |
| "grad_norm": 0.39497804641723633, | |
| "learning_rate": 8.084094947478556e-05, | |
| "loss": 0.0193, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 2.0595690747782003, | |
| "grad_norm": 0.2211255133152008, | |
| "learning_rate": 8.07106356344834e-05, | |
| "loss": 0.0155, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 2.0659062103929022, | |
| "grad_norm": 0.3327501118183136, | |
| "learning_rate": 8.057998594759022e-05, | |
| "loss": 0.0154, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 2.0722433460076046, | |
| "grad_norm": 0.3289473056793213, | |
| "learning_rate": 8.044900184287007e-05, | |
| "loss": 0.0199, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 2.0785804816223066, | |
| "grad_norm": 0.36586564779281616, | |
| "learning_rate": 8.031768475274413e-05, | |
| "loss": 0.0164, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 2.084917617237009, | |
| "grad_norm": 0.3189631402492523, | |
| "learning_rate": 8.018603611327504e-05, | |
| "loss": 0.0168, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 2.091254752851711, | |
| "grad_norm": 0.36633414030075073, | |
| "learning_rate": 8.005405736415126e-05, | |
| "loss": 0.0317, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 2.0975918884664133, | |
| "grad_norm": 0.2375175505876541, | |
| "learning_rate": 7.992174994867123e-05, | |
| "loss": 0.0123, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 2.1039290240811153, | |
| "grad_norm": 0.39194151759147644, | |
| "learning_rate": 7.978911531372765e-05, | |
| "loss": 0.0148, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 2.1102661596958177, | |
| "grad_norm": 0.4107789099216461, | |
| "learning_rate": 7.965615490979163e-05, | |
| "loss": 0.0205, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 2.1166032953105196, | |
| "grad_norm": 0.3235548436641693, | |
| "learning_rate": 7.952287019089685e-05, | |
| "loss": 0.0159, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 2.122940430925222, | |
| "grad_norm": 0.2603886127471924, | |
| "learning_rate": 7.938926261462366e-05, | |
| "loss": 0.0188, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 2.129277566539924, | |
| "grad_norm": 0.387168288230896, | |
| "learning_rate": 7.925533364208309e-05, | |
| "loss": 0.0223, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 2.1356147021546263, | |
| "grad_norm": 0.4185580015182495, | |
| "learning_rate": 7.912108473790092e-05, | |
| "loss": 0.0242, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 2.1419518377693283, | |
| "grad_norm": 0.25904420018196106, | |
| "learning_rate": 7.898651737020166e-05, | |
| "loss": 0.0169, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 2.1482889733840302, | |
| "grad_norm": 0.29916971921920776, | |
| "learning_rate": 7.88516330105925e-05, | |
| "loss": 0.0155, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 2.1546261089987326, | |
| "grad_norm": 0.3949578106403351, | |
| "learning_rate": 7.871643313414718e-05, | |
| "loss": 0.018, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 2.1609632446134346, | |
| "grad_norm": 0.31313779950141907, | |
| "learning_rate": 7.858091921938988e-05, | |
| "loss": 0.0139, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 2.167300380228137, | |
| "grad_norm": 0.31156864762306213, | |
| "learning_rate": 7.844509274827907e-05, | |
| "loss": 0.0231, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 2.173637515842839, | |
| "grad_norm": 0.27319496870040894, | |
| "learning_rate": 7.830895520619128e-05, | |
| "loss": 0.0173, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 2.1799746514575413, | |
| "grad_norm": 0.5082638263702393, | |
| "learning_rate": 7.817250808190483e-05, | |
| "loss": 0.0206, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 2.1863117870722433, | |
| "grad_norm": 0.3610362410545349, | |
| "learning_rate": 7.803575286758364e-05, | |
| "loss": 0.0187, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 2.1926489226869457, | |
| "grad_norm": 0.6687695384025574, | |
| "learning_rate": 7.789869105876083e-05, | |
| "loss": 0.0188, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 2.1989860583016476, | |
| "grad_norm": 0.5592470765113831, | |
| "learning_rate": 7.776132415432234e-05, | |
| "loss": 0.0178, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 2.20532319391635, | |
| "grad_norm": 0.27790650725364685, | |
| "learning_rate": 7.762365365649067e-05, | |
| "loss": 0.0171, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 2.211660329531052, | |
| "grad_norm": 0.4534870684146881, | |
| "learning_rate": 7.748568107080832e-05, | |
| "loss": 0.0219, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 2.2179974651457544, | |
| "grad_norm": 0.33770209550857544, | |
| "learning_rate": 7.734740790612136e-05, | |
| "loss": 0.0184, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 2.2243346007604563, | |
| "grad_norm": 0.29577895998954773, | |
| "learning_rate": 7.720883567456298e-05, | |
| "loss": 0.0194, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 2.2306717363751583, | |
| "grad_norm": 0.45840632915496826, | |
| "learning_rate": 7.70699658915369e-05, | |
| "loss": 0.023, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 2.2370088719898606, | |
| "grad_norm": 0.2918465733528137, | |
| "learning_rate": 7.693080007570084e-05, | |
| "loss": 0.0176, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 2.2433460076045626, | |
| "grad_norm": 0.3843630254268646, | |
| "learning_rate": 7.679133974894983e-05, | |
| "loss": 0.0265, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 2.249683143219265, | |
| "grad_norm": 0.24515217542648315, | |
| "learning_rate": 7.66515864363997e-05, | |
| "loss": 0.0174, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 2.256020278833967, | |
| "grad_norm": 0.19862592220306396, | |
| "learning_rate": 7.651154166637025e-05, | |
| "loss": 0.0215, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 2.2623574144486693, | |
| "grad_norm": 0.17352795600891113, | |
| "learning_rate": 7.637120697036866e-05, | |
| "loss": 0.0278, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 2.2686945500633713, | |
| "grad_norm": 0.34404289722442627, | |
| "learning_rate": 7.623058388307269e-05, | |
| "loss": 0.0208, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 2.2750316856780737, | |
| "grad_norm": 0.2361377477645874, | |
| "learning_rate": 7.608967394231387e-05, | |
| "loss": 0.0215, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 2.2813688212927756, | |
| "grad_norm": 0.24867765605449677, | |
| "learning_rate": 7.594847868906076e-05, | |
| "loss": 0.0197, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 2.2877059569074776, | |
| "grad_norm": 0.33056747913360596, | |
| "learning_rate": 7.580699966740201e-05, | |
| "loss": 0.0206, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 2.29404309252218, | |
| "grad_norm": 0.4908096194267273, | |
| "learning_rate": 7.566523842452958e-05, | |
| "loss": 0.0193, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 2.3003802281368824, | |
| "grad_norm": 0.3107684254646301, | |
| "learning_rate": 7.552319651072164e-05, | |
| "loss": 0.0179, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 2.3067173637515843, | |
| "grad_norm": 0.2108030468225479, | |
| "learning_rate": 7.538087547932585e-05, | |
| "loss": 0.0167, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 2.3130544993662863, | |
| "grad_norm": 0.3508223295211792, | |
| "learning_rate": 7.52382768867422e-05, | |
| "loss": 0.0226, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 2.3193916349809887, | |
| "grad_norm": 0.33375847339630127, | |
| "learning_rate": 7.509540229240601e-05, | |
| "loss": 0.0159, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 2.3257287705956906, | |
| "grad_norm": 0.3554519712924957, | |
| "learning_rate": 7.495225325877103e-05, | |
| "loss": 0.0182, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 2.332065906210393, | |
| "grad_norm": 0.38898682594299316, | |
| "learning_rate": 7.480883135129211e-05, | |
| "loss": 0.0145, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 2.338403041825095, | |
| "grad_norm": 0.33638787269592285, | |
| "learning_rate": 7.466513813840825e-05, | |
| "loss": 0.0211, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 2.3447401774397973, | |
| "grad_norm": 0.40711718797683716, | |
| "learning_rate": 7.452117519152542e-05, | |
| "loss": 0.0196, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 2.3510773130544993, | |
| "grad_norm": 0.3523763418197632, | |
| "learning_rate": 7.437694408499933e-05, | |
| "loss": 0.0216, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 2.3574144486692017, | |
| "grad_norm": 0.2812027633190155, | |
| "learning_rate": 7.423244639611826e-05, | |
| "loss": 0.0207, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 2.3637515842839036, | |
| "grad_norm": 0.2593352496623993, | |
| "learning_rate": 7.408768370508576e-05, | |
| "loss": 0.0198, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 2.3700887198986056, | |
| "grad_norm": 0.5507994890213013, | |
| "learning_rate": 7.394265759500348e-05, | |
| "loss": 0.0249, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 2.376425855513308, | |
| "grad_norm": 0.33799615502357483, | |
| "learning_rate": 7.379736965185368e-05, | |
| "loss": 0.0222, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 2.3827629911280104, | |
| "grad_norm": 0.3209291398525238, | |
| "learning_rate": 7.365182146448205e-05, | |
| "loss": 0.02, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 2.3891001267427123, | |
| "grad_norm": 0.34429389238357544, | |
| "learning_rate": 7.350601462458024e-05, | |
| "loss": 0.0156, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 2.3954372623574143, | |
| "grad_norm": 0.29221874475479126, | |
| "learning_rate": 7.335995072666848e-05, | |
| "loss": 0.0141, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 2.4017743979721167, | |
| "grad_norm": 0.37552404403686523, | |
| "learning_rate": 7.32136313680782e-05, | |
| "loss": 0.0193, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 2.4081115335868186, | |
| "grad_norm": 0.25026583671569824, | |
| "learning_rate": 7.30670581489344e-05, | |
| "loss": 0.0215, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 2.414448669201521, | |
| "grad_norm": 0.3728311061859131, | |
| "learning_rate": 7.292023267213835e-05, | |
| "loss": 0.0173, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 2.420785804816223, | |
| "grad_norm": 0.35615667700767517, | |
| "learning_rate": 7.277315654334997e-05, | |
| "loss": 0.021, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 2.4271229404309254, | |
| "grad_norm": 0.25696536898612976, | |
| "learning_rate": 7.262583137097018e-05, | |
| "loss": 0.018, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 2.4334600760456273, | |
| "grad_norm": 0.3925075829029083, | |
| "learning_rate": 7.247825876612353e-05, | |
| "loss": 0.0168, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 2.4397972116603297, | |
| "grad_norm": 0.2698250710964203, | |
| "learning_rate": 7.233044034264034e-05, | |
| "loss": 0.0178, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 2.4461343472750317, | |
| "grad_norm": 0.29354268312454224, | |
| "learning_rate": 7.218237771703921e-05, | |
| "loss": 0.0183, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 2.4524714828897336, | |
| "grad_norm": 0.3076222240924835, | |
| "learning_rate": 7.203407250850928e-05, | |
| "loss": 0.0167, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 2.458808618504436, | |
| "grad_norm": 0.23654355108737946, | |
| "learning_rate": 7.188552633889259e-05, | |
| "loss": 0.0165, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 2.4651457541191384, | |
| "grad_norm": 0.2536214292049408, | |
| "learning_rate": 7.173674083266624e-05, | |
| "loss": 0.0138, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 2.4714828897338403, | |
| "grad_norm": 0.2710602879524231, | |
| "learning_rate": 7.158771761692464e-05, | |
| "loss": 0.0148, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 2.4778200253485423, | |
| "grad_norm": 0.15324455499649048, | |
| "learning_rate": 7.143845832136188e-05, | |
| "loss": 0.0142, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 2.4841571609632447, | |
| "grad_norm": 0.13363996148109436, | |
| "learning_rate": 7.128896457825364e-05, | |
| "loss": 0.0144, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 2.4904942965779466, | |
| "grad_norm": 0.2499360889196396, | |
| "learning_rate": 7.113923802243957e-05, | |
| "loss": 0.0177, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 2.496831432192649, | |
| "grad_norm": 0.1742558777332306, | |
| "learning_rate": 7.09892802913053e-05, | |
| "loss": 0.0161, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 2.503168567807351, | |
| "grad_norm": 0.33764421939849854, | |
| "learning_rate": 7.083909302476453e-05, | |
| "loss": 0.0189, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 2.5095057034220534, | |
| "grad_norm": 0.35038939118385315, | |
| "learning_rate": 7.068867786524116e-05, | |
| "loss": 0.0236, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 2.5158428390367553, | |
| "grad_norm": 0.16515985131263733, | |
| "learning_rate": 7.053803645765128e-05, | |
| "loss": 0.0169, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 2.5221799746514577, | |
| "grad_norm": 0.49210289120674133, | |
| "learning_rate": 7.038717044938519e-05, | |
| "loss": 0.0188, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 2.5285171102661597, | |
| "grad_norm": 0.2284458875656128, | |
| "learning_rate": 7.023608149028937e-05, | |
| "loss": 0.0159, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 2.5348542458808616, | |
| "grad_norm": 0.19401752948760986, | |
| "learning_rate": 7.008477123264848e-05, | |
| "loss": 0.0168, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 2.541191381495564, | |
| "grad_norm": 0.14309851825237274, | |
| "learning_rate": 6.993324133116726e-05, | |
| "loss": 0.0118, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 2.5475285171102664, | |
| "grad_norm": 0.22239387035369873, | |
| "learning_rate": 6.978149344295242e-05, | |
| "loss": 0.0141, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 2.5538656527249683, | |
| "grad_norm": 0.4320535957813263, | |
| "learning_rate": 6.962952922749457e-05, | |
| "loss": 0.0178, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 2.5602027883396703, | |
| "grad_norm": 0.32748010754585266, | |
| "learning_rate": 6.947735034665002e-05, | |
| "loss": 0.0137, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 2.5665399239543727, | |
| "grad_norm": 0.1492803394794464, | |
| "learning_rate": 6.932495846462261e-05, | |
| "loss": 0.0219, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 2.5728770595690746, | |
| "grad_norm": 0.3322237432003021, | |
| "learning_rate": 6.917235524794558e-05, | |
| "loss": 0.0188, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 2.579214195183777, | |
| "grad_norm": 0.40152451395988464, | |
| "learning_rate": 6.901954236546323e-05, | |
| "loss": 0.0263, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 2.585551330798479, | |
| "grad_norm": 0.43540433049201965, | |
| "learning_rate": 6.886652148831279e-05, | |
| "loss": 0.0146, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 2.5918884664131814, | |
| "grad_norm": 0.24758180975914001, | |
| "learning_rate": 6.871329428990602e-05, | |
| "loss": 0.0146, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 2.5982256020278833, | |
| "grad_norm": 0.34365662932395935, | |
| "learning_rate": 6.855986244591104e-05, | |
| "loss": 0.0166, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 2.6045627376425857, | |
| "grad_norm": 0.2038048356771469, | |
| "learning_rate": 6.840622763423391e-05, | |
| "loss": 0.0151, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 2.6108998732572877, | |
| "grad_norm": 0.182295024394989, | |
| "learning_rate": 6.825239153500029e-05, | |
| "loss": 0.0145, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 2.6172370088719896, | |
| "grad_norm": 0.3425464928150177, | |
| "learning_rate": 6.809835583053715e-05, | |
| "loss": 0.0121, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 2.623574144486692, | |
| "grad_norm": 0.2870509624481201, | |
| "learning_rate": 6.794412220535426e-05, | |
| "loss": 0.0127, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 2.6299112801013944, | |
| "grad_norm": 0.34516963362693787, | |
| "learning_rate": 6.778969234612584e-05, | |
| "loss": 0.0158, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 2.6362484157160964, | |
| "grad_norm": 0.27682995796203613, | |
| "learning_rate": 6.763506794167208e-05, | |
| "loss": 0.0132, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 2.6425855513307983, | |
| "grad_norm": 0.3806769549846649, | |
| "learning_rate": 6.748025068294067e-05, | |
| "loss": 0.0158, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 2.6489226869455007, | |
| "grad_norm": 0.20911303162574768, | |
| "learning_rate": 6.732524226298841e-05, | |
| "loss": 0.0164, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 2.6552598225602027, | |
| "grad_norm": 0.263217568397522, | |
| "learning_rate": 6.71700443769625e-05, | |
| "loss": 0.0174, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 2.661596958174905, | |
| "grad_norm": 0.17768670618534088, | |
| "learning_rate": 6.701465872208216e-05, | |
| "loss": 0.0156, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 2.667934093789607, | |
| "grad_norm": 0.30574101209640503, | |
| "learning_rate": 6.685908699762002e-05, | |
| "loss": 0.0127, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 2.6742712294043094, | |
| "grad_norm": 0.2742626965045929, | |
| "learning_rate": 6.670333090488356e-05, | |
| "loss": 0.0168, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 2.6806083650190113, | |
| "grad_norm": 0.24799436330795288, | |
| "learning_rate": 6.654739214719641e-05, | |
| "loss": 0.0141, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 2.6869455006337137, | |
| "grad_norm": 0.33699798583984375, | |
| "learning_rate": 6.639127242987988e-05, | |
| "loss": 0.0115, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 2.6932826362484157, | |
| "grad_norm": 0.4366147816181183, | |
| "learning_rate": 6.623497346023418e-05, | |
| "loss": 0.0201, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 2.6996197718631176, | |
| "grad_norm": 0.29953205585479736, | |
| "learning_rate": 6.607849694751977e-05, | |
| "loss": 0.0143, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 2.70595690747782, | |
| "grad_norm": 0.33452221751213074, | |
| "learning_rate": 6.592184460293877e-05, | |
| "loss": 0.0144, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 2.7122940430925224, | |
| "grad_norm": 0.2584233283996582, | |
| "learning_rate": 6.576501813961609e-05, | |
| "loss": 0.0161, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 2.7186311787072244, | |
| "grad_norm": 0.3979489505290985, | |
| "learning_rate": 6.56080192725808e-05, | |
| "loss": 0.0282, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 2.7249683143219263, | |
| "grad_norm": 0.4084644615650177, | |
| "learning_rate": 6.545084971874738e-05, | |
| "loss": 0.0155, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 2.7313054499366287, | |
| "grad_norm": 0.3948853611946106, | |
| "learning_rate": 6.529351119689688e-05, | |
| "loss": 0.014, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 2.7376425855513307, | |
| "grad_norm": 0.3730246424674988, | |
| "learning_rate": 6.513600542765817e-05, | |
| "loss": 0.0205, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 2.743979721166033, | |
| "grad_norm": 0.30106091499328613, | |
| "learning_rate": 6.497833413348909e-05, | |
| "loss": 0.0162, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 2.750316856780735, | |
| "grad_norm": 0.14317883551120758, | |
| "learning_rate": 6.48204990386577e-05, | |
| "loss": 0.0134, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 2.7566539923954374, | |
| "grad_norm": 0.33047810196876526, | |
| "learning_rate": 6.466250186922325e-05, | |
| "loss": 0.0138, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 2.7629911280101394, | |
| "grad_norm": 0.29450923204421997, | |
| "learning_rate": 6.450434435301751e-05, | |
| "loss": 0.0166, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 2.7693282636248417, | |
| "grad_norm": 0.22328028082847595, | |
| "learning_rate": 6.43460282196257e-05, | |
| "loss": 0.0159, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 2.7756653992395437, | |
| "grad_norm": 0.170170858502388, | |
| "learning_rate": 6.418755520036775e-05, | |
| "loss": 0.0147, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 2.7820025348542456, | |
| "grad_norm": 0.3056323826313019, | |
| "learning_rate": 6.402892702827916e-05, | |
| "loss": 0.0193, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 2.788339670468948, | |
| "grad_norm": 0.19363383948802948, | |
| "learning_rate": 6.387014543809223e-05, | |
| "loss": 0.0134, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 2.7946768060836504, | |
| "grad_norm": 0.4496561586856842, | |
| "learning_rate": 6.371121216621698e-05, | |
| "loss": 0.0132, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 2.8010139416983524, | |
| "grad_norm": 0.17226959764957428, | |
| "learning_rate": 6.355212895072223e-05, | |
| "loss": 0.0119, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 2.8073510773130543, | |
| "grad_norm": 0.2828828692436218, | |
| "learning_rate": 6.339289753131649e-05, | |
| "loss": 0.016, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 2.8136882129277567, | |
| "grad_norm": 0.28239116072654724, | |
| "learning_rate": 6.323351964932908e-05, | |
| "loss": 0.015, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 2.8200253485424587, | |
| "grad_norm": 0.2226468324661255, | |
| "learning_rate": 6.307399704769099e-05, | |
| "loss": 0.0135, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 2.826362484157161, | |
| "grad_norm": 0.23484410345554352, | |
| "learning_rate": 6.291433147091583e-05, | |
| "loss": 0.015, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 2.832699619771863, | |
| "grad_norm": 0.2702759802341461, | |
| "learning_rate": 6.275452466508077e-05, | |
| "loss": 0.016, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 2.8390367553865654, | |
| "grad_norm": 0.39190873503685, | |
| "learning_rate": 6.259457837780742e-05, | |
| "loss": 0.0116, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 2.8453738910012674, | |
| "grad_norm": 0.4797994792461395, | |
| "learning_rate": 6.243449435824276e-05, | |
| "loss": 0.0148, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 2.8517110266159698, | |
| "grad_norm": 0.29361021518707275, | |
| "learning_rate": 6.227427435703997e-05, | |
| "loss": 0.025, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 2.8580481622306717, | |
| "grad_norm": 0.26431670784950256, | |
| "learning_rate": 6.211392012633932e-05, | |
| "loss": 0.0116, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 2.8643852978453737, | |
| "grad_norm": 0.2542441189289093, | |
| "learning_rate": 6.195343341974899e-05, | |
| "loss": 0.0191, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 2.870722433460076, | |
| "grad_norm": 0.21877507865428925, | |
| "learning_rate": 6.179281599232591e-05, | |
| "loss": 0.016, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 2.8770595690747784, | |
| "grad_norm": 0.3358696401119232, | |
| "learning_rate": 6.163206960055651e-05, | |
| "loss": 0.0152, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 2.8833967046894804, | |
| "grad_norm": 0.3265288174152374, | |
| "learning_rate": 6.147119600233758e-05, | |
| "loss": 0.0199, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 2.8897338403041823, | |
| "grad_norm": 0.3462512195110321, | |
| "learning_rate": 6.131019695695702e-05, | |
| "loss": 0.0141, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 2.8960709759188847, | |
| "grad_norm": 0.24054935574531555, | |
| "learning_rate": 6.11490742250746e-05, | |
| "loss": 0.0133, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 2.9024081115335867, | |
| "grad_norm": 0.2769206166267395, | |
| "learning_rate": 6.0987829568702656e-05, | |
| "loss": 0.0177, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 2.908745247148289, | |
| "grad_norm": 0.23850680887699127, | |
| "learning_rate": 6.0826464751186994e-05, | |
| "loss": 0.0112, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 2.915082382762991, | |
| "grad_norm": 0.20703135430812836, | |
| "learning_rate": 6.066498153718735e-05, | |
| "loss": 0.0117, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 2.9214195183776934, | |
| "grad_norm": 0.22506844997406006, | |
| "learning_rate": 6.05033816926583e-05, | |
| "loss": 0.0162, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 2.9277566539923954, | |
| "grad_norm": 0.30291593074798584, | |
| "learning_rate": 6.034166698482984e-05, | |
| "loss": 0.0147, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 2.9340937896070978, | |
| "grad_norm": 0.24053440988063812, | |
| "learning_rate": 6.017983918218812e-05, | |
| "loss": 0.0091, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 2.9404309252217997, | |
| "grad_norm": 0.27635499835014343, | |
| "learning_rate": 6.001790005445607e-05, | |
| "loss": 0.0154, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 2.9467680608365017, | |
| "grad_norm": 0.29039266705513, | |
| "learning_rate": 5.985585137257401e-05, | |
| "loss": 0.0165, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 2.953105196451204, | |
| "grad_norm": 0.1605728417634964, | |
| "learning_rate": 5.969369490868042e-05, | |
| "loss": 0.0154, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 2.9594423320659065, | |
| "grad_norm": 0.2104729562997818, | |
| "learning_rate": 5.953143243609235e-05, | |
| "loss": 0.0224, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 2.9657794676806084, | |
| "grad_norm": 0.21257384121418, | |
| "learning_rate": 5.9369065729286245e-05, | |
| "loss": 0.0118, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 2.9721166032953104, | |
| "grad_norm": 0.2738170027732849, | |
| "learning_rate": 5.9206596563878357e-05, | |
| "loss": 0.0183, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 2.9784537389100127, | |
| "grad_norm": 0.2844460904598236, | |
| "learning_rate": 5.90440267166055e-05, | |
| "loss": 0.0156, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 2.9847908745247147, | |
| "grad_norm": 0.3566923439502716, | |
| "learning_rate": 5.888135796530544e-05, | |
| "loss": 0.0134, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 2.991128010139417, | |
| "grad_norm": 0.1965407133102417, | |
| "learning_rate": 5.871859208889759e-05, | |
| "loss": 0.013, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 2.997465145754119, | |
| "grad_norm": 0.18356238305568695, | |
| "learning_rate": 5.85557308673635e-05, | |
| "loss": 0.0113, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 3.0038022813688214, | |
| "grad_norm": 0.23230914771556854, | |
| "learning_rate": 5.8392776081727385e-05, | |
| "loss": 0.0164, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 3.0101394169835234, | |
| "grad_norm": 0.27572014927864075, | |
| "learning_rate": 5.8229729514036705e-05, | |
| "loss": 0.0113, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 3.016476552598226, | |
| "grad_norm": 0.36763879656791687, | |
| "learning_rate": 5.8066592947342555e-05, | |
| "loss": 0.0124, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 3.0228136882129277, | |
| "grad_norm": 0.18714648485183716, | |
| "learning_rate": 5.7903368165680327e-05, | |
| "loss": 0.0143, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 3.02915082382763, | |
| "grad_norm": 0.22437100112438202, | |
| "learning_rate": 5.7740056954050084e-05, | |
| "loss": 0.0176, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 3.035487959442332, | |
| "grad_norm": 0.33260634541511536, | |
| "learning_rate": 5.757666109839702e-05, | |
| "loss": 0.0195, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 3.041825095057034, | |
| "grad_norm": 0.29105865955352783, | |
| "learning_rate": 5.74131823855921e-05, | |
| "loss": 0.015, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 3.0481622306717364, | |
| "grad_norm": 0.20448502898216248, | |
| "learning_rate": 5.72496226034123e-05, | |
| "loss": 0.0141, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 3.0544993662864384, | |
| "grad_norm": 0.2469986230134964, | |
| "learning_rate": 5.7085983540521216e-05, | |
| "loss": 0.0117, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 3.0608365019011408, | |
| "grad_norm": 0.2351512908935547, | |
| "learning_rate": 5.692226698644938e-05, | |
| "loss": 0.0138, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 3.0671736375158427, | |
| "grad_norm": 0.2490169107913971, | |
| "learning_rate": 5.675847473157485e-05, | |
| "loss": 0.0144, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 3.073510773130545, | |
| "grad_norm": 0.3461778461933136, | |
| "learning_rate": 5.6594608567103456e-05, | |
| "loss": 0.0126, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 3.079847908745247, | |
| "grad_norm": 0.28383558988571167, | |
| "learning_rate": 5.6430670285049314e-05, | |
| "loss": 0.0117, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 3.0861850443599494, | |
| "grad_norm": 0.25864022970199585, | |
| "learning_rate": 5.6266661678215216e-05, | |
| "loss": 0.0146, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 3.0925221799746514, | |
| "grad_norm": 0.4212886095046997, | |
| "learning_rate": 5.6102584540173006e-05, | |
| "loss": 0.0135, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 3.098859315589354, | |
| "grad_norm": 0.2258593887090683, | |
| "learning_rate": 5.5938440665244006e-05, | |
| "loss": 0.0164, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 3.1051964512040557, | |
| "grad_norm": 0.3095914423465729, | |
| "learning_rate": 5.577423184847932e-05, | |
| "loss": 0.0134, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 3.111533586818758, | |
| "grad_norm": 0.14614951610565186, | |
| "learning_rate": 5.560995988564023e-05, | |
| "loss": 0.0131, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 3.11787072243346, | |
| "grad_norm": 0.30523428320884705, | |
| "learning_rate": 5.544562657317863e-05, | |
| "loss": 0.015, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 3.124207858048162, | |
| "grad_norm": 0.17284521460533142, | |
| "learning_rate": 5.52812337082173e-05, | |
| "loss": 0.0126, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 3.1305449936628644, | |
| "grad_norm": 0.33044707775115967, | |
| "learning_rate": 5.511678308853026e-05, | |
| "loss": 0.0142, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 3.1368821292775664, | |
| "grad_norm": 0.313587486743927, | |
| "learning_rate": 5.495227651252315e-05, | |
| "loss": 0.011, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 3.1432192648922688, | |
| "grad_norm": 0.3139721751213074, | |
| "learning_rate": 5.478771577921351e-05, | |
| "loss": 0.0132, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 3.1495564005069707, | |
| "grad_norm": 0.313909649848938, | |
| "learning_rate": 5.462310268821118e-05, | |
| "loss": 0.0129, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 3.155893536121673, | |
| "grad_norm": 0.17512694001197815, | |
| "learning_rate": 5.445843903969854e-05, | |
| "loss": 0.0102, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 3.162230671736375, | |
| "grad_norm": 0.21401318907737732, | |
| "learning_rate": 5.4293726634410855e-05, | |
| "loss": 0.012, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 3.1685678073510775, | |
| "grad_norm": 0.15441134572029114, | |
| "learning_rate": 5.4128967273616625e-05, | |
| "loss": 0.0119, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 3.1749049429657794, | |
| "grad_norm": 0.23418472707271576, | |
| "learning_rate": 5.396416275909779e-05, | |
| "loss": 0.015, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 3.181242078580482, | |
| "grad_norm": 0.2487867772579193, | |
| "learning_rate": 5.379931489313016e-05, | |
| "loss": 0.0133, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 3.1875792141951838, | |
| "grad_norm": 0.2844524681568146, | |
| "learning_rate": 5.363442547846356e-05, | |
| "loss": 0.014, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 3.1939163498098857, | |
| "grad_norm": 0.20346881449222565, | |
| "learning_rate": 5.3469496318302204e-05, | |
| "loss": 0.0114, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 3.200253485424588, | |
| "grad_norm": 0.3049124479293823, | |
| "learning_rate": 5.330452921628497e-05, | |
| "loss": 0.0153, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 3.20659062103929, | |
| "grad_norm": 0.29535266757011414, | |
| "learning_rate": 5.313952597646568e-05, | |
| "loss": 0.0132, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 3.2129277566539924, | |
| "grad_norm": 0.3959789276123047, | |
| "learning_rate": 5.297448840329329e-05, | |
| "loss": 0.0157, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 3.2192648922686944, | |
| "grad_norm": 0.23339158296585083, | |
| "learning_rate": 5.280941830159227e-05, | |
| "loss": 0.0128, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 3.225602027883397, | |
| "grad_norm": 0.2398107349872589, | |
| "learning_rate": 5.264431747654284e-05, | |
| "loss": 0.0137, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 3.2319391634980987, | |
| "grad_norm": 0.24322153627872467, | |
| "learning_rate": 5.247918773366112e-05, | |
| "loss": 0.0184, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 3.238276299112801, | |
| "grad_norm": 0.15239882469177246, | |
| "learning_rate": 5.231403087877955e-05, | |
| "loss": 0.0161, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 3.244613434727503, | |
| "grad_norm": 0.31975507736206055, | |
| "learning_rate": 5.214884871802703e-05, | |
| "loss": 0.0147, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 3.2509505703422055, | |
| "grad_norm": 0.22015072405338287, | |
| "learning_rate": 5.198364305780922e-05, | |
| "loss": 0.0105, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 3.2572877059569074, | |
| "grad_norm": 0.34388023614883423, | |
| "learning_rate": 5.1818415704788725e-05, | |
| "loss": 0.0156, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 3.26362484157161, | |
| "grad_norm": 0.3033609092235565, | |
| "learning_rate": 5.165316846586541e-05, | |
| "loss": 0.0148, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 3.2699619771863118, | |
| "grad_norm": 0.1909927874803543, | |
| "learning_rate": 5.148790314815663e-05, | |
| "loss": 0.0123, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 3.2762991128010137, | |
| "grad_norm": 0.23113128542900085, | |
| "learning_rate": 5.132262155897739e-05, | |
| "loss": 0.0164, | |
| "step": 5170 | |
| }, | |
| { | |
| "epoch": 3.282636248415716, | |
| "grad_norm": 0.19848641753196716, | |
| "learning_rate": 5.1157325505820694e-05, | |
| "loss": 0.0173, | |
| "step": 5180 | |
| }, | |
| { | |
| "epoch": 3.288973384030418, | |
| "grad_norm": 0.27411961555480957, | |
| "learning_rate": 5.0992016796337686e-05, | |
| "loss": 0.0105, | |
| "step": 5190 | |
| }, | |
| { | |
| "epoch": 3.2953105196451205, | |
| "grad_norm": 0.19896908104419708, | |
| "learning_rate": 5.0826697238317935e-05, | |
| "loss": 0.0153, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 3.3016476552598224, | |
| "grad_norm": 0.20173603296279907, | |
| "learning_rate": 5.066136863966963e-05, | |
| "loss": 0.0121, | |
| "step": 5210 | |
| }, | |
| { | |
| "epoch": 3.307984790874525, | |
| "grad_norm": 0.1532195508480072, | |
| "learning_rate": 5.0496032808399815e-05, | |
| "loss": 0.0116, | |
| "step": 5220 | |
| }, | |
| { | |
| "epoch": 3.3143219264892267, | |
| "grad_norm": 0.14388404786586761, | |
| "learning_rate": 5.033069155259471e-05, | |
| "loss": 0.0134, | |
| "step": 5230 | |
| }, | |
| { | |
| "epoch": 3.320659062103929, | |
| "grad_norm": 0.2755822241306305, | |
| "learning_rate": 5.016534668039976e-05, | |
| "loss": 0.0103, | |
| "step": 5240 | |
| }, | |
| { | |
| "epoch": 3.326996197718631, | |
| "grad_norm": 0.2270147204399109, | |
| "learning_rate": 5e-05, | |
| "loss": 0.0093, | |
| "step": 5250 | |
| }, | |
| { | |
| "epoch": 3.3333333333333335, | |
| "grad_norm": 0.17542065680027008, | |
| "learning_rate": 4.9834653319600246e-05, | |
| "loss": 0.0119, | |
| "step": 5260 | |
| }, | |
| { | |
| "epoch": 3.3396704689480354, | |
| "grad_norm": 0.26830822229385376, | |
| "learning_rate": 4.96693084474053e-05, | |
| "loss": 0.0139, | |
| "step": 5270 | |
| }, | |
| { | |
| "epoch": 3.346007604562738, | |
| "grad_norm": 0.23319189250469208, | |
| "learning_rate": 4.950396719160018e-05, | |
| "loss": 0.0115, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 3.3523447401774398, | |
| "grad_norm": 0.28234973549842834, | |
| "learning_rate": 4.93386313603304e-05, | |
| "loss": 0.0096, | |
| "step": 5290 | |
| }, | |
| { | |
| "epoch": 3.3586818757921417, | |
| "grad_norm": 0.2358221560716629, | |
| "learning_rate": 4.917330276168208e-05, | |
| "loss": 0.0126, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 3.365019011406844, | |
| "grad_norm": 0.29058966040611267, | |
| "learning_rate": 4.9007983203662326e-05, | |
| "loss": 0.0281, | |
| "step": 5310 | |
| }, | |
| { | |
| "epoch": 3.371356147021546, | |
| "grad_norm": 0.23800086975097656, | |
| "learning_rate": 4.884267449417931e-05, | |
| "loss": 0.0177, | |
| "step": 5320 | |
| }, | |
| { | |
| "epoch": 3.3776932826362485, | |
| "grad_norm": 0.26259422302246094, | |
| "learning_rate": 4.867737844102261e-05, | |
| "loss": 0.014, | |
| "step": 5330 | |
| }, | |
| { | |
| "epoch": 3.3840304182509504, | |
| "grad_norm": 0.2746807336807251, | |
| "learning_rate": 4.851209685184338e-05, | |
| "loss": 0.0102, | |
| "step": 5340 | |
| }, | |
| { | |
| "epoch": 3.390367553865653, | |
| "grad_norm": 0.46735528111457825, | |
| "learning_rate": 4.834683153413459e-05, | |
| "loss": 0.0207, | |
| "step": 5350 | |
| }, | |
| { | |
| "epoch": 3.3967046894803548, | |
| "grad_norm": 0.2719608247280121, | |
| "learning_rate": 4.818158429521129e-05, | |
| "loss": 0.0129, | |
| "step": 5360 | |
| }, | |
| { | |
| "epoch": 3.403041825095057, | |
| "grad_norm": 0.28547102212905884, | |
| "learning_rate": 4.801635694219079e-05, | |
| "loss": 0.0101, | |
| "step": 5370 | |
| }, | |
| { | |
| "epoch": 3.409378960709759, | |
| "grad_norm": 0.37190404534339905, | |
| "learning_rate": 4.785115128197298e-05, | |
| "loss": 0.0156, | |
| "step": 5380 | |
| }, | |
| { | |
| "epoch": 3.4157160963244615, | |
| "grad_norm": 0.3981722295284271, | |
| "learning_rate": 4.7685969121220456e-05, | |
| "loss": 0.0119, | |
| "step": 5390 | |
| }, | |
| { | |
| "epoch": 3.4220532319391634, | |
| "grad_norm": 0.325492262840271, | |
| "learning_rate": 4.7520812266338885e-05, | |
| "loss": 0.0135, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 3.428390367553866, | |
| "grad_norm": 0.24744857847690582, | |
| "learning_rate": 4.735568252345718e-05, | |
| "loss": 0.0152, | |
| "step": 5410 | |
| }, | |
| { | |
| "epoch": 3.434727503168568, | |
| "grad_norm": 0.15038323402404785, | |
| "learning_rate": 4.7190581698407725e-05, | |
| "loss": 0.0161, | |
| "step": 5420 | |
| }, | |
| { | |
| "epoch": 3.4410646387832697, | |
| "grad_norm": 0.2310582995414734, | |
| "learning_rate": 4.702551159670672e-05, | |
| "loss": 0.0121, | |
| "step": 5430 | |
| }, | |
| { | |
| "epoch": 3.447401774397972, | |
| "grad_norm": 0.17245414853096008, | |
| "learning_rate": 4.6860474023534335e-05, | |
| "loss": 0.0127, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 3.453738910012674, | |
| "grad_norm": 0.2583564817905426, | |
| "learning_rate": 4.669547078371504e-05, | |
| "loss": 0.0171, | |
| "step": 5450 | |
| }, | |
| { | |
| "epoch": 3.4600760456273765, | |
| "grad_norm": 0.3162192702293396, | |
| "learning_rate": 4.65305036816978e-05, | |
| "loss": 0.0113, | |
| "step": 5460 | |
| }, | |
| { | |
| "epoch": 3.4664131812420784, | |
| "grad_norm": 0.29524022340774536, | |
| "learning_rate": 4.6365574521536445e-05, | |
| "loss": 0.0126, | |
| "step": 5470 | |
| }, | |
| { | |
| "epoch": 3.472750316856781, | |
| "grad_norm": 0.14694812893867493, | |
| "learning_rate": 4.620068510686985e-05, | |
| "loss": 0.0119, | |
| "step": 5480 | |
| }, | |
| { | |
| "epoch": 3.4790874524714828, | |
| "grad_norm": 0.1337551772594452, | |
| "learning_rate": 4.60358372409022e-05, | |
| "loss": 0.0088, | |
| "step": 5490 | |
| }, | |
| { | |
| "epoch": 3.485424588086185, | |
| "grad_norm": 0.3237602114677429, | |
| "learning_rate": 4.5871032726383386e-05, | |
| "loss": 0.0091, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 3.491761723700887, | |
| "grad_norm": 0.1705574095249176, | |
| "learning_rate": 4.570627336558915e-05, | |
| "loss": 0.0097, | |
| "step": 5510 | |
| }, | |
| { | |
| "epoch": 3.4980988593155895, | |
| "grad_norm": 0.2950294613838196, | |
| "learning_rate": 4.554156096030149e-05, | |
| "loss": 0.0117, | |
| "step": 5520 | |
| }, | |
| { | |
| "epoch": 3.5044359949302915, | |
| "grad_norm": 0.27918121218681335, | |
| "learning_rate": 4.537689731178883e-05, | |
| "loss": 0.0104, | |
| "step": 5530 | |
| }, | |
| { | |
| "epoch": 3.510773130544994, | |
| "grad_norm": 0.2752488851547241, | |
| "learning_rate": 4.5212284220786494e-05, | |
| "loss": 0.009, | |
| "step": 5540 | |
| }, | |
| { | |
| "epoch": 3.517110266159696, | |
| "grad_norm": 0.24830137193202972, | |
| "learning_rate": 4.504772348747687e-05, | |
| "loss": 0.0133, | |
| "step": 5550 | |
| }, | |
| { | |
| "epoch": 3.5234474017743977, | |
| "grad_norm": 0.38127654790878296, | |
| "learning_rate": 4.488321691146975e-05, | |
| "loss": 0.0142, | |
| "step": 5560 | |
| }, | |
| { | |
| "epoch": 3.5297845373891, | |
| "grad_norm": 0.3115534782409668, | |
| "learning_rate": 4.471876629178273e-05, | |
| "loss": 0.0158, | |
| "step": 5570 | |
| }, | |
| { | |
| "epoch": 3.5361216730038025, | |
| "grad_norm": 0.18441027402877808, | |
| "learning_rate": 4.4554373426821374e-05, | |
| "loss": 0.0087, | |
| "step": 5580 | |
| }, | |
| { | |
| "epoch": 3.5424588086185045, | |
| "grad_norm": 0.19032280147075653, | |
| "learning_rate": 4.439004011435979e-05, | |
| "loss": 0.0128, | |
| "step": 5590 | |
| }, | |
| { | |
| "epoch": 3.5487959442332064, | |
| "grad_norm": 0.16672304272651672, | |
| "learning_rate": 4.4225768151520694e-05, | |
| "loss": 0.0182, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 3.555133079847909, | |
| "grad_norm": 0.24737948179244995, | |
| "learning_rate": 4.406155933475599e-05, | |
| "loss": 0.0112, | |
| "step": 5610 | |
| }, | |
| { | |
| "epoch": 3.5614702154626108, | |
| "grad_norm": 0.17746677994728088, | |
| "learning_rate": 4.3897415459827e-05, | |
| "loss": 0.0151, | |
| "step": 5620 | |
| }, | |
| { | |
| "epoch": 3.567807351077313, | |
| "grad_norm": 0.2580298185348511, | |
| "learning_rate": 4.373333832178478e-05, | |
| "loss": 0.0092, | |
| "step": 5630 | |
| }, | |
| { | |
| "epoch": 3.574144486692015, | |
| "grad_norm": 0.24834871292114258, | |
| "learning_rate": 4.3569329714950704e-05, | |
| "loss": 0.0133, | |
| "step": 5640 | |
| }, | |
| { | |
| "epoch": 3.5804816223067175, | |
| "grad_norm": 0.23635822534561157, | |
| "learning_rate": 4.3405391432896555e-05, | |
| "loss": 0.0132, | |
| "step": 5650 | |
| }, | |
| { | |
| "epoch": 3.5868187579214195, | |
| "grad_norm": 0.2585609555244446, | |
| "learning_rate": 4.324152526842517e-05, | |
| "loss": 0.014, | |
| "step": 5660 | |
| }, | |
| { | |
| "epoch": 3.593155893536122, | |
| "grad_norm": 0.2859804034233093, | |
| "learning_rate": 4.307773301355062e-05, | |
| "loss": 0.0101, | |
| "step": 5670 | |
| }, | |
| { | |
| "epoch": 3.599493029150824, | |
| "grad_norm": 0.18319383263587952, | |
| "learning_rate": 4.291401645947879e-05, | |
| "loss": 0.0202, | |
| "step": 5680 | |
| }, | |
| { | |
| "epoch": 3.6058301647655258, | |
| "grad_norm": 0.2706187665462494, | |
| "learning_rate": 4.275037739658771e-05, | |
| "loss": 0.0067, | |
| "step": 5690 | |
| }, | |
| { | |
| "epoch": 3.612167300380228, | |
| "grad_norm": 0.2773619592189789, | |
| "learning_rate": 4.2586817614407895e-05, | |
| "loss": 0.0114, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 3.6185044359949305, | |
| "grad_norm": 0.24154767394065857, | |
| "learning_rate": 4.2423338901602985e-05, | |
| "loss": 0.0178, | |
| "step": 5710 | |
| }, | |
| { | |
| "epoch": 3.6248415716096325, | |
| "grad_norm": 0.1268390715122223, | |
| "learning_rate": 4.2259943045949934e-05, | |
| "loss": 0.0098, | |
| "step": 5720 | |
| }, | |
| { | |
| "epoch": 3.6311787072243344, | |
| "grad_norm": 0.14447900652885437, | |
| "learning_rate": 4.209663183431969e-05, | |
| "loss": 0.0121, | |
| "step": 5730 | |
| }, | |
| { | |
| "epoch": 3.637515842839037, | |
| "grad_norm": 0.18617920577526093, | |
| "learning_rate": 4.1933407052657456e-05, | |
| "loss": 0.0092, | |
| "step": 5740 | |
| }, | |
| { | |
| "epoch": 3.643852978453739, | |
| "grad_norm": 0.3276727497577667, | |
| "learning_rate": 4.17702704859633e-05, | |
| "loss": 0.0109, | |
| "step": 5750 | |
| }, | |
| { | |
| "epoch": 3.650190114068441, | |
| "grad_norm": 0.20072652399539948, | |
| "learning_rate": 4.160722391827262e-05, | |
| "loss": 0.0106, | |
| "step": 5760 | |
| }, | |
| { | |
| "epoch": 3.656527249683143, | |
| "grad_norm": 0.2064548283815384, | |
| "learning_rate": 4.14442691326365e-05, | |
| "loss": 0.0097, | |
| "step": 5770 | |
| }, | |
| { | |
| "epoch": 3.6628643852978455, | |
| "grad_norm": 0.20051850378513336, | |
| "learning_rate": 4.1281407911102425e-05, | |
| "loss": 0.0127, | |
| "step": 5780 | |
| }, | |
| { | |
| "epoch": 3.6692015209125475, | |
| "grad_norm": 0.27055758237838745, | |
| "learning_rate": 4.111864203469457e-05, | |
| "loss": 0.0125, | |
| "step": 5790 | |
| }, | |
| { | |
| "epoch": 3.67553865652725, | |
| "grad_norm": 0.3249766230583191, | |
| "learning_rate": 4.095597328339452e-05, | |
| "loss": 0.013, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 3.681875792141952, | |
| "grad_norm": 0.36412471532821655, | |
| "learning_rate": 4.079340343612165e-05, | |
| "loss": 0.0135, | |
| "step": 5810 | |
| }, | |
| { | |
| "epoch": 3.6882129277566538, | |
| "grad_norm": 0.43938231468200684, | |
| "learning_rate": 4.063093427071376e-05, | |
| "loss": 0.0142, | |
| "step": 5820 | |
| }, | |
| { | |
| "epoch": 3.694550063371356, | |
| "grad_norm": 0.2211577147245407, | |
| "learning_rate": 4.046856756390767e-05, | |
| "loss": 0.0092, | |
| "step": 5830 | |
| }, | |
| { | |
| "epoch": 3.7008871989860586, | |
| "grad_norm": 0.2317885160446167, | |
| "learning_rate": 4.0306305091319595e-05, | |
| "loss": 0.0103, | |
| "step": 5840 | |
| }, | |
| { | |
| "epoch": 3.7072243346007605, | |
| "grad_norm": 0.30124250054359436, | |
| "learning_rate": 4.0144148627425993e-05, | |
| "loss": 0.0128, | |
| "step": 5850 | |
| }, | |
| { | |
| "epoch": 3.7135614702154625, | |
| "grad_norm": 0.22802747786045074, | |
| "learning_rate": 3.9982099945543945e-05, | |
| "loss": 0.0092, | |
| "step": 5860 | |
| }, | |
| { | |
| "epoch": 3.719898605830165, | |
| "grad_norm": 0.3372552990913391, | |
| "learning_rate": 3.982016081781189e-05, | |
| "loss": 0.009, | |
| "step": 5870 | |
| }, | |
| { | |
| "epoch": 3.726235741444867, | |
| "grad_norm": 0.3483387529850006, | |
| "learning_rate": 3.965833301517017e-05, | |
| "loss": 0.015, | |
| "step": 5880 | |
| }, | |
| { | |
| "epoch": 3.732572877059569, | |
| "grad_norm": 0.17584814131259918, | |
| "learning_rate": 3.949661830734172e-05, | |
| "loss": 0.0077, | |
| "step": 5890 | |
| }, | |
| { | |
| "epoch": 3.738910012674271, | |
| "grad_norm": 0.26330336928367615, | |
| "learning_rate": 3.933501846281267e-05, | |
| "loss": 0.0101, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 3.7452471482889735, | |
| "grad_norm": 0.19230346381664276, | |
| "learning_rate": 3.917353524881302e-05, | |
| "loss": 0.014, | |
| "step": 5910 | |
| }, | |
| { | |
| "epoch": 3.7515842839036755, | |
| "grad_norm": 0.19629403948783875, | |
| "learning_rate": 3.901217043129735e-05, | |
| "loss": 0.0124, | |
| "step": 5920 | |
| }, | |
| { | |
| "epoch": 3.757921419518378, | |
| "grad_norm": 0.22321918606758118, | |
| "learning_rate": 3.8850925774925425e-05, | |
| "loss": 0.0116, | |
| "step": 5930 | |
| }, | |
| { | |
| "epoch": 3.76425855513308, | |
| "grad_norm": 0.20382849872112274, | |
| "learning_rate": 3.8689803043043e-05, | |
| "loss": 0.0142, | |
| "step": 5940 | |
| }, | |
| { | |
| "epoch": 3.770595690747782, | |
| "grad_norm": 0.16952167451381683, | |
| "learning_rate": 3.852880399766243e-05, | |
| "loss": 0.0125, | |
| "step": 5950 | |
| }, | |
| { | |
| "epoch": 3.776932826362484, | |
| "grad_norm": 0.23627594113349915, | |
| "learning_rate": 3.836793039944349e-05, | |
| "loss": 0.0132, | |
| "step": 5960 | |
| }, | |
| { | |
| "epoch": 3.7832699619771866, | |
| "grad_norm": 0.20818836987018585, | |
| "learning_rate": 3.820718400767409e-05, | |
| "loss": 0.0121, | |
| "step": 5970 | |
| }, | |
| { | |
| "epoch": 3.7896070975918885, | |
| "grad_norm": 0.37023457884788513, | |
| "learning_rate": 3.8046566580251e-05, | |
| "loss": 0.0093, | |
| "step": 5980 | |
| }, | |
| { | |
| "epoch": 3.7959442332065905, | |
| "grad_norm": 0.24624159932136536, | |
| "learning_rate": 3.788607987366069e-05, | |
| "loss": 0.0105, | |
| "step": 5990 | |
| }, | |
| { | |
| "epoch": 3.802281368821293, | |
| "grad_norm": 0.2265588641166687, | |
| "learning_rate": 3.772572564296005e-05, | |
| "loss": 0.0148, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 3.808618504435995, | |
| "grad_norm": 0.24321384727954865, | |
| "learning_rate": 3.756550564175727e-05, | |
| "loss": 0.014, | |
| "step": 6010 | |
| }, | |
| { | |
| "epoch": 3.814955640050697, | |
| "grad_norm": 0.15189670026302338, | |
| "learning_rate": 3.74054216221926e-05, | |
| "loss": 0.0112, | |
| "step": 6020 | |
| }, | |
| { | |
| "epoch": 3.821292775665399, | |
| "grad_norm": 0.366676926612854, | |
| "learning_rate": 3.7245475334919246e-05, | |
| "loss": 0.0163, | |
| "step": 6030 | |
| }, | |
| { | |
| "epoch": 3.8276299112801015, | |
| "grad_norm": 0.22733189165592194, | |
| "learning_rate": 3.7085668529084184e-05, | |
| "loss": 0.014, | |
| "step": 6040 | |
| }, | |
| { | |
| "epoch": 3.8339670468948035, | |
| "grad_norm": 0.17254877090454102, | |
| "learning_rate": 3.6926002952309016e-05, | |
| "loss": 0.0112, | |
| "step": 6050 | |
| }, | |
| { | |
| "epoch": 3.840304182509506, | |
| "grad_norm": 0.22515353560447693, | |
| "learning_rate": 3.676648035067093e-05, | |
| "loss": 0.0113, | |
| "step": 6060 | |
| }, | |
| { | |
| "epoch": 3.846641318124208, | |
| "grad_norm": 0.239216610789299, | |
| "learning_rate": 3.6607102468683526e-05, | |
| "loss": 0.011, | |
| "step": 6070 | |
| }, | |
| { | |
| "epoch": 3.85297845373891, | |
| "grad_norm": 0.11387116461992264, | |
| "learning_rate": 3.6447871049277796e-05, | |
| "loss": 0.0072, | |
| "step": 6080 | |
| }, | |
| { | |
| "epoch": 3.859315589353612, | |
| "grad_norm": 0.18705078959465027, | |
| "learning_rate": 3.628878783378302e-05, | |
| "loss": 0.0118, | |
| "step": 6090 | |
| }, | |
| { | |
| "epoch": 3.8656527249683146, | |
| "grad_norm": 0.1804051548242569, | |
| "learning_rate": 3.612985456190778e-05, | |
| "loss": 0.0134, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 3.8719898605830165, | |
| "grad_norm": 0.4679585099220276, | |
| "learning_rate": 3.597107297172084e-05, | |
| "loss": 0.011, | |
| "step": 6110 | |
| }, | |
| { | |
| "epoch": 3.8783269961977185, | |
| "grad_norm": 0.14044460654258728, | |
| "learning_rate": 3.581244479963225e-05, | |
| "loss": 0.0106, | |
| "step": 6120 | |
| }, | |
| { | |
| "epoch": 3.884664131812421, | |
| "grad_norm": 0.332378089427948, | |
| "learning_rate": 3.5653971780374295e-05, | |
| "loss": 0.0166, | |
| "step": 6130 | |
| }, | |
| { | |
| "epoch": 3.891001267427123, | |
| "grad_norm": 0.2639565169811249, | |
| "learning_rate": 3.5495655646982505e-05, | |
| "loss": 0.0139, | |
| "step": 6140 | |
| }, | |
| { | |
| "epoch": 3.897338403041825, | |
| "grad_norm": 0.2334318906068802, | |
| "learning_rate": 3.533749813077677e-05, | |
| "loss": 0.0137, | |
| "step": 6150 | |
| }, | |
| { | |
| "epoch": 3.903675538656527, | |
| "grad_norm": 0.3366606831550598, | |
| "learning_rate": 3.517950096134232e-05, | |
| "loss": 0.0121, | |
| "step": 6160 | |
| }, | |
| { | |
| "epoch": 3.9100126742712296, | |
| "grad_norm": 0.32218649983406067, | |
| "learning_rate": 3.5021665866510925e-05, | |
| "loss": 0.0113, | |
| "step": 6170 | |
| }, | |
| { | |
| "epoch": 3.9163498098859315, | |
| "grad_norm": 0.14572682976722717, | |
| "learning_rate": 3.4863994572341843e-05, | |
| "loss": 0.0147, | |
| "step": 6180 | |
| }, | |
| { | |
| "epoch": 3.922686945500634, | |
| "grad_norm": 0.2537533938884735, | |
| "learning_rate": 3.470648880310313e-05, | |
| "loss": 0.011, | |
| "step": 6190 | |
| }, | |
| { | |
| "epoch": 3.929024081115336, | |
| "grad_norm": 0.25570058822631836, | |
| "learning_rate": 3.4549150281252636e-05, | |
| "loss": 0.0107, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 3.935361216730038, | |
| "grad_norm": 0.18532606959342957, | |
| "learning_rate": 3.439198072741921e-05, | |
| "loss": 0.0109, | |
| "step": 6210 | |
| }, | |
| { | |
| "epoch": 3.94169835234474, | |
| "grad_norm": 0.1890391707420349, | |
| "learning_rate": 3.423498186038393e-05, | |
| "loss": 0.0168, | |
| "step": 6220 | |
| }, | |
| { | |
| "epoch": 3.9480354879594426, | |
| "grad_norm": 0.17653165757656097, | |
| "learning_rate": 3.407815539706124e-05, | |
| "loss": 0.01, | |
| "step": 6230 | |
| }, | |
| { | |
| "epoch": 3.9543726235741445, | |
| "grad_norm": 0.127501979470253, | |
| "learning_rate": 3.392150305248024e-05, | |
| "loss": 0.0122, | |
| "step": 6240 | |
| }, | |
| { | |
| "epoch": 3.9607097591888465, | |
| "grad_norm": 0.16189290583133698, | |
| "learning_rate": 3.3765026539765834e-05, | |
| "loss": 0.0114, | |
| "step": 6250 | |
| }, | |
| { | |
| "epoch": 3.967046894803549, | |
| "grad_norm": 0.17998936772346497, | |
| "learning_rate": 3.360872757012011e-05, | |
| "loss": 0.0148, | |
| "step": 6260 | |
| }, | |
| { | |
| "epoch": 3.973384030418251, | |
| "grad_norm": 0.17054511606693268, | |
| "learning_rate": 3.3452607852803584e-05, | |
| "loss": 0.0112, | |
| "step": 6270 | |
| }, | |
| { | |
| "epoch": 3.9797211660329532, | |
| "grad_norm": 0.2786466181278229, | |
| "learning_rate": 3.329666909511645e-05, | |
| "loss": 0.0123, | |
| "step": 6280 | |
| }, | |
| { | |
| "epoch": 3.986058301647655, | |
| "grad_norm": 0.21183429658412933, | |
| "learning_rate": 3.3140913002379995e-05, | |
| "loss": 0.0099, | |
| "step": 6290 | |
| }, | |
| { | |
| "epoch": 3.9923954372623576, | |
| "grad_norm": 0.2610870897769928, | |
| "learning_rate": 3.298534127791785e-05, | |
| "loss": 0.0136, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 3.9987325728770595, | |
| "grad_norm": 0.22637638449668884, | |
| "learning_rate": 3.282995562303754e-05, | |
| "loss": 0.011, | |
| "step": 6310 | |
| }, | |
| { | |
| "epoch": 4.005069708491762, | |
| "grad_norm": 0.18758799135684967, | |
| "learning_rate": 3.267475773701161e-05, | |
| "loss": 0.009, | |
| "step": 6320 | |
| }, | |
| { | |
| "epoch": 4.011406844106464, | |
| "grad_norm": 0.22371801733970642, | |
| "learning_rate": 3.251974931705933e-05, | |
| "loss": 0.0107, | |
| "step": 6330 | |
| }, | |
| { | |
| "epoch": 4.017743979721166, | |
| "grad_norm": 0.22646395862102509, | |
| "learning_rate": 3.236493205832795e-05, | |
| "loss": 0.0117, | |
| "step": 6340 | |
| }, | |
| { | |
| "epoch": 4.024081115335868, | |
| "grad_norm": 0.19698502123355865, | |
| "learning_rate": 3.221030765387417e-05, | |
| "loss": 0.0119, | |
| "step": 6350 | |
| }, | |
| { | |
| "epoch": 4.030418250950571, | |
| "grad_norm": 0.22287245094776154, | |
| "learning_rate": 3.205587779464576e-05, | |
| "loss": 0.0093, | |
| "step": 6360 | |
| }, | |
| { | |
| "epoch": 4.0367553865652726, | |
| "grad_norm": 0.27699437737464905, | |
| "learning_rate": 3.190164416946285e-05, | |
| "loss": 0.0093, | |
| "step": 6370 | |
| }, | |
| { | |
| "epoch": 4.0430925221799745, | |
| "grad_norm": 0.1392444372177124, | |
| "learning_rate": 3.1747608464999725e-05, | |
| "loss": 0.0072, | |
| "step": 6380 | |
| }, | |
| { | |
| "epoch": 4.0494296577946765, | |
| "grad_norm": 0.2287701815366745, | |
| "learning_rate": 3.1593772365766105e-05, | |
| "loss": 0.0154, | |
| "step": 6390 | |
| }, | |
| { | |
| "epoch": 4.055766793409379, | |
| "grad_norm": 0.24777956306934357, | |
| "learning_rate": 3.144013755408895e-05, | |
| "loss": 0.012, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 4.062103929024081, | |
| "grad_norm": 0.1426210254430771, | |
| "learning_rate": 3.128670571009399e-05, | |
| "loss": 0.012, | |
| "step": 6410 | |
| }, | |
| { | |
| "epoch": 4.068441064638783, | |
| "grad_norm": 0.11875156313180923, | |
| "learning_rate": 3.113347851168721e-05, | |
| "loss": 0.0088, | |
| "step": 6420 | |
| }, | |
| { | |
| "epoch": 4.074778200253485, | |
| "grad_norm": 0.1875237077474594, | |
| "learning_rate": 3.098045763453678e-05, | |
| "loss": 0.0074, | |
| "step": 6430 | |
| }, | |
| { | |
| "epoch": 4.081115335868188, | |
| "grad_norm": 0.09495358169078827, | |
| "learning_rate": 3.082764475205442e-05, | |
| "loss": 0.0095, | |
| "step": 6440 | |
| }, | |
| { | |
| "epoch": 4.08745247148289, | |
| "grad_norm": 0.2620109021663666, | |
| "learning_rate": 3.0675041535377405e-05, | |
| "loss": 0.014, | |
| "step": 6450 | |
| }, | |
| { | |
| "epoch": 4.093789607097592, | |
| "grad_norm": 0.30358102917671204, | |
| "learning_rate": 3.052264965335e-05, | |
| "loss": 0.0068, | |
| "step": 6460 | |
| }, | |
| { | |
| "epoch": 4.100126742712294, | |
| "grad_norm": 0.2840622663497925, | |
| "learning_rate": 3.0370470772505433e-05, | |
| "loss": 0.008, | |
| "step": 6470 | |
| }, | |
| { | |
| "epoch": 4.106463878326996, | |
| "grad_norm": 0.12551774084568024, | |
| "learning_rate": 3.0218506557047598e-05, | |
| "loss": 0.0069, | |
| "step": 6480 | |
| }, | |
| { | |
| "epoch": 4.112801013941699, | |
| "grad_norm": 0.2494884580373764, | |
| "learning_rate": 3.006675866883275e-05, | |
| "loss": 0.0076, | |
| "step": 6490 | |
| }, | |
| { | |
| "epoch": 4.119138149556401, | |
| "grad_norm": 0.11765393614768982, | |
| "learning_rate": 2.991522876735154e-05, | |
| "loss": 0.0065, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 4.1254752851711025, | |
| "grad_norm": 0.1556195318698883, | |
| "learning_rate": 2.976391850971065e-05, | |
| "loss": 0.0155, | |
| "step": 6510 | |
| }, | |
| { | |
| "epoch": 4.1318124207858045, | |
| "grad_norm": 0.24738621711730957, | |
| "learning_rate": 2.9612829550614836e-05, | |
| "loss": 0.0107, | |
| "step": 6520 | |
| }, | |
| { | |
| "epoch": 4.138149556400507, | |
| "grad_norm": 0.2733817398548126, | |
| "learning_rate": 2.9461963542348737e-05, | |
| "loss": 0.0083, | |
| "step": 6530 | |
| }, | |
| { | |
| "epoch": 4.144486692015209, | |
| "grad_norm": 0.4232415556907654, | |
| "learning_rate": 2.931132213475884e-05, | |
| "loss": 0.0154, | |
| "step": 6540 | |
| }, | |
| { | |
| "epoch": 4.150823827629911, | |
| "grad_norm": 0.16608037054538727, | |
| "learning_rate": 2.916090697523549e-05, | |
| "loss": 0.0117, | |
| "step": 6550 | |
| }, | |
| { | |
| "epoch": 4.157160963244613, | |
| "grad_norm": 0.226557657122612, | |
| "learning_rate": 2.9010719708694722e-05, | |
| "loss": 0.0087, | |
| "step": 6560 | |
| }, | |
| { | |
| "epoch": 4.163498098859316, | |
| "grad_norm": 0.34729430079460144, | |
| "learning_rate": 2.8860761977560436e-05, | |
| "loss": 0.012, | |
| "step": 6570 | |
| }, | |
| { | |
| "epoch": 4.169835234474018, | |
| "grad_norm": 0.2842123806476593, | |
| "learning_rate": 2.8711035421746367e-05, | |
| "loss": 0.0109, | |
| "step": 6580 | |
| }, | |
| { | |
| "epoch": 4.17617237008872, | |
| "grad_norm": 0.24015074968338013, | |
| "learning_rate": 2.8561541678638142e-05, | |
| "loss": 0.01, | |
| "step": 6590 | |
| }, | |
| { | |
| "epoch": 4.182509505703422, | |
| "grad_norm": 0.2127242535352707, | |
| "learning_rate": 2.8412282383075363e-05, | |
| "loss": 0.0092, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 4.188846641318124, | |
| "grad_norm": 0.1556386947631836, | |
| "learning_rate": 2.8263259167333777e-05, | |
| "loss": 0.0066, | |
| "step": 6610 | |
| }, | |
| { | |
| "epoch": 4.195183776932827, | |
| "grad_norm": 0.2527133524417877, | |
| "learning_rate": 2.811447366110741e-05, | |
| "loss": 0.0121, | |
| "step": 6620 | |
| }, | |
| { | |
| "epoch": 4.201520912547529, | |
| "grad_norm": 0.13160255551338196, | |
| "learning_rate": 2.7965927491490705e-05, | |
| "loss": 0.0069, | |
| "step": 6630 | |
| }, | |
| { | |
| "epoch": 4.2078580481622305, | |
| "grad_norm": 0.29007360339164734, | |
| "learning_rate": 2.7817622282960815e-05, | |
| "loss": 0.0107, | |
| "step": 6640 | |
| }, | |
| { | |
| "epoch": 4.2141951837769325, | |
| "grad_norm": 0.09741051495075226, | |
| "learning_rate": 2.766955965735968e-05, | |
| "loss": 0.0069, | |
| "step": 6650 | |
| }, | |
| { | |
| "epoch": 4.220532319391635, | |
| "grad_norm": 0.2465200573205948, | |
| "learning_rate": 2.7521741233876496e-05, | |
| "loss": 0.0105, | |
| "step": 6660 | |
| }, | |
| { | |
| "epoch": 4.226869455006337, | |
| "grad_norm": 0.18087448179721832, | |
| "learning_rate": 2.7374168629029813e-05, | |
| "loss": 0.0122, | |
| "step": 6670 | |
| }, | |
| { | |
| "epoch": 4.233206590621039, | |
| "grad_norm": 0.15381130576133728, | |
| "learning_rate": 2.7226843456650037e-05, | |
| "loss": 0.0091, | |
| "step": 6680 | |
| }, | |
| { | |
| "epoch": 4.239543726235741, | |
| "grad_norm": 0.16442646086215973, | |
| "learning_rate": 2.707976732786166e-05, | |
| "loss": 0.0094, | |
| "step": 6690 | |
| }, | |
| { | |
| "epoch": 4.245880861850444, | |
| "grad_norm": 0.14896883070468903, | |
| "learning_rate": 2.693294185106562e-05, | |
| "loss": 0.0158, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 4.252217997465146, | |
| "grad_norm": 0.1330922693014145, | |
| "learning_rate": 2.6786368631921836e-05, | |
| "loss": 0.0097, | |
| "step": 6710 | |
| }, | |
| { | |
| "epoch": 4.258555133079848, | |
| "grad_norm": 0.18497434258460999, | |
| "learning_rate": 2.6640049273331515e-05, | |
| "loss": 0.0082, | |
| "step": 6720 | |
| }, | |
| { | |
| "epoch": 4.26489226869455, | |
| "grad_norm": 0.12878283858299255, | |
| "learning_rate": 2.6493985375419778e-05, | |
| "loss": 0.0141, | |
| "step": 6730 | |
| }, | |
| { | |
| "epoch": 4.271229404309253, | |
| "grad_norm": 0.1790645867586136, | |
| "learning_rate": 2.6348178535517966e-05, | |
| "loss": 0.012, | |
| "step": 6740 | |
| }, | |
| { | |
| "epoch": 4.277566539923955, | |
| "grad_norm": 0.29659292101860046, | |
| "learning_rate": 2.6202630348146324e-05, | |
| "loss": 0.0141, | |
| "step": 6750 | |
| }, | |
| { | |
| "epoch": 4.283903675538657, | |
| "grad_norm": 0.2492123246192932, | |
| "learning_rate": 2.6057342404996522e-05, | |
| "loss": 0.0133, | |
| "step": 6760 | |
| }, | |
| { | |
| "epoch": 4.2902408111533585, | |
| "grad_norm": 0.11853443086147308, | |
| "learning_rate": 2.591231629491423e-05, | |
| "loss": 0.01, | |
| "step": 6770 | |
| }, | |
| { | |
| "epoch": 4.2965779467680605, | |
| "grad_norm": 0.24464382231235504, | |
| "learning_rate": 2.5767553603881767e-05, | |
| "loss": 0.0088, | |
| "step": 6780 | |
| }, | |
| { | |
| "epoch": 4.302915082382763, | |
| "grad_norm": 0.2709729075431824, | |
| "learning_rate": 2.562305591500069e-05, | |
| "loss": 0.0075, | |
| "step": 6790 | |
| }, | |
| { | |
| "epoch": 4.309252217997465, | |
| "grad_norm": 0.19660340249538422, | |
| "learning_rate": 2.547882480847461e-05, | |
| "loss": 0.0136, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 4.315589353612167, | |
| "grad_norm": 0.21766068041324615, | |
| "learning_rate": 2.5334861861591753e-05, | |
| "loss": 0.0073, | |
| "step": 6810 | |
| }, | |
| { | |
| "epoch": 4.321926489226869, | |
| "grad_norm": 0.27058249711990356, | |
| "learning_rate": 2.5191168648707887e-05, | |
| "loss": 0.0128, | |
| "step": 6820 | |
| }, | |
| { | |
| "epoch": 4.328263624841572, | |
| "grad_norm": 0.18688562512397766, | |
| "learning_rate": 2.5047746741228978e-05, | |
| "loss": 0.0146, | |
| "step": 6830 | |
| }, | |
| { | |
| "epoch": 4.334600760456274, | |
| "grad_norm": 0.23607704043388367, | |
| "learning_rate": 2.490459770759398e-05, | |
| "loss": 0.0089, | |
| "step": 6840 | |
| }, | |
| { | |
| "epoch": 4.340937896070976, | |
| "grad_norm": 0.1704394668340683, | |
| "learning_rate": 2.476172311325783e-05, | |
| "loss": 0.009, | |
| "step": 6850 | |
| }, | |
| { | |
| "epoch": 4.347275031685678, | |
| "grad_norm": 0.11735841631889343, | |
| "learning_rate": 2.4619124520674146e-05, | |
| "loss": 0.0099, | |
| "step": 6860 | |
| }, | |
| { | |
| "epoch": 4.35361216730038, | |
| "grad_norm": 0.17828071117401123, | |
| "learning_rate": 2.447680348927837e-05, | |
| "loss": 0.0069, | |
| "step": 6870 | |
| }, | |
| { | |
| "epoch": 4.359949302915083, | |
| "grad_norm": 0.09438079595565796, | |
| "learning_rate": 2.433476157547044e-05, | |
| "loss": 0.0077, | |
| "step": 6880 | |
| }, | |
| { | |
| "epoch": 4.366286438529785, | |
| "grad_norm": 0.14729665219783783, | |
| "learning_rate": 2.419300033259798e-05, | |
| "loss": 0.0131, | |
| "step": 6890 | |
| }, | |
| { | |
| "epoch": 4.3726235741444865, | |
| "grad_norm": 0.21107903122901917, | |
| "learning_rate": 2.405152131093926e-05, | |
| "loss": 0.0089, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 4.3789607097591885, | |
| "grad_norm": 0.26359254121780396, | |
| "learning_rate": 2.3910326057686127e-05, | |
| "loss": 0.012, | |
| "step": 6910 | |
| }, | |
| { | |
| "epoch": 4.385297845373891, | |
| "grad_norm": 0.22561860084533691, | |
| "learning_rate": 2.3769416116927335e-05, | |
| "loss": 0.0112, | |
| "step": 6920 | |
| }, | |
| { | |
| "epoch": 4.391634980988593, | |
| "grad_norm": 0.253752201795578, | |
| "learning_rate": 2.362879302963135e-05, | |
| "loss": 0.0087, | |
| "step": 6930 | |
| }, | |
| { | |
| "epoch": 4.397972116603295, | |
| "grad_norm": 0.1931048333644867, | |
| "learning_rate": 2.3488458333629777e-05, | |
| "loss": 0.0078, | |
| "step": 6940 | |
| }, | |
| { | |
| "epoch": 4.404309252217997, | |
| "grad_norm": 0.13094092905521393, | |
| "learning_rate": 2.3348413563600325e-05, | |
| "loss": 0.0088, | |
| "step": 6950 | |
| }, | |
| { | |
| "epoch": 4.4106463878327, | |
| "grad_norm": 0.14549599587917328, | |
| "learning_rate": 2.3208660251050158e-05, | |
| "loss": 0.0099, | |
| "step": 6960 | |
| }, | |
| { | |
| "epoch": 4.416983523447402, | |
| "grad_norm": 0.10897214710712433, | |
| "learning_rate": 2.3069199924299174e-05, | |
| "loss": 0.0084, | |
| "step": 6970 | |
| }, | |
| { | |
| "epoch": 4.423320659062104, | |
| "grad_norm": 0.13430652022361755, | |
| "learning_rate": 2.29300341084631e-05, | |
| "loss": 0.0076, | |
| "step": 6980 | |
| }, | |
| { | |
| "epoch": 4.429657794676806, | |
| "grad_norm": 0.3317708969116211, | |
| "learning_rate": 2.279116432543705e-05, | |
| "loss": 0.0178, | |
| "step": 6990 | |
| }, | |
| { | |
| "epoch": 4.435994930291509, | |
| "grad_norm": 0.29772767424583435, | |
| "learning_rate": 2.2652592093878666e-05, | |
| "loss": 0.012, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 4.442332065906211, | |
| "grad_norm": 0.287555068731308, | |
| "learning_rate": 2.251431892919171e-05, | |
| "loss": 0.0131, | |
| "step": 7010 | |
| }, | |
| { | |
| "epoch": 4.448669201520913, | |
| "grad_norm": 0.22690831124782562, | |
| "learning_rate": 2.237634634350934e-05, | |
| "loss": 0.0105, | |
| "step": 7020 | |
| }, | |
| { | |
| "epoch": 4.455006337135615, | |
| "grad_norm": 0.2328868806362152, | |
| "learning_rate": 2.2238675845677663e-05, | |
| "loss": 0.011, | |
| "step": 7030 | |
| }, | |
| { | |
| "epoch": 4.4613434727503165, | |
| "grad_norm": 0.17060483992099762, | |
| "learning_rate": 2.2101308941239203e-05, | |
| "loss": 0.0099, | |
| "step": 7040 | |
| }, | |
| { | |
| "epoch": 4.467680608365019, | |
| "grad_norm": 0.3242473006248474, | |
| "learning_rate": 2.196424713241637e-05, | |
| "loss": 0.0107, | |
| "step": 7050 | |
| }, | |
| { | |
| "epoch": 4.474017743979721, | |
| "grad_norm": 0.2539024353027344, | |
| "learning_rate": 2.182749191809518e-05, | |
| "loss": 0.0069, | |
| "step": 7060 | |
| }, | |
| { | |
| "epoch": 4.480354879594423, | |
| "grad_norm": 0.2088938057422638, | |
| "learning_rate": 2.1691044793808734e-05, | |
| "loss": 0.0118, | |
| "step": 7070 | |
| }, | |
| { | |
| "epoch": 4.486692015209125, | |
| "grad_norm": 0.24568264186382294, | |
| "learning_rate": 2.1554907251720945e-05, | |
| "loss": 0.0091, | |
| "step": 7080 | |
| }, | |
| { | |
| "epoch": 4.493029150823828, | |
| "grad_norm": 0.14442333579063416, | |
| "learning_rate": 2.1419080780610123e-05, | |
| "loss": 0.0088, | |
| "step": 7090 | |
| }, | |
| { | |
| "epoch": 4.49936628643853, | |
| "grad_norm": 0.22089751064777374, | |
| "learning_rate": 2.128356686585282e-05, | |
| "loss": 0.013, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 4.505703422053232, | |
| "grad_norm": 0.1270669400691986, | |
| "learning_rate": 2.1148366989407496e-05, | |
| "loss": 0.0113, | |
| "step": 7110 | |
| }, | |
| { | |
| "epoch": 4.512040557667934, | |
| "grad_norm": 0.11936408281326294, | |
| "learning_rate": 2.1013482629798333e-05, | |
| "loss": 0.0072, | |
| "step": 7120 | |
| }, | |
| { | |
| "epoch": 4.518377693282636, | |
| "grad_norm": 0.1418917030096054, | |
| "learning_rate": 2.0878915262099098e-05, | |
| "loss": 0.0097, | |
| "step": 7130 | |
| }, | |
| { | |
| "epoch": 4.524714828897339, | |
| "grad_norm": 0.22373194992542267, | |
| "learning_rate": 2.0744666357916925e-05, | |
| "loss": 0.0107, | |
| "step": 7140 | |
| }, | |
| { | |
| "epoch": 4.531051964512041, | |
| "grad_norm": 0.16027112305164337, | |
| "learning_rate": 2.061073738537635e-05, | |
| "loss": 0.0076, | |
| "step": 7150 | |
| }, | |
| { | |
| "epoch": 4.537389100126743, | |
| "grad_norm": 0.205812469124794, | |
| "learning_rate": 2.0477129809103147e-05, | |
| "loss": 0.0156, | |
| "step": 7160 | |
| }, | |
| { | |
| "epoch": 4.5437262357414445, | |
| "grad_norm": 0.07858346402645111, | |
| "learning_rate": 2.0343845090208368e-05, | |
| "loss": 0.0069, | |
| "step": 7170 | |
| }, | |
| { | |
| "epoch": 4.550063371356147, | |
| "grad_norm": 0.173725426197052, | |
| "learning_rate": 2.0210884686272368e-05, | |
| "loss": 0.0107, | |
| "step": 7180 | |
| }, | |
| { | |
| "epoch": 4.556400506970849, | |
| "grad_norm": 0.14513538777828217, | |
| "learning_rate": 2.0078250051328784e-05, | |
| "loss": 0.0096, | |
| "step": 7190 | |
| }, | |
| { | |
| "epoch": 4.562737642585551, | |
| "grad_norm": 0.14513027667999268, | |
| "learning_rate": 1.9945942635848748e-05, | |
| "loss": 0.0155, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 4.569074778200253, | |
| "grad_norm": 0.12547020614147186, | |
| "learning_rate": 1.981396388672496e-05, | |
| "loss": 0.007, | |
| "step": 7210 | |
| }, | |
| { | |
| "epoch": 4.575411913814955, | |
| "grad_norm": 0.09598460793495178, | |
| "learning_rate": 1.9682315247255894e-05, | |
| "loss": 0.0077, | |
| "step": 7220 | |
| }, | |
| { | |
| "epoch": 4.581749049429658, | |
| "grad_norm": 0.09776268899440765, | |
| "learning_rate": 1.9550998157129946e-05, | |
| "loss": 0.0088, | |
| "step": 7230 | |
| }, | |
| { | |
| "epoch": 4.58808618504436, | |
| "grad_norm": 0.14646144211292267, | |
| "learning_rate": 1.942001405240979e-05, | |
| "loss": 0.0066, | |
| "step": 7240 | |
| }, | |
| { | |
| "epoch": 4.594423320659062, | |
| "grad_norm": 0.09714064747095108, | |
| "learning_rate": 1.928936436551661e-05, | |
| "loss": 0.0068, | |
| "step": 7250 | |
| }, | |
| { | |
| "epoch": 4.600760456273765, | |
| "grad_norm": 0.09656299650669098, | |
| "learning_rate": 1.9159050525214452e-05, | |
| "loss": 0.0088, | |
| "step": 7260 | |
| }, | |
| { | |
| "epoch": 4.607097591888467, | |
| "grad_norm": 0.11101258546113968, | |
| "learning_rate": 1.9029073956594606e-05, | |
| "loss": 0.0125, | |
| "step": 7270 | |
| }, | |
| { | |
| "epoch": 4.613434727503169, | |
| "grad_norm": 0.24712498486042023, | |
| "learning_rate": 1.8899436081059975e-05, | |
| "loss": 0.0079, | |
| "step": 7280 | |
| }, | |
| { | |
| "epoch": 4.619771863117871, | |
| "grad_norm": 0.27654388546943665, | |
| "learning_rate": 1.877013831630961e-05, | |
| "loss": 0.0086, | |
| "step": 7290 | |
| }, | |
| { | |
| "epoch": 4.6261089987325725, | |
| "grad_norm": 0.30745381116867065, | |
| "learning_rate": 1.8641182076323148e-05, | |
| "loss": 0.0115, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 4.632446134347275, | |
| "grad_norm": 0.21805402636528015, | |
| "learning_rate": 1.851256877134538e-05, | |
| "loss": 0.0086, | |
| "step": 7310 | |
| }, | |
| { | |
| "epoch": 4.638783269961977, | |
| "grad_norm": 0.14376069605350494, | |
| "learning_rate": 1.838429980787081e-05, | |
| "loss": 0.0078, | |
| "step": 7320 | |
| }, | |
| { | |
| "epoch": 4.645120405576679, | |
| "grad_norm": 0.15503360331058502, | |
| "learning_rate": 1.8256376588628238e-05, | |
| "loss": 0.0061, | |
| "step": 7330 | |
| }, | |
| { | |
| "epoch": 4.651457541191381, | |
| "grad_norm": 0.21702007949352264, | |
| "learning_rate": 1.8128800512565513e-05, | |
| "loss": 0.0078, | |
| "step": 7340 | |
| }, | |
| { | |
| "epoch": 4.657794676806084, | |
| "grad_norm": 0.16204503178596497, | |
| "learning_rate": 1.800157297483417e-05, | |
| "loss": 0.0084, | |
| "step": 7350 | |
| }, | |
| { | |
| "epoch": 4.664131812420786, | |
| "grad_norm": 0.16800753772258759, | |
| "learning_rate": 1.787469536677419e-05, | |
| "loss": 0.0123, | |
| "step": 7360 | |
| }, | |
| { | |
| "epoch": 4.670468948035488, | |
| "grad_norm": 0.26554012298583984, | |
| "learning_rate": 1.774816907589873e-05, | |
| "loss": 0.0155, | |
| "step": 7370 | |
| }, | |
| { | |
| "epoch": 4.67680608365019, | |
| "grad_norm": 0.18996752798557281, | |
| "learning_rate": 1.7621995485879062e-05, | |
| "loss": 0.0093, | |
| "step": 7380 | |
| }, | |
| { | |
| "epoch": 4.683143219264892, | |
| "grad_norm": 0.08091560751199722, | |
| "learning_rate": 1.749617597652934e-05, | |
| "loss": 0.0087, | |
| "step": 7390 | |
| }, | |
| { | |
| "epoch": 4.689480354879595, | |
| "grad_norm": 0.16652986407279968, | |
| "learning_rate": 1.7370711923791567e-05, | |
| "loss": 0.0068, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 4.695817490494297, | |
| "grad_norm": 0.06975802779197693, | |
| "learning_rate": 1.7245604699720535e-05, | |
| "loss": 0.0098, | |
| "step": 7410 | |
| }, | |
| { | |
| "epoch": 4.702154626108999, | |
| "grad_norm": 0.2281133234500885, | |
| "learning_rate": 1.712085567246878e-05, | |
| "loss": 0.0213, | |
| "step": 7420 | |
| }, | |
| { | |
| "epoch": 4.7084917617237005, | |
| "grad_norm": 0.12594124674797058, | |
| "learning_rate": 1.699646620627168e-05, | |
| "loss": 0.0082, | |
| "step": 7430 | |
| }, | |
| { | |
| "epoch": 4.714828897338403, | |
| "grad_norm": 0.18425187468528748, | |
| "learning_rate": 1.6872437661432517e-05, | |
| "loss": 0.0068, | |
| "step": 7440 | |
| }, | |
| { | |
| "epoch": 4.721166032953105, | |
| "grad_norm": 0.3101319670677185, | |
| "learning_rate": 1.6748771394307585e-05, | |
| "loss": 0.0076, | |
| "step": 7450 | |
| }, | |
| { | |
| "epoch": 4.727503168567807, | |
| "grad_norm": 0.1615813672542572, | |
| "learning_rate": 1.662546875729138e-05, | |
| "loss": 0.0078, | |
| "step": 7460 | |
| }, | |
| { | |
| "epoch": 4.733840304182509, | |
| "grad_norm": 0.22787398099899292, | |
| "learning_rate": 1.6502531098801753e-05, | |
| "loss": 0.0112, | |
| "step": 7470 | |
| }, | |
| { | |
| "epoch": 4.740177439797211, | |
| "grad_norm": 0.08283796161413193, | |
| "learning_rate": 1.637995976326527e-05, | |
| "loss": 0.0157, | |
| "step": 7480 | |
| }, | |
| { | |
| "epoch": 4.746514575411914, | |
| "grad_norm": 0.20619574189186096, | |
| "learning_rate": 1.62577560911024e-05, | |
| "loss": 0.0119, | |
| "step": 7490 | |
| }, | |
| { | |
| "epoch": 4.752851711026616, | |
| "grad_norm": 0.35674479603767395, | |
| "learning_rate": 1.6135921418712956e-05, | |
| "loss": 0.0087, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 4.759188846641318, | |
| "grad_norm": 0.2025870382785797, | |
| "learning_rate": 1.6014457078461353e-05, | |
| "loss": 0.0062, | |
| "step": 7510 | |
| }, | |
| { | |
| "epoch": 4.765525982256021, | |
| "grad_norm": 0.13968420028686523, | |
| "learning_rate": 1.5893364398662176e-05, | |
| "loss": 0.0132, | |
| "step": 7520 | |
| }, | |
| { | |
| "epoch": 4.771863117870723, | |
| "grad_norm": 0.18326647579669952, | |
| "learning_rate": 1.5772644703565565e-05, | |
| "loss": 0.0096, | |
| "step": 7530 | |
| }, | |
| { | |
| "epoch": 4.778200253485425, | |
| "grad_norm": 0.2547348141670227, | |
| "learning_rate": 1.5652299313342773e-05, | |
| "loss": 0.0084, | |
| "step": 7540 | |
| }, | |
| { | |
| "epoch": 4.784537389100127, | |
| "grad_norm": 0.12518715858459473, | |
| "learning_rate": 1.553232954407171e-05, | |
| "loss": 0.01, | |
| "step": 7550 | |
| }, | |
| { | |
| "epoch": 4.7908745247148286, | |
| "grad_norm": 0.12570436298847198, | |
| "learning_rate": 1.5412736707722537e-05, | |
| "loss": 0.0094, | |
| "step": 7560 | |
| }, | |
| { | |
| "epoch": 4.797211660329531, | |
| "grad_norm": 0.15311287343502045, | |
| "learning_rate": 1.5293522112143373e-05, | |
| "loss": 0.0057, | |
| "step": 7570 | |
| }, | |
| { | |
| "epoch": 4.803548795944233, | |
| "grad_norm": 0.08851443231105804, | |
| "learning_rate": 1.517468706104589e-05, | |
| "loss": 0.0083, | |
| "step": 7580 | |
| }, | |
| { | |
| "epoch": 4.809885931558935, | |
| "grad_norm": 0.14807093143463135, | |
| "learning_rate": 1.5056232853991209e-05, | |
| "loss": 0.0097, | |
| "step": 7590 | |
| }, | |
| { | |
| "epoch": 4.816223067173637, | |
| "grad_norm": 0.20004534721374512, | |
| "learning_rate": 1.4938160786375572e-05, | |
| "loss": 0.0101, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 4.82256020278834, | |
| "grad_norm": 0.2074967622756958, | |
| "learning_rate": 1.4820472149416154e-05, | |
| "loss": 0.0073, | |
| "step": 7610 | |
| }, | |
| { | |
| "epoch": 4.828897338403042, | |
| "grad_norm": 0.1131831631064415, | |
| "learning_rate": 1.470316823013707e-05, | |
| "loss": 0.0067, | |
| "step": 7620 | |
| }, | |
| { | |
| "epoch": 4.835234474017744, | |
| "grad_norm": 0.21209746599197388, | |
| "learning_rate": 1.4586250311355132e-05, | |
| "loss": 0.014, | |
| "step": 7630 | |
| }, | |
| { | |
| "epoch": 4.841571609632446, | |
| "grad_norm": 0.14775687456130981, | |
| "learning_rate": 1.4469719671666043e-05, | |
| "loss": 0.0081, | |
| "step": 7640 | |
| }, | |
| { | |
| "epoch": 4.847908745247148, | |
| "grad_norm": 0.20488443970680237, | |
| "learning_rate": 1.435357758543015e-05, | |
| "loss": 0.0073, | |
| "step": 7650 | |
| }, | |
| { | |
| "epoch": 4.854245880861851, | |
| "grad_norm": 0.06777448952198029, | |
| "learning_rate": 1.4237825322758736e-05, | |
| "loss": 0.0131, | |
| "step": 7660 | |
| }, | |
| { | |
| "epoch": 4.860583016476553, | |
| "grad_norm": 0.13396242260932922, | |
| "learning_rate": 1.412246414949997e-05, | |
| "loss": 0.005, | |
| "step": 7670 | |
| }, | |
| { | |
| "epoch": 4.866920152091255, | |
| "grad_norm": 0.21779701113700867, | |
| "learning_rate": 1.4007495327225162e-05, | |
| "loss": 0.0076, | |
| "step": 7680 | |
| }, | |
| { | |
| "epoch": 4.873257287705957, | |
| "grad_norm": 0.2239043265581131, | |
| "learning_rate": 1.389292011321498e-05, | |
| "loss": 0.0091, | |
| "step": 7690 | |
| }, | |
| { | |
| "epoch": 4.879594423320659, | |
| "grad_norm": 0.06960437446832657, | |
| "learning_rate": 1.3778739760445552e-05, | |
| "loss": 0.007, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 4.885931558935361, | |
| "grad_norm": 0.14277929067611694, | |
| "learning_rate": 1.3664955517574968e-05, | |
| "loss": 0.0078, | |
| "step": 7710 | |
| }, | |
| { | |
| "epoch": 4.892268694550063, | |
| "grad_norm": 0.17942270636558533, | |
| "learning_rate": 1.3551568628929434e-05, | |
| "loss": 0.0093, | |
| "step": 7720 | |
| }, | |
| { | |
| "epoch": 4.898605830164765, | |
| "grad_norm": 0.23902519047260284, | |
| "learning_rate": 1.343858033448982e-05, | |
| "loss": 0.0087, | |
| "step": 7730 | |
| }, | |
| { | |
| "epoch": 4.904942965779467, | |
| "grad_norm": 0.33525460958480835, | |
| "learning_rate": 1.3325991869878013e-05, | |
| "loss": 0.0064, | |
| "step": 7740 | |
| }, | |
| { | |
| "epoch": 4.91128010139417, | |
| "grad_norm": 0.11698520183563232, | |
| "learning_rate": 1.3213804466343421e-05, | |
| "loss": 0.0088, | |
| "step": 7750 | |
| }, | |
| { | |
| "epoch": 4.917617237008872, | |
| "grad_norm": 0.1842937171459198, | |
| "learning_rate": 1.3102019350749528e-05, | |
| "loss": 0.0092, | |
| "step": 7760 | |
| }, | |
| { | |
| "epoch": 4.923954372623574, | |
| "grad_norm": 0.09481093287467957, | |
| "learning_rate": 1.299063774556042e-05, | |
| "loss": 0.0059, | |
| "step": 7770 | |
| }, | |
| { | |
| "epoch": 4.930291508238277, | |
| "grad_norm": 0.17098549008369446, | |
| "learning_rate": 1.2879660868827508e-05, | |
| "loss": 0.0084, | |
| "step": 7780 | |
| }, | |
| { | |
| "epoch": 4.936628643852979, | |
| "grad_norm": 0.3938882648944855, | |
| "learning_rate": 1.2769089934176126e-05, | |
| "loss": 0.0113, | |
| "step": 7790 | |
| }, | |
| { | |
| "epoch": 4.942965779467681, | |
| "grad_norm": 0.21974030137062073, | |
| "learning_rate": 1.2658926150792322e-05, | |
| "loss": 0.0074, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 4.949302915082383, | |
| "grad_norm": 0.1278766393661499, | |
| "learning_rate": 1.2549170723409549e-05, | |
| "loss": 0.01, | |
| "step": 7810 | |
| }, | |
| { | |
| "epoch": 4.955640050697085, | |
| "grad_norm": 0.18058639764785767, | |
| "learning_rate": 1.243982485229559e-05, | |
| "loss": 0.0075, | |
| "step": 7820 | |
| }, | |
| { | |
| "epoch": 4.961977186311787, | |
| "grad_norm": 0.20735391974449158, | |
| "learning_rate": 1.233088973323937e-05, | |
| "loss": 0.0074, | |
| "step": 7830 | |
| }, | |
| { | |
| "epoch": 4.968314321926489, | |
| "grad_norm": 0.12043243646621704, | |
| "learning_rate": 1.2222366557537911e-05, | |
| "loss": 0.0093, | |
| "step": 7840 | |
| }, | |
| { | |
| "epoch": 4.974651457541191, | |
| "grad_norm": 0.28951793909072876, | |
| "learning_rate": 1.2114256511983274e-05, | |
| "loss": 0.007, | |
| "step": 7850 | |
| }, | |
| { | |
| "epoch": 4.980988593155893, | |
| "grad_norm": 0.1455940455198288, | |
| "learning_rate": 1.2006560778849578e-05, | |
| "loss": 0.0067, | |
| "step": 7860 | |
| }, | |
| { | |
| "epoch": 4.987325728770596, | |
| "grad_norm": 0.1419738233089447, | |
| "learning_rate": 1.1899280535880119e-05, | |
| "loss": 0.0115, | |
| "step": 7870 | |
| }, | |
| { | |
| "epoch": 4.993662864385298, | |
| "grad_norm": 0.3839736878871918, | |
| "learning_rate": 1.1792416956274444e-05, | |
| "loss": 0.0103, | |
| "step": 7880 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 0.12225103378295898, | |
| "learning_rate": 1.1685971208675539e-05, | |
| "loss": 0.0083, | |
| "step": 7890 | |
| }, | |
| { | |
| "epoch": 5.006337135614702, | |
| "grad_norm": 0.22479389607906342, | |
| "learning_rate": 1.157994445715706e-05, | |
| "loss": 0.0082, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 5.012674271229404, | |
| "grad_norm": 0.2502928674221039, | |
| "learning_rate": 1.1474337861210543e-05, | |
| "loss": 0.0073, | |
| "step": 7910 | |
| }, | |
| { | |
| "epoch": 5.019011406844107, | |
| "grad_norm": 0.11566631495952606, | |
| "learning_rate": 1.1369152575732822e-05, | |
| "loss": 0.0058, | |
| "step": 7920 | |
| }, | |
| { | |
| "epoch": 5.025348542458809, | |
| "grad_norm": 0.11840217560529709, | |
| "learning_rate": 1.1264389751013326e-05, | |
| "loss": 0.007, | |
| "step": 7930 | |
| }, | |
| { | |
| "epoch": 5.031685678073511, | |
| "grad_norm": 0.30334770679473877, | |
| "learning_rate": 1.1160050532721528e-05, | |
| "loss": 0.0116, | |
| "step": 7940 | |
| }, | |
| { | |
| "epoch": 5.038022813688213, | |
| "grad_norm": 0.17277342081069946, | |
| "learning_rate": 1.1056136061894384e-05, | |
| "loss": 0.0069, | |
| "step": 7950 | |
| }, | |
| { | |
| "epoch": 5.044359949302915, | |
| "grad_norm": 0.09852743148803711, | |
| "learning_rate": 1.095264747492391e-05, | |
| "loss": 0.0056, | |
| "step": 7960 | |
| }, | |
| { | |
| "epoch": 5.050697084917617, | |
| "grad_norm": 0.10726805031299591, | |
| "learning_rate": 1.0849585903544706e-05, | |
| "loss": 0.0148, | |
| "step": 7970 | |
| }, | |
| { | |
| "epoch": 5.057034220532319, | |
| "grad_norm": 0.08306515961885452, | |
| "learning_rate": 1.0746952474821614e-05, | |
| "loss": 0.005, | |
| "step": 7980 | |
| }, | |
| { | |
| "epoch": 5.063371356147021, | |
| "grad_norm": 0.13456346094608307, | |
| "learning_rate": 1.0644748311137376e-05, | |
| "loss": 0.0071, | |
| "step": 7990 | |
| }, | |
| { | |
| "epoch": 5.069708491761724, | |
| "grad_norm": 0.2890625, | |
| "learning_rate": 1.0542974530180327e-05, | |
| "loss": 0.0063, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 5.076045627376426, | |
| "grad_norm": 0.15506230294704437, | |
| "learning_rate": 1.0441632244932237e-05, | |
| "loss": 0.0077, | |
| "step": 8010 | |
| }, | |
| { | |
| "epoch": 5.082382762991128, | |
| "grad_norm": 0.17413881421089172, | |
| "learning_rate": 1.0340722563656107e-05, | |
| "loss": 0.0066, | |
| "step": 8020 | |
| }, | |
| { | |
| "epoch": 5.08871989860583, | |
| "grad_norm": 0.15269486606121063, | |
| "learning_rate": 1.0240246589884044e-05, | |
| "loss": 0.0056, | |
| "step": 8030 | |
| }, | |
| { | |
| "epoch": 5.095057034220532, | |
| "grad_norm": 0.17267774045467377, | |
| "learning_rate": 1.0140205422405214e-05, | |
| "loss": 0.0062, | |
| "step": 8040 | |
| }, | |
| { | |
| "epoch": 5.101394169835235, | |
| "grad_norm": 0.20491677522659302, | |
| "learning_rate": 1.0040600155253765e-05, | |
| "loss": 0.0082, | |
| "step": 8050 | |
| }, | |
| { | |
| "epoch": 5.107731305449937, | |
| "grad_norm": 0.15973028540611267, | |
| "learning_rate": 9.941431877696955e-06, | |
| "loss": 0.0095, | |
| "step": 8060 | |
| }, | |
| { | |
| "epoch": 5.114068441064639, | |
| "grad_norm": 0.13956964015960693, | |
| "learning_rate": 9.842701674223187e-06, | |
| "loss": 0.008, | |
| "step": 8070 | |
| }, | |
| { | |
| "epoch": 5.120405576679341, | |
| "grad_norm": 0.29668980836868286, | |
| "learning_rate": 9.744410624530148e-06, | |
| "loss": 0.009, | |
| "step": 8080 | |
| }, | |
| { | |
| "epoch": 5.126742712294043, | |
| "grad_norm": 0.1549568921327591, | |
| "learning_rate": 9.646559803512994e-06, | |
| "loss": 0.0073, | |
| "step": 8090 | |
| }, | |
| { | |
| "epoch": 5.133079847908745, | |
| "grad_norm": 0.15987129509449005, | |
| "learning_rate": 9.549150281252633e-06, | |
| "loss": 0.0078, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 5.139416983523447, | |
| "grad_norm": 0.0577310174703598, | |
| "learning_rate": 9.452183123004e-06, | |
| "loss": 0.0098, | |
| "step": 8110 | |
| }, | |
| { | |
| "epoch": 5.145754119138149, | |
| "grad_norm": 0.1426042914390564, | |
| "learning_rate": 9.355659389184396e-06, | |
| "loss": 0.0078, | |
| "step": 8120 | |
| }, | |
| { | |
| "epoch": 5.152091254752852, | |
| "grad_norm": 0.13074810802936554, | |
| "learning_rate": 9.259580135361929e-06, | |
| "loss": 0.0081, | |
| "step": 8130 | |
| }, | |
| { | |
| "epoch": 5.158428390367554, | |
| "grad_norm": 0.19135090708732605, | |
| "learning_rate": 9.163946412243896e-06, | |
| "loss": 0.0098, | |
| "step": 8140 | |
| }, | |
| { | |
| "epoch": 5.164765525982256, | |
| "grad_norm": 0.15256141126155853, | |
| "learning_rate": 9.068759265665384e-06, | |
| "loss": 0.0053, | |
| "step": 8150 | |
| }, | |
| { | |
| "epoch": 5.171102661596958, | |
| "grad_norm": 0.15117891132831573, | |
| "learning_rate": 8.974019736577777e-06, | |
| "loss": 0.0079, | |
| "step": 8160 | |
| }, | |
| { | |
| "epoch": 5.17743979721166, | |
| "grad_norm": 0.09071781486272812, | |
| "learning_rate": 8.879728861037384e-06, | |
| "loss": 0.0099, | |
| "step": 8170 | |
| }, | |
| { | |
| "epoch": 5.183776932826363, | |
| "grad_norm": 0.24567152559757233, | |
| "learning_rate": 8.785887670194138e-06, | |
| "loss": 0.0055, | |
| "step": 8180 | |
| }, | |
| { | |
| "epoch": 5.190114068441065, | |
| "grad_norm": 0.19762946665287018, | |
| "learning_rate": 8.692497190280224e-06, | |
| "loss": 0.0092, | |
| "step": 8190 | |
| }, | |
| { | |
| "epoch": 5.196451204055767, | |
| "grad_norm": 0.24934445321559906, | |
| "learning_rate": 8.599558442598998e-06, | |
| "loss": 0.0094, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 5.202788339670469, | |
| "grad_norm": 0.1780957281589508, | |
| "learning_rate": 8.507072443513702e-06, | |
| "loss": 0.0083, | |
| "step": 8210 | |
| }, | |
| { | |
| "epoch": 5.2091254752851714, | |
| "grad_norm": 0.23371122777462006, | |
| "learning_rate": 8.415040204436426e-06, | |
| "loss": 0.0082, | |
| "step": 8220 | |
| }, | |
| { | |
| "epoch": 5.215462610899873, | |
| "grad_norm": 0.06640614569187164, | |
| "learning_rate": 8.323462731816961e-06, | |
| "loss": 0.0052, | |
| "step": 8230 | |
| }, | |
| { | |
| "epoch": 5.221799746514575, | |
| "grad_norm": 0.118456169962883, | |
| "learning_rate": 8.232341027131885e-06, | |
| "loss": 0.0081, | |
| "step": 8240 | |
| }, | |
| { | |
| "epoch": 5.228136882129277, | |
| "grad_norm": 0.1477072387933731, | |
| "learning_rate": 8.141676086873572e-06, | |
| "loss": 0.0081, | |
| "step": 8250 | |
| }, | |
| { | |
| "epoch": 5.23447401774398, | |
| "grad_norm": 0.1423831284046173, | |
| "learning_rate": 8.051468902539272e-06, | |
| "loss": 0.0058, | |
| "step": 8260 | |
| }, | |
| { | |
| "epoch": 5.240811153358682, | |
| "grad_norm": 0.14859850704669952, | |
| "learning_rate": 7.96172046062032e-06, | |
| "loss": 0.0094, | |
| "step": 8270 | |
| }, | |
| { | |
| "epoch": 5.247148288973384, | |
| "grad_norm": 0.22598117589950562, | |
| "learning_rate": 7.872431742591268e-06, | |
| "loss": 0.0079, | |
| "step": 8280 | |
| }, | |
| { | |
| "epoch": 5.253485424588086, | |
| "grad_norm": 0.20482905209064484, | |
| "learning_rate": 7.783603724899257e-06, | |
| "loss": 0.0079, | |
| "step": 8290 | |
| }, | |
| { | |
| "epoch": 5.259822560202788, | |
| "grad_norm": 0.21788014471530914, | |
| "learning_rate": 7.695237378953223e-06, | |
| "loss": 0.0055, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 5.266159695817491, | |
| "grad_norm": 0.09685339778661728, | |
| "learning_rate": 7.607333671113409e-06, | |
| "loss": 0.0038, | |
| "step": 8310 | |
| }, | |
| { | |
| "epoch": 5.272496831432193, | |
| "grad_norm": 0.175571009516716, | |
| "learning_rate": 7.519893562680663e-06, | |
| "loss": 0.0062, | |
| "step": 8320 | |
| }, | |
| { | |
| "epoch": 5.278833967046895, | |
| "grad_norm": 0.20628724992275238, | |
| "learning_rate": 7.432918009885997e-06, | |
| "loss": 0.0088, | |
| "step": 8330 | |
| }, | |
| { | |
| "epoch": 5.285171102661597, | |
| "grad_norm": 0.0737687200307846, | |
| "learning_rate": 7.3464079638801365e-06, | |
| "loss": 0.0088, | |
| "step": 8340 | |
| }, | |
| { | |
| "epoch": 5.2915082382762995, | |
| "grad_norm": 0.10191112756729126, | |
| "learning_rate": 7.260364370723044e-06, | |
| "loss": 0.0081, | |
| "step": 8350 | |
| }, | |
| { | |
| "epoch": 5.297845373891001, | |
| "grad_norm": 0.35432806611061096, | |
| "learning_rate": 7.174788171373731e-06, | |
| "loss": 0.0096, | |
| "step": 8360 | |
| }, | |
| { | |
| "epoch": 5.304182509505703, | |
| "grad_norm": 0.11743071675300598, | |
| "learning_rate": 7.089680301679752e-06, | |
| "loss": 0.0132, | |
| "step": 8370 | |
| }, | |
| { | |
| "epoch": 5.310519645120405, | |
| "grad_norm": 0.15060439705848694, | |
| "learning_rate": 7.005041692367154e-06, | |
| "loss": 0.008, | |
| "step": 8380 | |
| }, | |
| { | |
| "epoch": 5.316856780735108, | |
| "grad_norm": 0.19431732594966888, | |
| "learning_rate": 6.92087326903022e-06, | |
| "loss": 0.0062, | |
| "step": 8390 | |
| }, | |
| { | |
| "epoch": 5.32319391634981, | |
| "grad_norm": 0.10982516407966614, | |
| "learning_rate": 6.837175952121306e-06, | |
| "loss": 0.0057, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 5.329531051964512, | |
| "grad_norm": 0.09768649935722351, | |
| "learning_rate": 6.753950656940905e-06, | |
| "loss": 0.0047, | |
| "step": 8410 | |
| }, | |
| { | |
| "epoch": 5.335868187579214, | |
| "grad_norm": 0.2184268981218338, | |
| "learning_rate": 6.671198293627479e-06, | |
| "loss": 0.0061, | |
| "step": 8420 | |
| }, | |
| { | |
| "epoch": 5.342205323193916, | |
| "grad_norm": 0.11367691308259964, | |
| "learning_rate": 6.588919767147639e-06, | |
| "loss": 0.0048, | |
| "step": 8430 | |
| }, | |
| { | |
| "epoch": 5.348542458808619, | |
| "grad_norm": 0.11176607757806778, | |
| "learning_rate": 6.5071159772861436e-06, | |
| "loss": 0.0092, | |
| "step": 8440 | |
| }, | |
| { | |
| "epoch": 5.354879594423321, | |
| "grad_norm": 0.12881191074848175, | |
| "learning_rate": 6.425787818636131e-06, | |
| "loss": 0.0056, | |
| "step": 8450 | |
| }, | |
| { | |
| "epoch": 5.361216730038023, | |
| "grad_norm": 0.19846558570861816, | |
| "learning_rate": 6.344936180589351e-06, | |
| "loss": 0.0059, | |
| "step": 8460 | |
| }, | |
| { | |
| "epoch": 5.367553865652725, | |
| "grad_norm": 0.2730921506881714, | |
| "learning_rate": 6.264561947326331e-06, | |
| "loss": 0.0074, | |
| "step": 8470 | |
| }, | |
| { | |
| "epoch": 5.3738910012674275, | |
| "grad_norm": 0.1230594590306282, | |
| "learning_rate": 6.184665997806832e-06, | |
| "loss": 0.0076, | |
| "step": 8480 | |
| }, | |
| { | |
| "epoch": 5.380228136882129, | |
| "grad_norm": 0.12817886471748352, | |
| "learning_rate": 6.1052492057601275e-06, | |
| "loss": 0.0069, | |
| "step": 8490 | |
| }, | |
| { | |
| "epoch": 5.386565272496831, | |
| "grad_norm": 0.08666050434112549, | |
| "learning_rate": 6.026312439675552e-06, | |
| "loss": 0.0059, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 5.392902408111533, | |
| "grad_norm": 0.2247517853975296, | |
| "learning_rate": 5.947856562792925e-06, | |
| "loss": 0.0147, | |
| "step": 8510 | |
| }, | |
| { | |
| "epoch": 5.399239543726236, | |
| "grad_norm": 0.08660010248422623, | |
| "learning_rate": 5.869882433093155e-06, | |
| "loss": 0.0073, | |
| "step": 8520 | |
| }, | |
| { | |
| "epoch": 5.405576679340938, | |
| "grad_norm": 0.0791560560464859, | |
| "learning_rate": 5.79239090328883e-06, | |
| "loss": 0.0061, | |
| "step": 8530 | |
| }, | |
| { | |
| "epoch": 5.41191381495564, | |
| "grad_norm": 0.07316572964191437, | |
| "learning_rate": 5.715382820814885e-06, | |
| "loss": 0.0084, | |
| "step": 8540 | |
| }, | |
| { | |
| "epoch": 5.418250950570342, | |
| "grad_norm": 0.12857115268707275, | |
| "learning_rate": 5.6388590278194096e-06, | |
| "loss": 0.0068, | |
| "step": 8550 | |
| }, | |
| { | |
| "epoch": 5.424588086185044, | |
| "grad_norm": 0.21100053191184998, | |
| "learning_rate": 5.562820361154314e-06, | |
| "loss": 0.0075, | |
| "step": 8560 | |
| }, | |
| { | |
| "epoch": 5.430925221799747, | |
| "grad_norm": 0.08549866825342178, | |
| "learning_rate": 5.48726765236629e-06, | |
| "loss": 0.006, | |
| "step": 8570 | |
| }, | |
| { | |
| "epoch": 5.437262357414449, | |
| "grad_norm": 0.15550892055034637, | |
| "learning_rate": 5.412201727687644e-06, | |
| "loss": 0.0063, | |
| "step": 8580 | |
| }, | |
| { | |
| "epoch": 5.443599493029151, | |
| "grad_norm": 0.18836095929145813, | |
| "learning_rate": 5.337623408027293e-06, | |
| "loss": 0.0087, | |
| "step": 8590 | |
| }, | |
| { | |
| "epoch": 5.449936628643853, | |
| "grad_norm": 0.08456358313560486, | |
| "learning_rate": 5.263533508961827e-06, | |
| "loss": 0.006, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 5.4562737642585555, | |
| "grad_norm": 0.054700568318367004, | |
| "learning_rate": 5.1899328407264855e-06, | |
| "loss": 0.0053, | |
| "step": 8610 | |
| }, | |
| { | |
| "epoch": 5.462610899873257, | |
| "grad_norm": 0.13627506792545319, | |
| "learning_rate": 5.116822208206396e-06, | |
| "loss": 0.006, | |
| "step": 8620 | |
| }, | |
| { | |
| "epoch": 5.468948035487959, | |
| "grad_norm": 0.1252082735300064, | |
| "learning_rate": 5.044202410927706e-06, | |
| "loss": 0.0125, | |
| "step": 8630 | |
| }, | |
| { | |
| "epoch": 5.475285171102661, | |
| "grad_norm": 0.2162148505449295, | |
| "learning_rate": 4.972074243048897e-06, | |
| "loss": 0.0065, | |
| "step": 8640 | |
| }, | |
| { | |
| "epoch": 5.481622306717364, | |
| "grad_norm": 0.12487296015024185, | |
| "learning_rate": 4.900438493352055e-06, | |
| "loss": 0.0063, | |
| "step": 8650 | |
| }, | |
| { | |
| "epoch": 5.487959442332066, | |
| "grad_norm": 0.14202018082141876, | |
| "learning_rate": 4.829295945234258e-06, | |
| "loss": 0.0063, | |
| "step": 8660 | |
| }, | |
| { | |
| "epoch": 5.494296577946768, | |
| "grad_norm": 0.18836723268032074, | |
| "learning_rate": 4.758647376699032e-06, | |
| "loss": 0.0069, | |
| "step": 8670 | |
| }, | |
| { | |
| "epoch": 5.50063371356147, | |
| "grad_norm": 0.2227470427751541, | |
| "learning_rate": 4.688493560347773e-06, | |
| "loss": 0.0111, | |
| "step": 8680 | |
| }, | |
| { | |
| "epoch": 5.506970849176172, | |
| "grad_norm": 0.1131884977221489, | |
| "learning_rate": 4.618835263371396e-06, | |
| "loss": 0.0054, | |
| "step": 8690 | |
| }, | |
| { | |
| "epoch": 5.513307984790875, | |
| "grad_norm": 0.08110499382019043, | |
| "learning_rate": 4.549673247541875e-06, | |
| "loss": 0.0039, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 5.519645120405577, | |
| "grad_norm": 0.2664946913719177, | |
| "learning_rate": 4.48100826920394e-06, | |
| "loss": 0.0055, | |
| "step": 8710 | |
| }, | |
| { | |
| "epoch": 5.525982256020279, | |
| "grad_norm": 0.09621760994195938, | |
| "learning_rate": 4.412841079266777e-06, | |
| "loss": 0.0066, | |
| "step": 8720 | |
| }, | |
| { | |
| "epoch": 5.532319391634981, | |
| "grad_norm": 0.07459976524114609, | |
| "learning_rate": 4.3451724231958644e-06, | |
| "loss": 0.0103, | |
| "step": 8730 | |
| }, | |
| { | |
| "epoch": 5.5386565272496835, | |
| "grad_norm": 0.252323180437088, | |
| "learning_rate": 4.27800304100478e-06, | |
| "loss": 0.0041, | |
| "step": 8740 | |
| }, | |
| { | |
| "epoch": 5.544993662864385, | |
| "grad_norm": 0.11829400062561035, | |
| "learning_rate": 4.2113336672471245e-06, | |
| "loss": 0.0042, | |
| "step": 8750 | |
| }, | |
| { | |
| "epoch": 5.551330798479087, | |
| "grad_norm": 0.07013269513845444, | |
| "learning_rate": 4.145165031008508e-06, | |
| "loss": 0.0047, | |
| "step": 8760 | |
| }, | |
| { | |
| "epoch": 5.557667934093789, | |
| "grad_norm": 0.2011323720216751, | |
| "learning_rate": 4.079497855898501e-06, | |
| "loss": 0.0067, | |
| "step": 8770 | |
| }, | |
| { | |
| "epoch": 5.564005069708491, | |
| "grad_norm": 0.06639205664396286, | |
| "learning_rate": 4.01433286004283e-06, | |
| "loss": 0.0086, | |
| "step": 8780 | |
| }, | |
| { | |
| "epoch": 5.570342205323194, | |
| "grad_norm": 0.24140506982803345, | |
| "learning_rate": 3.949670756075447e-06, | |
| "loss": 0.0101, | |
| "step": 8790 | |
| }, | |
| { | |
| "epoch": 5.576679340937896, | |
| "grad_norm": 0.1788935363292694, | |
| "learning_rate": 3.885512251130763e-06, | |
| "loss": 0.0068, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 5.583016476552598, | |
| "grad_norm": 0.10148416459560394, | |
| "learning_rate": 3.821858046835913e-06, | |
| "loss": 0.011, | |
| "step": 8810 | |
| }, | |
| { | |
| "epoch": 5.589353612167301, | |
| "grad_norm": 0.25111889839172363, | |
| "learning_rate": 3.75870883930306e-06, | |
| "loss": 0.0052, | |
| "step": 8820 | |
| }, | |
| { | |
| "epoch": 5.595690747782003, | |
| "grad_norm": 0.1619209200143814, | |
| "learning_rate": 3.696065319121833e-06, | |
| "loss": 0.0061, | |
| "step": 8830 | |
| }, | |
| { | |
| "epoch": 5.602027883396705, | |
| "grad_norm": 0.1424710899591446, | |
| "learning_rate": 3.6339281713517303e-06, | |
| "loss": 0.0147, | |
| "step": 8840 | |
| }, | |
| { | |
| "epoch": 5.608365019011407, | |
| "grad_norm": 0.093357153236866, | |
| "learning_rate": 3.5722980755146517e-06, | |
| "loss": 0.01, | |
| "step": 8850 | |
| }, | |
| { | |
| "epoch": 5.614702154626109, | |
| "grad_norm": 0.09066253155469894, | |
| "learning_rate": 3.511175705587433e-06, | |
| "loss": 0.0064, | |
| "step": 8860 | |
| }, | |
| { | |
| "epoch": 5.6210392902408115, | |
| "grad_norm": 0.24886588752269745, | |
| "learning_rate": 3.4505617299945336e-06, | |
| "loss": 0.0072, | |
| "step": 8870 | |
| }, | |
| { | |
| "epoch": 5.6273764258555135, | |
| "grad_norm": 0.22231824696063995, | |
| "learning_rate": 3.390456811600673e-06, | |
| "loss": 0.0111, | |
| "step": 8880 | |
| }, | |
| { | |
| "epoch": 5.633713561470215, | |
| "grad_norm": 0.11994162201881409, | |
| "learning_rate": 3.3308616077036115e-06, | |
| "loss": 0.0088, | |
| "step": 8890 | |
| }, | |
| { | |
| "epoch": 5.640050697084917, | |
| "grad_norm": 0.11728494614362717, | |
| "learning_rate": 3.271776770026963e-06, | |
| "loss": 0.0095, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 5.64638783269962, | |
| "grad_norm": 0.13038089871406555, | |
| "learning_rate": 3.213202944713023e-06, | |
| "loss": 0.0053, | |
| "step": 8910 | |
| }, | |
| { | |
| "epoch": 5.652724968314322, | |
| "grad_norm": 0.09141723066568375, | |
| "learning_rate": 3.155140772315773e-06, | |
| "loss": 0.0062, | |
| "step": 8920 | |
| }, | |
| { | |
| "epoch": 5.659062103929024, | |
| "grad_norm": 0.07316479086875916, | |
| "learning_rate": 3.0975908877938277e-06, | |
| "loss": 0.0062, | |
| "step": 8930 | |
| }, | |
| { | |
| "epoch": 5.665399239543726, | |
| "grad_norm": 0.18643398582935333, | |
| "learning_rate": 3.040553920503503e-06, | |
| "loss": 0.0065, | |
| "step": 8940 | |
| }, | |
| { | |
| "epoch": 5.671736375158428, | |
| "grad_norm": 0.151958167552948, | |
| "learning_rate": 2.9840304941919415e-06, | |
| "loss": 0.0055, | |
| "step": 8950 | |
| }, | |
| { | |
| "epoch": 5.678073510773131, | |
| "grad_norm": 0.2611534297466278, | |
| "learning_rate": 2.928021226990263e-06, | |
| "loss": 0.0075, | |
| "step": 8960 | |
| }, | |
| { | |
| "epoch": 5.684410646387833, | |
| "grad_norm": 0.16426794230937958, | |
| "learning_rate": 2.8725267314068495e-06, | |
| "loss": 0.0046, | |
| "step": 8970 | |
| }, | |
| { | |
| "epoch": 5.690747782002535, | |
| "grad_norm": 0.11258076876401901, | |
| "learning_rate": 2.817547614320615e-06, | |
| "loss": 0.0095, | |
| "step": 8980 | |
| }, | |
| { | |
| "epoch": 5.697084917617237, | |
| "grad_norm": 0.15174216032028198, | |
| "learning_rate": 2.7630844769743757e-06, | |
| "loss": 0.0069, | |
| "step": 8990 | |
| }, | |
| { | |
| "epoch": 5.7034220532319395, | |
| "grad_norm": 0.15580610930919647, | |
| "learning_rate": 2.7091379149682685e-06, | |
| "loss": 0.0079, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 5.7097591888466415, | |
| "grad_norm": 0.12744936347007751, | |
| "learning_rate": 2.6557085182532582e-06, | |
| "loss": 0.0086, | |
| "step": 9010 | |
| }, | |
| { | |
| "epoch": 5.716096324461343, | |
| "grad_norm": 0.1542847603559494, | |
| "learning_rate": 2.602796871124663e-06, | |
| "loss": 0.0096, | |
| "step": 9020 | |
| }, | |
| { | |
| "epoch": 5.722433460076045, | |
| "grad_norm": 0.13958968222141266, | |
| "learning_rate": 2.5504035522157854e-06, | |
| "loss": 0.0089, | |
| "step": 9030 | |
| }, | |
| { | |
| "epoch": 5.728770595690747, | |
| "grad_norm": 0.20519228279590607, | |
| "learning_rate": 2.4985291344915674e-06, | |
| "loss": 0.0061, | |
| "step": 9040 | |
| }, | |
| { | |
| "epoch": 5.73510773130545, | |
| "grad_norm": 0.11365640163421631, | |
| "learning_rate": 2.4471741852423237e-06, | |
| "loss": 0.0046, | |
| "step": 9050 | |
| }, | |
| { | |
| "epoch": 5.741444866920152, | |
| "grad_norm": 0.0870782658457756, | |
| "learning_rate": 2.3963392660775575e-06, | |
| "loss": 0.0079, | |
| "step": 9060 | |
| }, | |
| { | |
| "epoch": 5.747782002534854, | |
| "grad_norm": 0.07872149348258972, | |
| "learning_rate": 2.3460249329197824e-06, | |
| "loss": 0.0055, | |
| "step": 9070 | |
| }, | |
| { | |
| "epoch": 5.754119138149557, | |
| "grad_norm": 0.08804658055305481, | |
| "learning_rate": 2.296231735998511e-06, | |
| "loss": 0.007, | |
| "step": 9080 | |
| }, | |
| { | |
| "epoch": 5.760456273764259, | |
| "grad_norm": 0.05982697755098343, | |
| "learning_rate": 2.2469602198441573e-06, | |
| "loss": 0.0044, | |
| "step": 9090 | |
| }, | |
| { | |
| "epoch": 5.766793409378961, | |
| "grad_norm": 0.10259458422660828, | |
| "learning_rate": 2.1982109232821178e-06, | |
| "loss": 0.006, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 5.773130544993663, | |
| "grad_norm": 0.1546311378479004, | |
| "learning_rate": 2.149984379426906e-06, | |
| "loss": 0.0067, | |
| "step": 9110 | |
| }, | |
| { | |
| "epoch": 5.779467680608365, | |
| "grad_norm": 0.14351686835289001, | |
| "learning_rate": 2.102281115676258e-06, | |
| "loss": 0.0056, | |
| "step": 9120 | |
| }, | |
| { | |
| "epoch": 5.7858048162230675, | |
| "grad_norm": 0.05316438898444176, | |
| "learning_rate": 2.0551016537054493e-06, | |
| "loss": 0.0061, | |
| "step": 9130 | |
| }, | |
| { | |
| "epoch": 5.7921419518377695, | |
| "grad_norm": 0.0845993310213089, | |
| "learning_rate": 2.008446509461498e-06, | |
| "loss": 0.006, | |
| "step": 9140 | |
| }, | |
| { | |
| "epoch": 5.798479087452471, | |
| "grad_norm": 0.156995952129364, | |
| "learning_rate": 1.962316193157593e-06, | |
| "loss": 0.0054, | |
| "step": 9150 | |
| }, | |
| { | |
| "epoch": 5.804816223067173, | |
| "grad_norm": 0.20765309035778046, | |
| "learning_rate": 1.91671120926748e-06, | |
| "loss": 0.0049, | |
| "step": 9160 | |
| }, | |
| { | |
| "epoch": 5.811153358681876, | |
| "grad_norm": 0.15031087398529053, | |
| "learning_rate": 1.8716320565199618e-06, | |
| "loss": 0.0069, | |
| "step": 9170 | |
| }, | |
| { | |
| "epoch": 5.817490494296578, | |
| "grad_norm": 0.08898857980966568, | |
| "learning_rate": 1.8270792278934302e-06, | |
| "loss": 0.0052, | |
| "step": 9180 | |
| }, | |
| { | |
| "epoch": 5.82382762991128, | |
| "grad_norm": 0.16047555208206177, | |
| "learning_rate": 1.7830532106104747e-06, | |
| "loss": 0.0061, | |
| "step": 9190 | |
| }, | |
| { | |
| "epoch": 5.830164765525982, | |
| "grad_norm": 0.24277397990226746, | |
| "learning_rate": 1.7395544861325718e-06, | |
| "loss": 0.0061, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 5.836501901140684, | |
| "grad_norm": 0.1472398340702057, | |
| "learning_rate": 1.696583530154794e-06, | |
| "loss": 0.0123, | |
| "step": 9210 | |
| }, | |
| { | |
| "epoch": 5.842839036755387, | |
| "grad_norm": 0.15417790412902832, | |
| "learning_rate": 1.6541408126006463e-06, | |
| "loss": 0.0058, | |
| "step": 9220 | |
| }, | |
| { | |
| "epoch": 5.849176172370089, | |
| "grad_norm": 0.10276266187429428, | |
| "learning_rate": 1.6122267976168781e-06, | |
| "loss": 0.0076, | |
| "step": 9230 | |
| }, | |
| { | |
| "epoch": 5.855513307984791, | |
| "grad_norm": 0.22714407742023468, | |
| "learning_rate": 1.5708419435684462e-06, | |
| "loss": 0.0079, | |
| "step": 9240 | |
| }, | |
| { | |
| "epoch": 5.861850443599493, | |
| "grad_norm": 0.08724337071180344, | |
| "learning_rate": 1.5299867030334814e-06, | |
| "loss": 0.006, | |
| "step": 9250 | |
| }, | |
| { | |
| "epoch": 5.8681875792141955, | |
| "grad_norm": 0.1354474574327469, | |
| "learning_rate": 1.4896615227983468e-06, | |
| "loss": 0.0063, | |
| "step": 9260 | |
| }, | |
| { | |
| "epoch": 5.8745247148288975, | |
| "grad_norm": 0.09567315876483917, | |
| "learning_rate": 1.4498668438527597e-06, | |
| "loss": 0.0073, | |
| "step": 9270 | |
| }, | |
| { | |
| "epoch": 5.880861850443599, | |
| "grad_norm": 0.25440752506256104, | |
| "learning_rate": 1.4106031013849496e-06, | |
| "loss": 0.0059, | |
| "step": 9280 | |
| }, | |
| { | |
| "epoch": 5.887198986058301, | |
| "grad_norm": 0.046442631632089615, | |
| "learning_rate": 1.3718707247769135e-06, | |
| "loss": 0.0085, | |
| "step": 9290 | |
| }, | |
| { | |
| "epoch": 5.893536121673003, | |
| "grad_norm": 0.202633798122406, | |
| "learning_rate": 1.333670137599713e-06, | |
| "loss": 0.0078, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 5.899873257287706, | |
| "grad_norm": 0.11459842324256897, | |
| "learning_rate": 1.2960017576088446e-06, | |
| "loss": 0.0049, | |
| "step": 9310 | |
| }, | |
| { | |
| "epoch": 5.906210392902408, | |
| "grad_norm": 0.06873945146799088, | |
| "learning_rate": 1.2588659967397e-06, | |
| "loss": 0.0172, | |
| "step": 9320 | |
| }, | |
| { | |
| "epoch": 5.91254752851711, | |
| "grad_norm": 0.17459562420845032, | |
| "learning_rate": 1.222263261102985e-06, | |
| "loss": 0.0071, | |
| "step": 9330 | |
| }, | |
| { | |
| "epoch": 5.918884664131813, | |
| "grad_norm": 0.15463189780712128, | |
| "learning_rate": 1.1861939509803687e-06, | |
| "loss": 0.0068, | |
| "step": 9340 | |
| }, | |
| { | |
| "epoch": 5.925221799746515, | |
| "grad_norm": 0.052368469536304474, | |
| "learning_rate": 1.1506584608200367e-06, | |
| "loss": 0.0063, | |
| "step": 9350 | |
| }, | |
| { | |
| "epoch": 5.931558935361217, | |
| "grad_norm": 0.08080845326185226, | |
| "learning_rate": 1.1156571792324211e-06, | |
| "loss": 0.0037, | |
| "step": 9360 | |
| }, | |
| { | |
| "epoch": 5.937896070975919, | |
| "grad_norm": 0.1452123522758484, | |
| "learning_rate": 1.0811904889859336e-06, | |
| "loss": 0.0106, | |
| "step": 9370 | |
| }, | |
| { | |
| "epoch": 5.944233206590621, | |
| "grad_norm": 0.18937888741493225, | |
| "learning_rate": 1.0472587670027678e-06, | |
| "loss": 0.0045, | |
| "step": 9380 | |
| }, | |
| { | |
| "epoch": 5.9505703422053235, | |
| "grad_norm": 0.1064092218875885, | |
| "learning_rate": 1.0138623843548078e-06, | |
| "loss": 0.01, | |
| "step": 9390 | |
| }, | |
| { | |
| "epoch": 5.9569074778200255, | |
| "grad_norm": 0.15949054062366486, | |
| "learning_rate": 9.810017062595322e-07, | |
| "loss": 0.0093, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 5.9632446134347274, | |
| "grad_norm": 0.11651349812746048, | |
| "learning_rate": 9.486770920760668e-07, | |
| "loss": 0.0062, | |
| "step": 9410 | |
| }, | |
| { | |
| "epoch": 5.969581749049429, | |
| "grad_norm": 0.16830092668533325, | |
| "learning_rate": 9.168888953011989e-07, | |
| "loss": 0.005, | |
| "step": 9420 | |
| }, | |
| { | |
| "epoch": 5.975918884664132, | |
| "grad_norm": 0.08310183882713318, | |
| "learning_rate": 8.856374635655695e-07, | |
| "loss": 0.0059, | |
| "step": 9430 | |
| }, | |
| { | |
| "epoch": 5.982256020278834, | |
| "grad_norm": 0.12785309553146362, | |
| "learning_rate": 8.549231386298151e-07, | |
| "loss": 0.0097, | |
| "step": 9440 | |
| }, | |
| { | |
| "epoch": 5.988593155893536, | |
| "grad_norm": 0.2574361264705658, | |
| "learning_rate": 8.247462563808817e-07, | |
| "loss": 0.0071, | |
| "step": 9450 | |
| }, | |
| { | |
| "epoch": 5.994930291508238, | |
| "grad_norm": 0.31402966380119324, | |
| "learning_rate": 7.951071468283167e-07, | |
| "loss": 0.0065, | |
| "step": 9460 | |
| }, | |
| { | |
| "epoch": 6.00126742712294, | |
| "grad_norm": 0.1356462836265564, | |
| "learning_rate": 7.66006134100672e-07, | |
| "loss": 0.0058, | |
| "step": 9470 | |
| }, | |
| { | |
| "epoch": 6.007604562737643, | |
| "grad_norm": 0.08312007039785385, | |
| "learning_rate": 7.374435364419674e-07, | |
| "loss": 0.004, | |
| "step": 9480 | |
| }, | |
| { | |
| "epoch": 6.013941698352345, | |
| "grad_norm": 0.25194162130355835, | |
| "learning_rate": 7.094196662081831e-07, | |
| "loss": 0.0105, | |
| "step": 9490 | |
| }, | |
| { | |
| "epoch": 6.020278833967047, | |
| "grad_norm": 0.11876359581947327, | |
| "learning_rate": 6.819348298638839e-07, | |
| "loss": 0.005, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 6.026615969581749, | |
| "grad_norm": 0.11712250113487244, | |
| "learning_rate": 6.549893279788277e-07, | |
| "loss": 0.0087, | |
| "step": 9510 | |
| }, | |
| { | |
| "epoch": 6.032953105196452, | |
| "grad_norm": 0.24556323885917664, | |
| "learning_rate": 6.285834552247128e-07, | |
| "loss": 0.0059, | |
| "step": 9520 | |
| }, | |
| { | |
| "epoch": 6.0392902408111535, | |
| "grad_norm": 0.12040778994560242, | |
| "learning_rate": 6.027175003719354e-07, | |
| "loss": 0.0062, | |
| "step": 9530 | |
| }, | |
| { | |
| "epoch": 6.0456273764258555, | |
| "grad_norm": 0.07460421323776245, | |
| "learning_rate": 5.773917462864264e-07, | |
| "loss": 0.0091, | |
| "step": 9540 | |
| }, | |
| { | |
| "epoch": 6.051964512040557, | |
| "grad_norm": 0.18476122617721558, | |
| "learning_rate": 5.526064699265753e-07, | |
| "loss": 0.01, | |
| "step": 9550 | |
| }, | |
| { | |
| "epoch": 6.05830164765526, | |
| "grad_norm": 0.06253950297832489, | |
| "learning_rate": 5.283619423401998e-07, | |
| "loss": 0.0076, | |
| "step": 9560 | |
| }, | |
| { | |
| "epoch": 6.064638783269962, | |
| "grad_norm": 0.07986325025558472, | |
| "learning_rate": 5.046584286615697e-07, | |
| "loss": 0.0111, | |
| "step": 9570 | |
| }, | |
| { | |
| "epoch": 6.070975918884664, | |
| "grad_norm": 0.25652700662612915, | |
| "learning_rate": 4.814961881085045e-07, | |
| "loss": 0.004, | |
| "step": 9580 | |
| }, | |
| { | |
| "epoch": 6.077313054499366, | |
| "grad_norm": 0.10732295364141464, | |
| "learning_rate": 4.5887547397955864e-07, | |
| "loss": 0.0052, | |
| "step": 9590 | |
| }, | |
| { | |
| "epoch": 6.083650190114068, | |
| "grad_norm": 0.09151678532361984, | |
| "learning_rate": 4.367965336512403e-07, | |
| "loss": 0.0109, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 6.089987325728771, | |
| "grad_norm": 0.057842839509248734, | |
| "learning_rate": 4.1525960857530243e-07, | |
| "loss": 0.0059, | |
| "step": 9610 | |
| }, | |
| { | |
| "epoch": 6.096324461343473, | |
| "grad_norm": 0.1317504197359085, | |
| "learning_rate": 3.9426493427611177e-07, | |
| "loss": 0.0067, | |
| "step": 9620 | |
| }, | |
| { | |
| "epoch": 6.102661596958175, | |
| "grad_norm": 0.0697597935795784, | |
| "learning_rate": 3.738127403480507e-07, | |
| "loss": 0.0064, | |
| "step": 9630 | |
| }, | |
| { | |
| "epoch": 6.108998732572877, | |
| "grad_norm": 0.3189285695552826, | |
| "learning_rate": 3.5390325045304706e-07, | |
| "loss": 0.0066, | |
| "step": 9640 | |
| }, | |
| { | |
| "epoch": 6.11533586818758, | |
| "grad_norm": 0.07586175948381424, | |
| "learning_rate": 3.3453668231809286e-07, | |
| "loss": 0.0061, | |
| "step": 9650 | |
| }, | |
| { | |
| "epoch": 6.1216730038022815, | |
| "grad_norm": 0.28144219517707825, | |
| "learning_rate": 3.157132477328628e-07, | |
| "loss": 0.0061, | |
| "step": 9660 | |
| }, | |
| { | |
| "epoch": 6.1280101394169835, | |
| "grad_norm": 0.21867600083351135, | |
| "learning_rate": 2.9743315254743833e-07, | |
| "loss": 0.016, | |
| "step": 9670 | |
| }, | |
| { | |
| "epoch": 6.134347275031685, | |
| "grad_norm": 0.08317314088344574, | |
| "learning_rate": 2.796965966699927e-07, | |
| "loss": 0.0143, | |
| "step": 9680 | |
| }, | |
| { | |
| "epoch": 6.140684410646388, | |
| "grad_norm": 0.13135367631912231, | |
| "learning_rate": 2.625037740646763e-07, | |
| "loss": 0.0058, | |
| "step": 9690 | |
| }, | |
| { | |
| "epoch": 6.14702154626109, | |
| "grad_norm": 0.06645505130290985, | |
| "learning_rate": 2.458548727494292e-07, | |
| "loss": 0.0109, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 6.153358681875792, | |
| "grad_norm": 0.09750212728977203, | |
| "learning_rate": 2.2975007479397738e-07, | |
| "loss": 0.0071, | |
| "step": 9710 | |
| }, | |
| { | |
| "epoch": 6.159695817490494, | |
| "grad_norm": 0.09561249613761902, | |
| "learning_rate": 2.1418955631781202e-07, | |
| "loss": 0.0076, | |
| "step": 9720 | |
| }, | |
| { | |
| "epoch": 6.166032953105196, | |
| "grad_norm": 0.1740742325782776, | |
| "learning_rate": 1.9917348748826335e-07, | |
| "loss": 0.006, | |
| "step": 9730 | |
| }, | |
| { | |
| "epoch": 6.172370088719899, | |
| "grad_norm": 0.18339525163173676, | |
| "learning_rate": 1.847020325186577e-07, | |
| "loss": 0.006, | |
| "step": 9740 | |
| }, | |
| { | |
| "epoch": 6.178707224334601, | |
| "grad_norm": 0.06436607986688614, | |
| "learning_rate": 1.7077534966650766e-07, | |
| "loss": 0.0122, | |
| "step": 9750 | |
| }, | |
| { | |
| "epoch": 6.185044359949303, | |
| "grad_norm": 0.05160180851817131, | |
| "learning_rate": 1.5739359123178587e-07, | |
| "loss": 0.008, | |
| "step": 9760 | |
| }, | |
| { | |
| "epoch": 6.191381495564005, | |
| "grad_norm": 0.0478476956486702, | |
| "learning_rate": 1.4455690355525964e-07, | |
| "loss": 0.0054, | |
| "step": 9770 | |
| }, | |
| { | |
| "epoch": 6.197718631178708, | |
| "grad_norm": 0.13458269834518433, | |
| "learning_rate": 1.3226542701689215e-07, | |
| "loss": 0.0051, | |
| "step": 9780 | |
| }, | |
| { | |
| "epoch": 6.2040557667934095, | |
| "grad_norm": 0.14054587483406067, | |
| "learning_rate": 1.2051929603428825e-07, | |
| "loss": 0.0066, | |
| "step": 9790 | |
| }, | |
| { | |
| "epoch": 6.2103929024081115, | |
| "grad_norm": 0.3035231828689575, | |
| "learning_rate": 1.0931863906127327e-07, | |
| "loss": 0.0057, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 6.216730038022813, | |
| "grad_norm": 0.3359600901603699, | |
| "learning_rate": 9.866357858642205e-08, | |
| "loss": 0.0088, | |
| "step": 9810 | |
| }, | |
| { | |
| "epoch": 6.223067173637516, | |
| "grad_norm": 0.13854598999023438, | |
| "learning_rate": 8.855423113177664e-08, | |
| "loss": 0.0057, | |
| "step": 9820 | |
| }, | |
| { | |
| "epoch": 6.229404309252218, | |
| "grad_norm": 0.061560165137052536, | |
| "learning_rate": 7.899070725153613e-08, | |
| "loss": 0.0047, | |
| "step": 9830 | |
| }, | |
| { | |
| "epoch": 6.23574144486692, | |
| "grad_norm": 0.1461103856563568, | |
| "learning_rate": 6.997311153086883e-08, | |
| "loss": 0.004, | |
| "step": 9840 | |
| }, | |
| { | |
| "epoch": 6.242078580481622, | |
| "grad_norm": 0.3121045231819153, | |
| "learning_rate": 6.150154258476315e-08, | |
| "loss": 0.0073, | |
| "step": 9850 | |
| }, | |
| { | |
| "epoch": 6.248415716096324, | |
| "grad_norm": 0.18830816447734833, | |
| "learning_rate": 5.3576093056922906e-08, | |
| "loss": 0.0067, | |
| "step": 9860 | |
| }, | |
| { | |
| "epoch": 6.254752851711027, | |
| "grad_norm": 0.060759589076042175, | |
| "learning_rate": 4.619684961881254e-08, | |
| "loss": 0.005, | |
| "step": 9870 | |
| }, | |
| { | |
| "epoch": 6.261089987325729, | |
| "grad_norm": 0.1718645989894867, | |
| "learning_rate": 3.936389296864129e-08, | |
| "loss": 0.0049, | |
| "step": 9880 | |
| }, | |
| { | |
| "epoch": 6.267427122940431, | |
| "grad_norm": 0.09834865480661392, | |
| "learning_rate": 3.3077297830541584e-08, | |
| "loss": 0.0066, | |
| "step": 9890 | |
| }, | |
| { | |
| "epoch": 6.273764258555133, | |
| "grad_norm": 0.11674518138170242, | |
| "learning_rate": 2.7337132953697554e-08, | |
| "loss": 0.0086, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 6.280101394169836, | |
| "grad_norm": 0.17682595551013947, | |
| "learning_rate": 2.214346111164556e-08, | |
| "loss": 0.0065, | |
| "step": 9910 | |
| }, | |
| { | |
| "epoch": 6.2864385297845375, | |
| "grad_norm": 0.07229286432266235, | |
| "learning_rate": 1.749633910153592e-08, | |
| "loss": 0.0038, | |
| "step": 9920 | |
| }, | |
| { | |
| "epoch": 6.2927756653992395, | |
| "grad_norm": 0.083204485476017, | |
| "learning_rate": 1.3395817743561134e-08, | |
| "loss": 0.0057, | |
| "step": 9930 | |
| }, | |
| { | |
| "epoch": 6.299112801013941, | |
| "grad_norm": 0.2328864485025406, | |
| "learning_rate": 9.841941880361916e-09, | |
| "loss": 0.0108, | |
| "step": 9940 | |
| }, | |
| { | |
| "epoch": 6.305449936628644, | |
| "grad_norm": 0.1641470342874527, | |
| "learning_rate": 6.834750376549792e-09, | |
| "loss": 0.007, | |
| "step": 9950 | |
| }, | |
| { | |
| "epoch": 6.311787072243346, | |
| "grad_norm": 0.08585844933986664, | |
| "learning_rate": 4.3742761183018784e-09, | |
| "loss": 0.012, | |
| "step": 9960 | |
| }, | |
| { | |
| "epoch": 6.318124207858048, | |
| "grad_norm": 0.10597538948059082, | |
| "learning_rate": 2.4605460129556445e-09, | |
| "loss": 0.004, | |
| "step": 9970 | |
| }, | |
| { | |
| "epoch": 6.32446134347275, | |
| "grad_norm": 0.21842822432518005, | |
| "learning_rate": 1.0935809887702154e-09, | |
| "loss": 0.006, | |
| "step": 9980 | |
| }, | |
| { | |
| "epoch": 6.330798479087452, | |
| "grad_norm": 0.0765758827328682, | |
| "learning_rate": 2.7339599464326627e-10, | |
| "loss": 0.0059, | |
| "step": 9990 | |
| }, | |
| { | |
| "epoch": 6.337135614702155, | |
| "grad_norm": 0.09744177758693695, | |
| "learning_rate": 0.0, | |
| "loss": 0.0057, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 6.337135614702155, | |
| "step": 10000, | |
| "total_flos": 3.5060140843731366e+17, | |
| "train_loss": 0.02414526521936059, | |
| "train_runtime": 5814.7906, | |
| "train_samples_per_second": 27.516, | |
| "train_steps_per_second": 1.72 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 10000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 7, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.5060140843731366e+17, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |