| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 1911, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0005232862375719519, | |
| "grad_norm": 0.7093115077895382, | |
| "learning_rate": 5.208333333333333e-08, | |
| "loss": 1.3326, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0026164311878597592, | |
| "grad_norm": 0.7177076522063459, | |
| "learning_rate": 2.604166666666667e-07, | |
| "loss": 1.3899, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0052328623757195184, | |
| "grad_norm": 0.7583191114646187, | |
| "learning_rate": 5.208333333333334e-07, | |
| "loss": 1.4091, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.007849293563579277, | |
| "grad_norm": 0.8096572396517622, | |
| "learning_rate": 7.8125e-07, | |
| "loss": 1.3974, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.010465724751439037, | |
| "grad_norm": 0.7135393613888278, | |
| "learning_rate": 1.0416666666666667e-06, | |
| "loss": 1.4201, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.013082155939298797, | |
| "grad_norm": 0.8237790378312874, | |
| "learning_rate": 1.3020833333333335e-06, | |
| "loss": 1.38, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.015698587127158554, | |
| "grad_norm": 0.6714119856024239, | |
| "learning_rate": 1.5625e-06, | |
| "loss": 1.351, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.018315018315018316, | |
| "grad_norm": 0.8239045782794375, | |
| "learning_rate": 1.8229166666666666e-06, | |
| "loss": 1.3952, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.020931449502878074, | |
| "grad_norm": 1.0046248611988764, | |
| "learning_rate": 2.0833333333333334e-06, | |
| "loss": 1.3865, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.023547880690737835, | |
| "grad_norm": 0.9510024082096722, | |
| "learning_rate": 2.3437500000000002e-06, | |
| "loss": 1.3277, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.026164311878597593, | |
| "grad_norm": 0.9091511707564879, | |
| "learning_rate": 2.604166666666667e-06, | |
| "loss": 1.3283, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.02878074306645735, | |
| "grad_norm": 0.6123784246790888, | |
| "learning_rate": 2.8645833333333334e-06, | |
| "loss": 1.2988, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.03139717425431711, | |
| "grad_norm": 0.4784674969150843, | |
| "learning_rate": 3.125e-06, | |
| "loss": 1.2584, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.034013605442176874, | |
| "grad_norm": 0.3377324101778873, | |
| "learning_rate": 3.385416666666667e-06, | |
| "loss": 1.301, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.03663003663003663, | |
| "grad_norm": 0.2500962071399914, | |
| "learning_rate": 3.6458333333333333e-06, | |
| "loss": 1.276, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.03924646781789639, | |
| "grad_norm": 0.24599412075991076, | |
| "learning_rate": 3.90625e-06, | |
| "loss": 1.2516, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.04186289900575615, | |
| "grad_norm": 0.22877062148669897, | |
| "learning_rate": 4.166666666666667e-06, | |
| "loss": 1.285, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.044479330193615906, | |
| "grad_norm": 0.23251824977802066, | |
| "learning_rate": 4.427083333333334e-06, | |
| "loss": 1.2306, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.04709576138147567, | |
| "grad_norm": 0.17838050643581013, | |
| "learning_rate": 4.6875000000000004e-06, | |
| "loss": 1.1997, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.04971219256933543, | |
| "grad_norm": 0.18600609316398758, | |
| "learning_rate": 4.947916666666667e-06, | |
| "loss": 1.2432, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.052328623757195186, | |
| "grad_norm": 0.16796016800278632, | |
| "learning_rate": 5.208333333333334e-06, | |
| "loss": 1.2136, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.054945054945054944, | |
| "grad_norm": 0.15844027288563875, | |
| "learning_rate": 5.468750000000001e-06, | |
| "loss": 1.3021, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.0575614861329147, | |
| "grad_norm": 0.1414394595758024, | |
| "learning_rate": 5.729166666666667e-06, | |
| "loss": 1.207, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.06017791732077447, | |
| "grad_norm": 0.13372102221955964, | |
| "learning_rate": 5.989583333333334e-06, | |
| "loss": 1.1969, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.06279434850863422, | |
| "grad_norm": 0.13133233903537844, | |
| "learning_rate": 6.25e-06, | |
| "loss": 1.2135, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.06541077969649398, | |
| "grad_norm": 0.13504128154908707, | |
| "learning_rate": 6.510416666666667e-06, | |
| "loss": 1.2203, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.06802721088435375, | |
| "grad_norm": 0.11673714164624863, | |
| "learning_rate": 6.770833333333334e-06, | |
| "loss": 1.1957, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.0706436420722135, | |
| "grad_norm": 0.11648588731735841, | |
| "learning_rate": 7.031250000000001e-06, | |
| "loss": 1.213, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.07326007326007326, | |
| "grad_norm": 0.10764121011945749, | |
| "learning_rate": 7.291666666666667e-06, | |
| "loss": 1.1774, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.07587650444793302, | |
| "grad_norm": 0.1207489841636154, | |
| "learning_rate": 7.552083333333334e-06, | |
| "loss": 1.188, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.07849293563579278, | |
| "grad_norm": 0.11189371737439505, | |
| "learning_rate": 7.8125e-06, | |
| "loss": 1.1865, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.08110936682365254, | |
| "grad_norm": 0.10576666590076779, | |
| "learning_rate": 8.072916666666667e-06, | |
| "loss": 1.2143, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.0837257980115123, | |
| "grad_norm": 0.10753980068597714, | |
| "learning_rate": 8.333333333333334e-06, | |
| "loss": 1.1912, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.08634222919937205, | |
| "grad_norm": 0.2087852496477166, | |
| "learning_rate": 8.59375e-06, | |
| "loss": 1.1904, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.08895866038723181, | |
| "grad_norm": 0.11274393602070457, | |
| "learning_rate": 8.854166666666667e-06, | |
| "loss": 1.1934, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.09157509157509157, | |
| "grad_norm": 0.10078115406285185, | |
| "learning_rate": 9.114583333333334e-06, | |
| "loss": 1.1925, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.09419152276295134, | |
| "grad_norm": 0.09778449783264744, | |
| "learning_rate": 9.375000000000001e-06, | |
| "loss": 1.2049, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.0968079539508111, | |
| "grad_norm": 0.09777505019147527, | |
| "learning_rate": 9.635416666666668e-06, | |
| "loss": 1.1996, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.09942438513867086, | |
| "grad_norm": 0.09301261390047985, | |
| "learning_rate": 9.895833333333334e-06, | |
| "loss": 1.201, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.10204081632653061, | |
| "grad_norm": 0.09164005113888346, | |
| "learning_rate": 9.999924849924331e-06, | |
| "loss": 1.2041, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.10465724751439037, | |
| "grad_norm": 0.09513443297611852, | |
| "learning_rate": 9.999465607642677e-06, | |
| "loss": 1.1851, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.10727367870225013, | |
| "grad_norm": 0.0924625015215378, | |
| "learning_rate": 9.998588911421522e-06, | |
| "loss": 1.1497, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.10989010989010989, | |
| "grad_norm": 0.0867096881663944, | |
| "learning_rate": 9.99729483446475e-06, | |
| "loss": 1.1878, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.11250654107796965, | |
| "grad_norm": 0.09113399973145837, | |
| "learning_rate": 9.995583484827415e-06, | |
| "loss": 1.2096, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.1151229722658294, | |
| "grad_norm": 0.09372219223412578, | |
| "learning_rate": 9.993455005406717e-06, | |
| "loss": 1.1434, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.11773940345368916, | |
| "grad_norm": 0.0888706071797912, | |
| "learning_rate": 9.990909573930075e-06, | |
| "loss": 1.1714, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.12035583464154893, | |
| "grad_norm": 0.08688437225632582, | |
| "learning_rate": 9.987947402940285e-06, | |
| "loss": 1.2416, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.12297226582940869, | |
| "grad_norm": 0.08844989176928374, | |
| "learning_rate": 9.984568739777776e-06, | |
| "loss": 1.1494, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.12558869701726844, | |
| "grad_norm": 0.09356745534616749, | |
| "learning_rate": 9.980773866559946e-06, | |
| "loss": 1.181, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.1282051282051282, | |
| "grad_norm": 0.08899586003278553, | |
| "learning_rate": 9.976563100157615e-06, | |
| "loss": 1.1746, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.13082155939298795, | |
| "grad_norm": 0.09647258606085989, | |
| "learning_rate": 9.971936792168569e-06, | |
| "loss": 1.1681, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.13343799058084774, | |
| "grad_norm": 0.09145921664695972, | |
| "learning_rate": 9.966895328888195e-06, | |
| "loss": 1.1759, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.1360544217687075, | |
| "grad_norm": 0.09211002833483616, | |
| "learning_rate": 9.961439131277223e-06, | |
| "loss": 1.179, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.13867085295656725, | |
| "grad_norm": 0.08869761747590328, | |
| "learning_rate": 9.955568654926582e-06, | |
| "loss": 1.1448, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.141287284144427, | |
| "grad_norm": 0.08433870852751438, | |
| "learning_rate": 9.949284390019362e-06, | |
| "loss": 1.1419, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.14390371533228677, | |
| "grad_norm": 0.0862719212805683, | |
| "learning_rate": 9.942586861289874e-06, | |
| "loss": 1.1877, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.14652014652014653, | |
| "grad_norm": 0.08374106195292667, | |
| "learning_rate": 9.935476627979837e-06, | |
| "loss": 1.1565, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.14913657770800628, | |
| "grad_norm": 0.09016812822944884, | |
| "learning_rate": 9.927954283791687e-06, | |
| "loss": 1.157, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.15175300889586604, | |
| "grad_norm": 0.09210986306145622, | |
| "learning_rate": 9.920020456838998e-06, | |
| "loss": 1.198, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.1543694400837258, | |
| "grad_norm": 0.08480839343551977, | |
| "learning_rate": 9.911675809594042e-06, | |
| "loss": 1.1847, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.15698587127158556, | |
| "grad_norm": 0.08344711172445468, | |
| "learning_rate": 9.902921038832456e-06, | |
| "loss": 1.1578, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.15960230245944532, | |
| "grad_norm": 0.09164623997355711, | |
| "learning_rate": 9.893756875575082e-06, | |
| "loss": 1.1462, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.16221873364730507, | |
| "grad_norm": 0.09255247041493149, | |
| "learning_rate": 9.884184085026918e-06, | |
| "loss": 1.1392, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.16483516483516483, | |
| "grad_norm": 0.0897808746322372, | |
| "learning_rate": 9.874203466513215e-06, | |
| "loss": 1.1641, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.1674515960230246, | |
| "grad_norm": 1.0170278609791918, | |
| "learning_rate": 9.863815853412748e-06, | |
| "loss": 1.1577, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.17006802721088435, | |
| "grad_norm": 0.08826036294074761, | |
| "learning_rate": 9.853022113088223e-06, | |
| "loss": 1.2158, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.1726844583987441, | |
| "grad_norm": 0.08802482576182481, | |
| "learning_rate": 9.84182314681385e-06, | |
| "loss": 1.178, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.17530088958660386, | |
| "grad_norm": 0.08428748258121495, | |
| "learning_rate": 9.83021988970009e-06, | |
| "loss": 1.1748, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.17791732077446362, | |
| "grad_norm": 0.0915158879642272, | |
| "learning_rate": 9.818213310615575e-06, | |
| "loss": 1.1485, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.18053375196232338, | |
| "grad_norm": 0.08416073413575607, | |
| "learning_rate": 9.805804412106197e-06, | |
| "loss": 1.1549, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.18315018315018314, | |
| "grad_norm": 0.24967945822809895, | |
| "learning_rate": 9.792994230311419e-06, | |
| "loss": 1.1103, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.1857666143380429, | |
| "grad_norm": 0.08534468582528122, | |
| "learning_rate": 9.779783834877727e-06, | |
| "loss": 1.1502, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.18838304552590268, | |
| "grad_norm": 0.08785714030468032, | |
| "learning_rate": 9.766174328869344e-06, | |
| "loss": 1.1203, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.19099947671376244, | |
| "grad_norm": 0.09104404101954715, | |
| "learning_rate": 9.752166848676101e-06, | |
| "loss": 1.1587, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.1936159079016222, | |
| "grad_norm": 0.08702616441035418, | |
| "learning_rate": 9.737762563918564e-06, | |
| "loss": 1.1517, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.19623233908948196, | |
| "grad_norm": 0.08351621618433638, | |
| "learning_rate": 9.722962677350367e-06, | |
| "loss": 1.1694, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.1988487702773417, | |
| "grad_norm": 0.08613990809273116, | |
| "learning_rate": 9.707768424757778e-06, | |
| "loss": 1.1412, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.20146520146520147, | |
| "grad_norm": 0.08504331047696817, | |
| "learning_rate": 9.692181074856515e-06, | |
| "loss": 1.2068, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.20408163265306123, | |
| "grad_norm": 0.08832625003357134, | |
| "learning_rate": 9.676201929185809e-06, | |
| "loss": 1.1562, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.206698063840921, | |
| "grad_norm": 0.08410128753490248, | |
| "learning_rate": 9.659832321999727e-06, | |
| "loss": 1.1808, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.20931449502878074, | |
| "grad_norm": 0.08437421775574223, | |
| "learning_rate": 9.643073620155755e-06, | |
| "loss": 1.1513, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.2119309262166405, | |
| "grad_norm": 0.08297823254670997, | |
| "learning_rate": 9.625927223000679e-06, | |
| "loss": 1.1992, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.21454735740450026, | |
| "grad_norm": 0.08762545372392672, | |
| "learning_rate": 9.608394562253724e-06, | |
| "loss": 1.1399, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.21716378859236002, | |
| "grad_norm": 0.08663186312112149, | |
| "learning_rate": 9.590477101887016e-06, | |
| "loss": 1.1991, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.21978021978021978, | |
| "grad_norm": 0.08978933643128004, | |
| "learning_rate": 9.572176338003341e-06, | |
| "loss": 1.1494, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.22239665096807953, | |
| "grad_norm": 0.08867801298992739, | |
| "learning_rate": 9.553493798711217e-06, | |
| "loss": 1.1474, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.2250130821559393, | |
| "grad_norm": 0.08900603406641072, | |
| "learning_rate": 9.534431043997298e-06, | |
| "loss": 1.1369, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.22762951334379905, | |
| "grad_norm": 0.09379226586738242, | |
| "learning_rate": 9.514989665596114e-06, | |
| "loss": 1.1704, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.2302459445316588, | |
| "grad_norm": 0.08713543794920169, | |
| "learning_rate": 9.495171286857171e-06, | |
| "loss": 1.1874, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.23286237571951857, | |
| "grad_norm": 0.08824396670528444, | |
| "learning_rate": 9.47497756260939e-06, | |
| "loss": 1.1212, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.23547880690737832, | |
| "grad_norm": 0.08660495234850499, | |
| "learning_rate": 9.454410179022932e-06, | |
| "loss": 1.163, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.23809523809523808, | |
| "grad_norm": 0.08750470885437896, | |
| "learning_rate": 9.433470853468409e-06, | |
| "loss": 1.1602, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.24071166928309787, | |
| "grad_norm": 0.08628399049083942, | |
| "learning_rate": 9.412161334373477e-06, | |
| "loss": 1.1678, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.24332810047095763, | |
| "grad_norm": 0.08571628924285413, | |
| "learning_rate": 9.39048340107685e-06, | |
| "loss": 1.1723, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.24594453165881738, | |
| "grad_norm": 0.08791633295219613, | |
| "learning_rate": 9.36843886367972e-06, | |
| "loss": 1.1566, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.24856096284667714, | |
| "grad_norm": 0.08738481746757637, | |
| "learning_rate": 9.346029562894616e-06, | |
| "loss": 1.1626, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.25117739403453687, | |
| "grad_norm": 0.08497656897045082, | |
| "learning_rate": 9.323257369891702e-06, | |
| "loss": 1.1282, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.25379382522239663, | |
| "grad_norm": 0.09461836398952288, | |
| "learning_rate": 9.300124186142542e-06, | |
| "loss": 1.1765, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.2564102564102564, | |
| "grad_norm": 0.08978125569108607, | |
| "learning_rate": 9.276631943261325e-06, | |
| "loss": 1.1511, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.25902668759811615, | |
| "grad_norm": 0.08888179419029707, | |
| "learning_rate": 9.252782602843565e-06, | |
| "loss": 1.1976, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.2616431187859759, | |
| "grad_norm": 0.08676326656909038, | |
| "learning_rate": 9.228578156302327e-06, | |
| "loss": 1.1431, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.26425954997383566, | |
| "grad_norm": 0.08911630868462432, | |
| "learning_rate": 9.204020624701932e-06, | |
| "loss": 1.1745, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.2668759811616955, | |
| "grad_norm": 0.0896364208788128, | |
| "learning_rate": 9.1791120585892e-06, | |
| "loss": 1.1191, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.26949241234955523, | |
| "grad_norm": 0.08715125789778647, | |
| "learning_rate": 9.153854537822235e-06, | |
| "loss": 1.1597, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.272108843537415, | |
| "grad_norm": 0.08940795227056254, | |
| "learning_rate": 9.12825017139675e-06, | |
| "loss": 1.1768, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.27472527472527475, | |
| "grad_norm": 0.08987300483710815, | |
| "learning_rate": 9.102301097269974e-06, | |
| "loss": 1.1848, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.2773417059131345, | |
| "grad_norm": 0.08688476589575474, | |
| "learning_rate": 9.076009482182132e-06, | |
| "loss": 1.155, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.27995813710099426, | |
| "grad_norm": 0.08866136812688989, | |
| "learning_rate": 9.049377521475514e-06, | |
| "loss": 1.1464, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.282574568288854, | |
| "grad_norm": 0.08688176593226869, | |
| "learning_rate": 9.022407438911177e-06, | |
| "loss": 1.1775, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.2851909994767138, | |
| "grad_norm": 0.08921372549008971, | |
| "learning_rate": 8.99510148648325e-06, | |
| "loss": 1.1447, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.28780743066457354, | |
| "grad_norm": 0.08881883359020548, | |
| "learning_rate": 8.967461944230908e-06, | |
| "loss": 1.1529, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.2904238618524333, | |
| "grad_norm": 0.09167492045293721, | |
| "learning_rate": 8.939491120047974e-06, | |
| "loss": 1.1934, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.29304029304029305, | |
| "grad_norm": 0.09803844949821347, | |
| "learning_rate": 8.911191349490215e-06, | |
| "loss": 1.1604, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.2956567242281528, | |
| "grad_norm": 0.08822343779730063, | |
| "learning_rate": 8.882564995580329e-06, | |
| "loss": 1.1652, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.29827315541601257, | |
| "grad_norm": 0.09087021944109164, | |
| "learning_rate": 8.85361444861063e-06, | |
| "loss": 1.152, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.3008895866038723, | |
| "grad_norm": 0.08966543759541298, | |
| "learning_rate": 8.824342125943461e-06, | |
| "loss": 1.1434, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.3035060177917321, | |
| "grad_norm": 0.08498500583112445, | |
| "learning_rate": 8.79475047180934e-06, | |
| "loss": 1.1821, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.30612244897959184, | |
| "grad_norm": 0.08988190311113226, | |
| "learning_rate": 8.764841957102866e-06, | |
| "loss": 1.1974, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.3087388801674516, | |
| "grad_norm": 0.09197928084619712, | |
| "learning_rate": 8.734619079176416e-06, | |
| "loss": 1.1856, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.31135531135531136, | |
| "grad_norm": 0.08755785801058809, | |
| "learning_rate": 8.704084361631597e-06, | |
| "loss": 1.1638, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.3139717425431711, | |
| "grad_norm": 0.08732771300829235, | |
| "learning_rate": 8.673240354108539e-06, | |
| "loss": 1.1286, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.3165881737310309, | |
| "grad_norm": 0.08526709940215751, | |
| "learning_rate": 8.642089632072992e-06, | |
| "loss": 1.113, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.31920460491889063, | |
| "grad_norm": 0.09343388151136044, | |
| "learning_rate": 8.61063479660128e-06, | |
| "loss": 1.1909, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.3218210361067504, | |
| "grad_norm": 0.09166163273899633, | |
| "learning_rate": 8.578878474163115e-06, | |
| "loss": 1.1998, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.32443746729461015, | |
| "grad_norm": 0.08646409623512233, | |
| "learning_rate": 8.546823316402282e-06, | |
| "loss": 1.1754, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.3270538984824699, | |
| "grad_norm": 0.08819056884980772, | |
| "learning_rate": 8.514471999915229e-06, | |
| "loss": 1.1826, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.32967032967032966, | |
| "grad_norm": 0.08749700044724819, | |
| "learning_rate": 8.48182722602757e-06, | |
| "loss": 1.1722, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.3322867608581894, | |
| "grad_norm": 0.09134268192230549, | |
| "learning_rate": 8.448891720568535e-06, | |
| "loss": 1.1441, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.3349031920460492, | |
| "grad_norm": 0.09018139513674642, | |
| "learning_rate": 8.415668233643346e-06, | |
| "loss": 1.1358, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.33751962323390894, | |
| "grad_norm": 0.08829658429170438, | |
| "learning_rate": 8.382159539403605e-06, | |
| "loss": 1.1945, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.3401360544217687, | |
| "grad_norm": 0.0897490752560212, | |
| "learning_rate": 8.348368435815636e-06, | |
| "loss": 1.1444, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.34275248560962845, | |
| "grad_norm": 0.08957345807921631, | |
| "learning_rate": 8.314297744426865e-06, | |
| "loss": 1.1493, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.3453689167974882, | |
| "grad_norm": 0.08701430974911231, | |
| "learning_rate": 8.279950310130218e-06, | |
| "loss": 1.1431, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.34798534798534797, | |
| "grad_norm": 0.09086431580546972, | |
| "learning_rate": 8.245329000926574e-06, | |
| "loss": 1.154, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.35060177917320773, | |
| "grad_norm": 0.0871637799963955, | |
| "learning_rate": 8.210436707685286e-06, | |
| "loss": 1.17, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.3532182103610675, | |
| "grad_norm": 0.09041417796185347, | |
| "learning_rate": 8.175276343902802e-06, | |
| "loss": 1.1578, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.35583464154892724, | |
| "grad_norm": 0.08625800774031583, | |
| "learning_rate": 8.139850845459378e-06, | |
| "loss": 1.152, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.358451072736787, | |
| "grad_norm": 0.08569835072754965, | |
| "learning_rate": 8.104163170373942e-06, | |
| "loss": 1.1474, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.36106750392464676, | |
| "grad_norm": 0.0862643619125561, | |
| "learning_rate": 8.068216298557088e-06, | |
| "loss": 1.1798, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.3636839351125065, | |
| "grad_norm": 0.08923553182172266, | |
| "learning_rate": 8.032013231562271e-06, | |
| "loss": 1.1774, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.3663003663003663, | |
| "grad_norm": 0.0938085723659071, | |
| "learning_rate": 7.995556992335168e-06, | |
| "loss": 1.1686, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.36891679748822603, | |
| "grad_norm": 0.09116157462150036, | |
| "learning_rate": 7.95885062496126e-06, | |
| "loss": 1.1596, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.3715332286760858, | |
| "grad_norm": 0.087377253429772, | |
| "learning_rate": 7.92189719441166e-06, | |
| "loss": 1.121, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.3741496598639456, | |
| "grad_norm": 0.08786198964388132, | |
| "learning_rate": 7.884699786287188e-06, | |
| "loss": 1.1874, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.37676609105180536, | |
| "grad_norm": 0.0876911684073287, | |
| "learning_rate": 7.847261506560716e-06, | |
| "loss": 1.1725, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.3793825222396651, | |
| "grad_norm": 0.08709531166064968, | |
| "learning_rate": 7.809585481317824e-06, | |
| "loss": 1.151, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.3819989534275249, | |
| "grad_norm": 0.08904419588580584, | |
| "learning_rate": 7.77167485649578e-06, | |
| "loss": 1.1723, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.38461538461538464, | |
| "grad_norm": 0.08714848924978698, | |
| "learning_rate": 7.733532797620849e-06, | |
| "loss": 1.1719, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.3872318158032444, | |
| "grad_norm": 0.0879637315772245, | |
| "learning_rate": 7.695162489543966e-06, | |
| "loss": 1.1916, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.38984824699110415, | |
| "grad_norm": 0.09131210712124205, | |
| "learning_rate": 7.656567136174817e-06, | |
| "loss": 1.1221, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.3924646781789639, | |
| "grad_norm": 0.08589014065766336, | |
| "learning_rate": 7.6177499602143e-06, | |
| "loss": 1.1239, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.39508110936682367, | |
| "grad_norm": 0.09203458025099896, | |
| "learning_rate": 7.578714202885436e-06, | |
| "loss": 1.189, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.3976975405546834, | |
| "grad_norm": 0.09039825298617554, | |
| "learning_rate": 7.53946312366273e-06, | |
| "loss": 1.1516, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.4003139717425432, | |
| "grad_norm": 0.0903818016412424, | |
| "learning_rate": 7.500000000000001e-06, | |
| "loss": 1.1021, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.40293040293040294, | |
| "grad_norm": 0.09041941058273394, | |
| "learning_rate": 7.460328127056718e-06, | |
| "loss": 1.1985, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.4055468341182627, | |
| "grad_norm": 0.09065571364174703, | |
| "learning_rate": 7.420450817422855e-06, | |
| "loss": 1.1658, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.40816326530612246, | |
| "grad_norm": 0.0906781027830754, | |
| "learning_rate": 7.38037140084229e-06, | |
| "loss": 1.1592, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.4107796964939822, | |
| "grad_norm": 0.08872684387079897, | |
| "learning_rate": 7.340093223934775e-06, | |
| "loss": 1.1568, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.413396127681842, | |
| "grad_norm": 0.08708882267716725, | |
| "learning_rate": 7.29961964991649e-06, | |
| "loss": 1.1302, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.41601255886970173, | |
| "grad_norm": 0.09987822856102017, | |
| "learning_rate": 7.2589540583192165e-06, | |
| "loss": 1.2165, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.4186289900575615, | |
| "grad_norm": 0.08691615677709154, | |
| "learning_rate": 7.218099844708152e-06, | |
| "loss": 1.1847, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.42124542124542125, | |
| "grad_norm": 0.0875710124804, | |
| "learning_rate": 7.177060420398376e-06, | |
| "loss": 1.155, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.423861852433281, | |
| "grad_norm": 0.09103596369664581, | |
| "learning_rate": 7.135839212170008e-06, | |
| "loss": 1.1482, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.42647828362114076, | |
| "grad_norm": 0.09142349753820916, | |
| "learning_rate": 7.094439661982072e-06, | |
| "loss": 1.1698, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.4290947148090005, | |
| "grad_norm": 0.09008053931300881, | |
| "learning_rate": 7.0528652266850935e-06, | |
| "loss": 1.159, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.4317111459968603, | |
| "grad_norm": 0.09147508459628673, | |
| "learning_rate": 7.011119377732459e-06, | |
| "loss": 1.1582, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.43432757718472004, | |
| "grad_norm": 0.08802145673026414, | |
| "learning_rate": 6.969205600890539e-06, | |
| "loss": 1.172, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.4369440083725798, | |
| "grad_norm": 0.0883394508080139, | |
| "learning_rate": 6.9271273959476415e-06, | |
| "loss": 1.1679, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.43956043956043955, | |
| "grad_norm": 0.09257358766531258, | |
| "learning_rate": 6.884888276421766e-06, | |
| "loss": 1.1459, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.4421768707482993, | |
| "grad_norm": 0.09097031092390022, | |
| "learning_rate": 6.842491769267241e-06, | |
| "loss": 1.1634, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.44479330193615907, | |
| "grad_norm": 0.09044290737240641, | |
| "learning_rate": 6.79994141458021e-06, | |
| "loss": 1.175, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.4474097331240188, | |
| "grad_norm": 0.08881468117501355, | |
| "learning_rate": 6.757240765303047e-06, | |
| "loss": 1.1345, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.4500261643118786, | |
| "grad_norm": 0.09065311873335255, | |
| "learning_rate": 6.7143933869276755e-06, | |
| "loss": 1.1279, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.45264259549973834, | |
| "grad_norm": 0.09184933289996511, | |
| "learning_rate": 6.671402857197864e-06, | |
| "loss": 1.1384, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.4552590266875981, | |
| "grad_norm": 0.08935816076743913, | |
| "learning_rate": 6.628272765810468e-06, | |
| "loss": 1.1944, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.45787545787545786, | |
| "grad_norm": 0.0865160813282612, | |
| "learning_rate": 6.585006714115709e-06, | |
| "loss": 1.1647, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.4604918890633176, | |
| "grad_norm": 0.0989149367662354, | |
| "learning_rate": 6.541608314816451e-06, | |
| "loss": 1.1822, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.4631083202511774, | |
| "grad_norm": 0.08957489979358704, | |
| "learning_rate": 6.498081191666549e-06, | |
| "loss": 1.099, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.46572475143903713, | |
| "grad_norm": 0.08577118327070016, | |
| "learning_rate": 6.454428979168257e-06, | |
| "loss": 1.1611, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.4683411826268969, | |
| "grad_norm": 0.09320922691441037, | |
| "learning_rate": 6.410655322268758e-06, | |
| "loss": 1.1723, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.47095761381475665, | |
| "grad_norm": 0.09240895994287311, | |
| "learning_rate": 6.3667638760558055e-06, | |
| "loss": 1.1925, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.4735740450026164, | |
| "grad_norm": 0.0923474851653278, | |
| "learning_rate": 6.3227583054525296e-06, | |
| "loss": 1.1375, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.47619047619047616, | |
| "grad_norm": 0.09601612097293628, | |
| "learning_rate": 6.2786422849114074e-06, | |
| "loss": 1.151, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.478806907378336, | |
| "grad_norm": 0.09251678910997652, | |
| "learning_rate": 6.2344194981074616e-06, | |
| "loss": 1.1752, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.48142333856619574, | |
| "grad_norm": 0.09499383199025971, | |
| "learning_rate": 6.190093637630662e-06, | |
| "loss": 1.181, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.4840397697540555, | |
| "grad_norm": 0.09218457504806246, | |
| "learning_rate": 6.145668404677604e-06, | |
| "loss": 1.1275, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.48665620094191525, | |
| "grad_norm": 0.09480903465950775, | |
| "learning_rate": 6.101147508742456e-06, | |
| "loss": 1.1352, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.489272632129775, | |
| "grad_norm": 0.08915545874295981, | |
| "learning_rate": 6.056534667307212e-06, | |
| "loss": 1.1671, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.49188906331763477, | |
| "grad_norm": 0.09205914759599232, | |
| "learning_rate": 6.011833605531295e-06, | |
| "loss": 1.1304, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.4945054945054945, | |
| "grad_norm": 0.09332848310400364, | |
| "learning_rate": 5.967048055940503e-06, | |
| "loss": 1.1408, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.4971219256933543, | |
| "grad_norm": 0.08939769403020309, | |
| "learning_rate": 5.922181758115333e-06, | |
| "loss": 1.191, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.49973835688121404, | |
| "grad_norm": 0.09642367429708014, | |
| "learning_rate": 5.8772384583787455e-06, | |
| "loss": 1.1692, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.5023547880690737, | |
| "grad_norm": 0.08887288118634253, | |
| "learning_rate": 5.832221909483334e-06, | |
| "loss": 1.2212, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.5049712192569336, | |
| "grad_norm": 0.09689598858333648, | |
| "learning_rate": 5.787135870297976e-06, | |
| "loss": 1.1289, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.5075876504447933, | |
| "grad_norm": 0.09107397379124381, | |
| "learning_rate": 5.741984105493967e-06, | |
| "loss": 1.1809, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.5102040816326531, | |
| "grad_norm": 0.09071246683679475, | |
| "learning_rate": 5.696770385230679e-06, | |
| "loss": 1.1393, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.5128205128205128, | |
| "grad_norm": 0.09036453577604019, | |
| "learning_rate": 5.651498484840737e-06, | |
| "loss": 1.1806, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.5154369440083726, | |
| "grad_norm": 0.0893188715090235, | |
| "learning_rate": 5.6061721845148e-06, | |
| "loss": 1.1399, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.5180533751962323, | |
| "grad_norm": 0.09085590942264107, | |
| "learning_rate": 5.560795268985899e-06, | |
| "loss": 1.142, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.5206698063840921, | |
| "grad_norm": 0.09019310702073005, | |
| "learning_rate": 5.515371527213422e-06, | |
| "loss": 1.133, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.5232862375719518, | |
| "grad_norm": 0.09126968115906356, | |
| "learning_rate": 5.469904752066736e-06, | |
| "loss": 1.1553, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.5259026687598116, | |
| "grad_norm": 0.09464747718065981, | |
| "learning_rate": 5.424398740008481e-06, | |
| "loss": 1.1718, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.5285190999476713, | |
| "grad_norm": 0.09436242020454211, | |
| "learning_rate": 5.378857290777566e-06, | |
| "loss": 1.155, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.5311355311355311, | |
| "grad_norm": 0.10001855825215443, | |
| "learning_rate": 5.333284207071901e-06, | |
| "loss": 1.1544, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.533751962323391, | |
| "grad_norm": 0.09354368422313043, | |
| "learning_rate": 5.287683294230855e-06, | |
| "loss": 1.1898, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.5363683935112507, | |
| "grad_norm": 0.09193771963055702, | |
| "learning_rate": 5.242058359917531e-06, | |
| "loss": 1.142, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.5389848246991105, | |
| "grad_norm": 0.09421770405872586, | |
| "learning_rate": 5.196413213800812e-06, | |
| "loss": 1.1405, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.5416012558869702, | |
| "grad_norm": 0.09278287279169352, | |
| "learning_rate": 5.150751667237266e-06, | |
| "loss": 1.1413, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.54421768707483, | |
| "grad_norm": 0.09200594925059793, | |
| "learning_rate": 5.1050775329528865e-06, | |
| "loss": 1.1763, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.5468341182626897, | |
| "grad_norm": 0.09202351377274724, | |
| "learning_rate": 5.059394624724749e-06, | |
| "loss": 1.1813, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.5494505494505495, | |
| "grad_norm": 0.09295646295778348, | |
| "learning_rate": 5.0137067570625345e-06, | |
| "loss": 1.147, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.5520669806384092, | |
| "grad_norm": 0.09596103826973694, | |
| "learning_rate": 4.968017744890052e-06, | |
| "loss": 1.1715, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.554683411826269, | |
| "grad_norm": 0.09280818950416957, | |
| "learning_rate": 4.922331403226667e-06, | |
| "loss": 1.1407, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.5572998430141287, | |
| "grad_norm": 0.08975080950200469, | |
| "learning_rate": 4.876651546868759e-06, | |
| "loss": 1.1852, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.5599162742019885, | |
| "grad_norm": 0.09265399542476259, | |
| "learning_rate": 4.830981990071193e-06, | |
| "loss": 1.1443, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.5625327053898482, | |
| "grad_norm": 0.09049838209270984, | |
| "learning_rate": 4.785326546228818e-06, | |
| "loss": 1.1858, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.565149136577708, | |
| "grad_norm": 0.08900263422386634, | |
| "learning_rate": 4.739689027558052e-06, | |
| "loss": 1.1406, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.5677655677655677, | |
| "grad_norm": 0.08902029592212933, | |
| "learning_rate": 4.694073244778571e-06, | |
| "loss": 1.179, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.5703819989534276, | |
| "grad_norm": 0.09275284871138449, | |
| "learning_rate": 4.648483006795115e-06, | |
| "loss": 1.1595, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.5729984301412873, | |
| "grad_norm": 0.09439177740704392, | |
| "learning_rate": 4.602922120379432e-06, | |
| "loss": 1.1233, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.5756148613291471, | |
| "grad_norm": 0.09274318315418675, | |
| "learning_rate": 4.557394389852427e-06, | |
| "loss": 1.1898, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.5782312925170068, | |
| "grad_norm": 0.09405997276406405, | |
| "learning_rate": 4.5119036167664966e-06, | |
| "loss": 1.1649, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.5808477237048666, | |
| "grad_norm": 0.09385980984186612, | |
| "learning_rate": 4.466453599588103e-06, | |
| "loss": 1.191, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.5834641548927263, | |
| "grad_norm": 0.0898832303576217, | |
| "learning_rate": 4.421048133380601e-06, | |
| "loss": 1.1165, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.5860805860805861, | |
| "grad_norm": 0.09360216710271695, | |
| "learning_rate": 4.375691009487351e-06, | |
| "loss": 1.178, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.5886970172684458, | |
| "grad_norm": 0.09683340115595745, | |
| "learning_rate": 4.330386015215145e-06, | |
| "loss": 1.2107, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.5913134484563056, | |
| "grad_norm": 0.09145018465422285, | |
| "learning_rate": 4.285136933517971e-06, | |
| "loss": 1.1315, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.5939298796441653, | |
| "grad_norm": 0.0896731050577423, | |
| "learning_rate": 4.239947542681125e-06, | |
| "loss": 1.1074, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.5965463108320251, | |
| "grad_norm": 0.09660990701138071, | |
| "learning_rate": 4.194821616005738e-06, | |
| "loss": 1.1084, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.5991627420198848, | |
| "grad_norm": 0.09209943029817762, | |
| "learning_rate": 4.1497629214937e-06, | |
| "loss": 1.1389, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.6017791732077447, | |
| "grad_norm": 0.1006325243229284, | |
| "learning_rate": 4.104775221533039e-06, | |
| "loss": 1.1816, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.6043956043956044, | |
| "grad_norm": 0.09349785393398444, | |
| "learning_rate": 4.059862272583755e-06, | |
| "loss": 1.1799, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.6070120355834642, | |
| "grad_norm": 0.0911894072302353, | |
| "learning_rate": 4.015027824864158e-06, | |
| "loss": 1.1787, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.6096284667713239, | |
| "grad_norm": 0.09090197142913302, | |
| "learning_rate": 3.97027562203773e-06, | |
| "loss": 1.1497, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.6122448979591837, | |
| "grad_norm": 0.09220893282730831, | |
| "learning_rate": 3.92560940090053e-06, | |
| "loss": 1.1556, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.6148613291470434, | |
| "grad_norm": 0.09513786147473494, | |
| "learning_rate": 3.881032891069169e-06, | |
| "loss": 1.1619, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.6174777603349032, | |
| "grad_norm": 0.09592949877983106, | |
| "learning_rate": 3.836549814669389e-06, | |
| "loss": 1.1717, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.6200941915227629, | |
| "grad_norm": 0.09162720160250148, | |
| "learning_rate": 3.7921638860252674e-06, | |
| "loss": 1.1496, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.6227106227106227, | |
| "grad_norm": 0.0929693110938376, | |
| "learning_rate": 3.747878811349075e-06, | |
| "loss": 1.179, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.6253270538984824, | |
| "grad_norm": 0.1717629210930998, | |
| "learning_rate": 3.703698288431801e-06, | |
| "loss": 1.1444, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.6279434850863422, | |
| "grad_norm": 0.09078602967297458, | |
| "learning_rate": 3.659626006334395e-06, | |
| "loss": 1.1602, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.6305599162742019, | |
| "grad_norm": 0.10359394840301954, | |
| "learning_rate": 3.615665645079728e-06, | |
| "loss": 1.1518, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.6331763474620618, | |
| "grad_norm": 0.09092011886226412, | |
| "learning_rate": 3.5718208753453166e-06, | |
| "loss": 1.1963, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.6357927786499215, | |
| "grad_norm": 0.09272641870883862, | |
| "learning_rate": 3.5280953581568155e-06, | |
| "loss": 1.16, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.6384092098377813, | |
| "grad_norm": 0.08834856572554059, | |
| "learning_rate": 3.484492744582325e-06, | |
| "loss": 1.1579, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.6410256410256411, | |
| "grad_norm": 0.09127555751260019, | |
| "learning_rate": 3.441016675427532e-06, | |
| "loss": 1.1536, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.6436420722135008, | |
| "grad_norm": 0.20406095763855414, | |
| "learning_rate": 3.397670780931699e-06, | |
| "loss": 1.1381, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.6462585034013606, | |
| "grad_norm": 0.0895588108461287, | |
| "learning_rate": 3.354458680464543e-06, | |
| "loss": 1.1363, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.6488749345892203, | |
| "grad_norm": 0.09164524609745374, | |
| "learning_rate": 3.311383982224017e-06, | |
| "loss": 1.1459, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.6514913657770801, | |
| "grad_norm": 0.09366284043931154, | |
| "learning_rate": 3.268450282935026e-06, | |
| "loss": 1.1663, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.6541077969649398, | |
| "grad_norm": 0.09332801200983233, | |
| "learning_rate": 3.2256611675491096e-06, | |
| "loss": 1.1582, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.6567242281527996, | |
| "grad_norm": 0.09420123178793989, | |
| "learning_rate": 3.183020208945086e-06, | |
| "loss": 1.1485, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.6593406593406593, | |
| "grad_norm": 0.09544963446821658, | |
| "learning_rate": 3.1405309676307283e-06, | |
| "loss": 1.1876, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.6619570905285191, | |
| "grad_norm": 0.09735866573127151, | |
| "learning_rate": 3.0981969914454555e-06, | |
| "loss": 1.1823, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.6645735217163788, | |
| "grad_norm": 0.09258003535852607, | |
| "learning_rate": 3.056021815264102e-06, | |
| "loss": 1.1615, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.6671899529042387, | |
| "grad_norm": 0.08811860630150058, | |
| "learning_rate": 3.0140089607017386e-06, | |
| "loss": 1.1935, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.6698063840920984, | |
| "grad_norm": 0.0926515991851408, | |
| "learning_rate": 2.972161935819632e-06, | |
| "loss": 1.1368, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.6724228152799582, | |
| "grad_norm": 0.09298206722745225, | |
| "learning_rate": 2.930484234832315e-06, | |
| "loss": 1.1613, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.6750392464678179, | |
| "grad_norm": 0.08718106601664169, | |
| "learning_rate": 2.8889793378158284e-06, | |
| "loss": 1.1373, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.6776556776556777, | |
| "grad_norm": 0.09364478524137042, | |
| "learning_rate": 2.8476507104171273e-06, | |
| "loss": 1.1599, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.6802721088435374, | |
| "grad_norm": 0.08648218131861361, | |
| "learning_rate": 2.806501803564708e-06, | |
| "loss": 1.1718, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.6828885400313972, | |
| "grad_norm": 0.09271938103028711, | |
| "learning_rate": 2.765536053180447e-06, | |
| "loss": 1.1501, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.6855049712192569, | |
| "grad_norm": 0.09416500959969552, | |
| "learning_rate": 2.724756879892717e-06, | |
| "loss": 1.1582, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.6881214024071167, | |
| "grad_norm": 0.09647225239602465, | |
| "learning_rate": 2.6841676887507505e-06, | |
| "loss": 1.1443, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.6907378335949764, | |
| "grad_norm": 0.09415539941506598, | |
| "learning_rate": 2.643771868940327e-06, | |
| "loss": 1.1565, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.6933542647828362, | |
| "grad_norm": 0.09124599697629637, | |
| "learning_rate": 2.603572793500775e-06, | |
| "loss": 1.1421, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.6959706959706959, | |
| "grad_norm": 0.09126639454111003, | |
| "learning_rate": 2.5635738190433252e-06, | |
| "loss": 1.1413, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.6985871271585558, | |
| "grad_norm": 0.09952564680953281, | |
| "learning_rate": 2.523778285470835e-06, | |
| "loss": 1.177, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.7012035583464155, | |
| "grad_norm": 0.09114528754560879, | |
| "learning_rate": 2.4841895156989047e-06, | |
| "loss": 1.1661, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.7038199895342753, | |
| "grad_norm": 0.09202095365080962, | |
| "learning_rate": 2.444810815378416e-06, | |
| "loss": 1.1186, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.706436420722135, | |
| "grad_norm": 0.09293569476066414, | |
| "learning_rate": 2.4056454726195166e-06, | |
| "loss": 1.1673, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.7090528519099948, | |
| "grad_norm": 0.0947692422323796, | |
| "learning_rate": 2.366696757717054e-06, | |
| "loss": 1.1345, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.7116692830978545, | |
| "grad_norm": 0.0928402764545078, | |
| "learning_rate": 2.327967922877515e-06, | |
| "loss": 1.2186, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 0.08953711007777569, | |
| "learning_rate": 2.28946220194746e-06, | |
| "loss": 1.159, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.716902145473574, | |
| "grad_norm": 0.0891685929865638, | |
| "learning_rate": 2.2511828101435105e-06, | |
| "loss": 1.1711, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.7195185766614338, | |
| "grad_norm": 0.08872499060461402, | |
| "learning_rate": 2.213132943783864e-06, | |
| "loss": 1.1262, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.7221350078492935, | |
| "grad_norm": 0.09018865422225913, | |
| "learning_rate": 2.1753157800214107e-06, | |
| "loss": 1.1448, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.7247514390371533, | |
| "grad_norm": 0.09255990516976745, | |
| "learning_rate": 2.137734476578443e-06, | |
| "loss": 1.1951, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.727367870225013, | |
| "grad_norm": 0.08814414164100583, | |
| "learning_rate": 2.1003921714829823e-06, | |
| "loss": 1.1657, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.7299843014128728, | |
| "grad_norm": 0.08989614430153806, | |
| "learning_rate": 2.063291982806759e-06, | |
| "loss": 1.1201, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.7326007326007326, | |
| "grad_norm": 0.09752021041617213, | |
| "learning_rate": 2.0264370084048498e-06, | |
| "loss": 1.1853, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.7352171637885924, | |
| "grad_norm": 0.0951117119155389, | |
| "learning_rate": 1.9898303256570093e-06, | |
| "loss": 1.1325, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.7378335949764521, | |
| "grad_norm": 0.09421101398847186, | |
| "learning_rate": 1.953474991210717e-06, | |
| "loss": 1.152, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.7404500261643119, | |
| "grad_norm": 0.09127814615367837, | |
| "learning_rate": 1.917374040725935e-06, | |
| "loss": 1.1695, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.7430664573521716, | |
| "grad_norm": 0.0929739978066435, | |
| "learning_rate": 1.8815304886216385e-06, | |
| "loss": 1.1487, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.7456828885400314, | |
| "grad_norm": 0.09984071894833883, | |
| "learning_rate": 1.8459473278241125e-06, | |
| "loss": 1.1208, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.7482993197278912, | |
| "grad_norm": 0.08978226240423473, | |
| "learning_rate": 1.8106275295170462e-06, | |
| "loss": 1.151, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.7509157509157509, | |
| "grad_norm": 0.08592008855291842, | |
| "learning_rate": 1.7755740428934333e-06, | |
| "loss": 1.1305, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.7535321821036107, | |
| "grad_norm": 0.08991613497803638, | |
| "learning_rate": 1.7407897949093184e-06, | |
| "loss": 1.1185, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.7561486132914704, | |
| "grad_norm": 0.0881827045855804, | |
| "learning_rate": 1.7062776900393979e-06, | |
| "loss": 1.1882, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.7587650444793302, | |
| "grad_norm": 0.09823529922897513, | |
| "learning_rate": 1.6720406100344977e-06, | |
| "loss": 1.159, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.7613814756671899, | |
| "grad_norm": 0.0918690597533081, | |
| "learning_rate": 1.6380814136809442e-06, | |
| "loss": 1.1465, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.7639979068550498, | |
| "grad_norm": 0.09514583884699968, | |
| "learning_rate": 1.6044029365618612e-06, | |
| "loss": 1.1075, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.7666143380429095, | |
| "grad_norm": 0.08938834682344118, | |
| "learning_rate": 1.571007990820394e-06, | |
| "loss": 1.1776, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.7692307692307693, | |
| "grad_norm": 0.09025983792300737, | |
| "learning_rate": 1.5378993649249053e-06, | |
| "loss": 1.1439, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.771847200418629, | |
| "grad_norm": 0.12420625501789342, | |
| "learning_rate": 1.5050798234361269e-06, | |
| "loss": 1.1878, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.7744636316064888, | |
| "grad_norm": 0.0916852143560027, | |
| "learning_rate": 1.4725521067763298e-06, | |
| "loss": 1.1365, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.7770800627943485, | |
| "grad_norm": 0.08990582712487037, | |
| "learning_rate": 1.4403189310004917e-06, | |
| "loss": 1.1606, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.7796964939822083, | |
| "grad_norm": 0.0898250912235613, | |
| "learning_rate": 1.4083829875695172e-06, | |
| "loss": 1.1369, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.782312925170068, | |
| "grad_norm": 0.08643871593208942, | |
| "learning_rate": 1.376746943125491e-06, | |
| "loss": 1.1521, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.7849293563579278, | |
| "grad_norm": 0.09699351922882188, | |
| "learning_rate": 1.34541343926902e-06, | |
| "loss": 1.1538, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.7875457875457875, | |
| "grad_norm": 0.09362779687742182, | |
| "learning_rate": 1.3143850923386586e-06, | |
| "loss": 1.1545, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.7901622187336473, | |
| "grad_norm": 0.0960893233426078, | |
| "learning_rate": 1.2836644931924469e-06, | |
| "loss": 1.1819, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.792778649921507, | |
| "grad_norm": 0.11158068373603516, | |
| "learning_rate": 1.2532542069915722e-06, | |
| "loss": 1.1416, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.7953950811093669, | |
| "grad_norm": 0.0891532569068602, | |
| "learning_rate": 1.2231567729861809e-06, | |
| "loss": 1.1318, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.7980115122972266, | |
| "grad_norm": 0.08941090773600295, | |
| "learning_rate": 1.1933747043033505e-06, | |
| "loss": 1.155, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.8006279434850864, | |
| "grad_norm": 0.09417150844445865, | |
| "learning_rate": 1.1639104877372475e-06, | |
| "loss": 1.1656, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.8032443746729461, | |
| "grad_norm": 0.09732531849636676, | |
| "learning_rate": 1.134766583541475e-06, | |
| "loss": 1.1511, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.8058608058608059, | |
| "grad_norm": 0.09041353322682366, | |
| "learning_rate": 1.1059454252236457e-06, | |
| "loss": 1.1633, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.8084772370486656, | |
| "grad_norm": 0.08924111402124033, | |
| "learning_rate": 1.0774494193421842e-06, | |
| "loss": 1.1638, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.8110936682365254, | |
| "grad_norm": 0.08906347087320848, | |
| "learning_rate": 1.0492809453053836e-06, | |
| "loss": 1.1428, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.8137100994243851, | |
| "grad_norm": 0.09931949226958557, | |
| "learning_rate": 1.0214423551727188e-06, | |
| "loss": 1.1551, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.8163265306122449, | |
| "grad_norm": 0.09231523884912021, | |
| "learning_rate": 9.939359734584552e-07, | |
| "loss": 1.1274, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.8189429618001046, | |
| "grad_norm": 0.09024048942683224, | |
| "learning_rate": 9.667640969375465e-07, | |
| "loss": 1.1292, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.8215593929879644, | |
| "grad_norm": 0.09114167737765824, | |
| "learning_rate": 9.399289944538664e-07, | |
| "loss": 1.1618, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.8241758241758241, | |
| "grad_norm": 0.09619116235928656, | |
| "learning_rate": 9.134329067307485e-07, | |
| "loss": 1.1815, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.826792255363684, | |
| "grad_norm": 0.09366351832282298, | |
| "learning_rate": 8.872780461838931e-07, | |
| "loss": 1.1743, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.8294086865515437, | |
| "grad_norm": 0.09032902184148513, | |
| "learning_rate": 8.614665967366276e-07, | |
| "loss": 1.1892, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.8320251177394035, | |
| "grad_norm": 0.09293264712898819, | |
| "learning_rate": 8.360007136375553e-07, | |
| "loss": 1.165, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.8346415489272632, | |
| "grad_norm": 0.09151115721297814, | |
| "learning_rate": 8.108825232805856e-07, | |
| "loss": 1.164, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.837257980115123, | |
| "grad_norm": 0.09489044639295917, | |
| "learning_rate": 7.861141230273839e-07, | |
| "loss": 1.1353, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.8398744113029827, | |
| "grad_norm": 0.09562784828651455, | |
| "learning_rate": 7.61697581032243e-07, | |
| "loss": 1.1759, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.8424908424908425, | |
| "grad_norm": 0.08998604269061755, | |
| "learning_rate": 7.376349360693952e-07, | |
| "loss": 1.1508, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.8451072736787022, | |
| "grad_norm": 0.09627443444242165, | |
| "learning_rate": 7.139281973627693e-07, | |
| "loss": 1.201, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.847723704866562, | |
| "grad_norm": 0.08931968675335339, | |
| "learning_rate": 6.905793444182257e-07, | |
| "loss": 1.1128, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.8503401360544217, | |
| "grad_norm": 0.08954411277957602, | |
| "learning_rate": 6.675903268582623e-07, | |
| "loss": 1.1492, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.8529565672422815, | |
| "grad_norm": 0.093953286316433, | |
| "learning_rate": 6.449630642592336e-07, | |
| "loss": 1.1368, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.8555729984301413, | |
| "grad_norm": 0.09373965859270121, | |
| "learning_rate": 6.22699445991054e-07, | |
| "loss": 1.198, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.858189429618001, | |
| "grad_norm": 0.09061630590131814, | |
| "learning_rate": 6.008013310594418e-07, | |
| "loss": 1.132, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.8608058608058609, | |
| "grad_norm": 0.09611482874800488, | |
| "learning_rate": 5.7927054795069e-07, | |
| "loss": 1.1624, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.8634222919937206, | |
| "grad_norm": 0.09051341176877918, | |
| "learning_rate": 5.581088944789953e-07, | |
| "loss": 1.1563, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.8660387231815804, | |
| "grad_norm": 0.08699408313720076, | |
| "learning_rate": 5.373181376363312e-07, | |
| "loss": 1.1935, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.8686551543694401, | |
| "grad_norm": 0.09489646346012261, | |
| "learning_rate": 5.169000134449115e-07, | |
| "loss": 1.185, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.8712715855572999, | |
| "grad_norm": 0.09257338125449141, | |
| "learning_rate": 4.968562268122285e-07, | |
| "loss": 1.1777, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.8738880167451596, | |
| "grad_norm": 0.09184746461118028, | |
| "learning_rate": 4.771884513886998e-07, | |
| "loss": 1.1872, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.8765044479330194, | |
| "grad_norm": 0.09013545936327065, | |
| "learning_rate": 4.578983294279138e-07, | |
| "loss": 1.1465, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.8791208791208791, | |
| "grad_norm": 0.09456190273676769, | |
| "learning_rate": 4.389874716495013e-07, | |
| "loss": 1.1729, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.8817373103087389, | |
| "grad_norm": 0.09236101919389716, | |
| "learning_rate": 4.204574571046438e-07, | |
| "loss": 1.1816, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.8843537414965986, | |
| "grad_norm": 0.09172117579338826, | |
| "learning_rate": 4.0230983304422543e-07, | |
| "loss": 1.1846, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.8869701726844584, | |
| "grad_norm": 0.09136763206816247, | |
| "learning_rate": 3.8454611478963235e-07, | |
| "loss": 1.1894, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.8895866038723181, | |
| "grad_norm": 0.09361682354937588, | |
| "learning_rate": 3.671677856062261e-07, | |
| "loss": 1.1359, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.892203035060178, | |
| "grad_norm": 0.09300698311619561, | |
| "learning_rate": 3.501762965794919e-07, | |
| "loss": 1.188, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.8948194662480377, | |
| "grad_norm": 0.08947489152597186, | |
| "learning_rate": 3.335730664938758e-07, | |
| "loss": 1.1547, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.8974358974358975, | |
| "grad_norm": 0.0910579074415647, | |
| "learning_rate": 3.1735948171431e-07, | |
| "loss": 1.1491, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.9000523286237572, | |
| "grad_norm": 0.0928618229612821, | |
| "learning_rate": 3.015368960704584e-07, | |
| "loss": 1.192, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.902668759811617, | |
| "grad_norm": 0.09756621044157429, | |
| "learning_rate": 2.8610663074366773e-07, | |
| "loss": 1.1696, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.9052851909994767, | |
| "grad_norm": 0.0921375489250745, | |
| "learning_rate": 2.7106997415665527e-07, | |
| "loss": 1.1309, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.9079016221873365, | |
| "grad_norm": 0.09207854242951917, | |
| "learning_rate": 2.564281818659159e-07, | |
| "loss": 1.1251, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 0.9105180533751962, | |
| "grad_norm": 0.09352988984475613, | |
| "learning_rate": 2.4218247645689306e-07, | |
| "loss": 1.1971, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.913134484563056, | |
| "grad_norm": 0.08856172913123674, | |
| "learning_rate": 2.2833404744188824e-07, | |
| "loss": 1.1427, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 0.9157509157509157, | |
| "grad_norm": 0.09697929384761596, | |
| "learning_rate": 2.1488405116074028e-07, | |
| "loss": 1.1201, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.9183673469387755, | |
| "grad_norm": 0.0960470883117448, | |
| "learning_rate": 2.0183361068426778e-07, | |
| "loss": 1.1678, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.9209837781266352, | |
| "grad_norm": 0.09009468323982603, | |
| "learning_rate": 1.8918381572049393e-07, | |
| "loss": 1.15, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.923600209314495, | |
| "grad_norm": 0.09038500937307103, | |
| "learning_rate": 1.7693572252365841e-07, | |
| "loss": 1.1524, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 0.9262166405023547, | |
| "grad_norm": 0.0935616350765029, | |
| "learning_rate": 1.650903538060189e-07, | |
| "loss": 1.1654, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.9288330716902146, | |
| "grad_norm": 0.09353048793106586, | |
| "learning_rate": 1.536486986524538e-07, | |
| "loss": 1.1752, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.9314495028780743, | |
| "grad_norm": 0.09494299784964721, | |
| "learning_rate": 1.426117124378762e-07, | |
| "loss": 1.1574, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.9340659340659341, | |
| "grad_norm": 0.09202594407561465, | |
| "learning_rate": 1.3198031674745814e-07, | |
| "loss": 1.1091, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.9366823652537938, | |
| "grad_norm": 0.09485952901545334, | |
| "learning_rate": 1.2175539929968117e-07, | |
| "loss": 1.1455, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.9392987964416536, | |
| "grad_norm": 0.0898806651171433, | |
| "learning_rate": 1.1193781387220936e-07, | |
| "loss": 1.1468, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 0.9419152276295133, | |
| "grad_norm": 0.09584452763045823, | |
| "learning_rate": 1.0252838023059985e-07, | |
| "loss": 1.1922, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.9445316588173731, | |
| "grad_norm": 0.09231922052171501, | |
| "learning_rate": 9.352788405985469e-08, | |
| "loss": 1.1465, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 0.9471480900052328, | |
| "grad_norm": 0.09015656002427216, | |
| "learning_rate": 8.493707689881448e-08, | |
| "loss": 1.149, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.9497645211930926, | |
| "grad_norm": 0.08904186680424883, | |
| "learning_rate": 7.675667607740356e-08, | |
| "loss": 1.1633, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.9523809523809523, | |
| "grad_norm": 0.09460621552754211, | |
| "learning_rate": 6.898736465673739e-08, | |
| "loss": 1.1649, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.9549973835688121, | |
| "grad_norm": 0.09701362107852395, | |
| "learning_rate": 6.162979137208314e-08, | |
| "loss": 1.1721, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.957613814756672, | |
| "grad_norm": 0.09119591337917335, | |
| "learning_rate": 5.468457057869358e-08, | |
| "loss": 1.1421, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.9602302459445317, | |
| "grad_norm": 0.09386063000292352, | |
| "learning_rate": 4.815228220050538e-08, | |
| "loss": 1.1494, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 0.9628466771323915, | |
| "grad_norm": 0.09314853572272408, | |
| "learning_rate": 4.2033471681718895e-08, | |
| "loss": 1.1461, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.9654631083202512, | |
| "grad_norm": 0.09482802830237459, | |
| "learning_rate": 3.632864994125129e-08, | |
| "loss": 1.1258, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.968079539508111, | |
| "grad_norm": 0.09214541317430823, | |
| "learning_rate": 3.103829333007624e-08, | |
| "loss": 1.1607, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.9706959706959707, | |
| "grad_norm": 0.09498845107605784, | |
| "learning_rate": 2.616284359144794e-08, | |
| "loss": 1.1273, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 0.9733124018838305, | |
| "grad_norm": 0.09632397271400663, | |
| "learning_rate": 2.1702707824017287e-08, | |
| "loss": 1.1715, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.9759288330716902, | |
| "grad_norm": 0.09265306344478029, | |
| "learning_rate": 1.7658258447836306e-08, | |
| "loss": 1.177, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 0.97854526425955, | |
| "grad_norm": 0.09534022115904282, | |
| "learning_rate": 1.4029833173264673e-08, | |
| "loss": 1.1683, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.9811616954474097, | |
| "grad_norm": 0.09336462707005236, | |
| "learning_rate": 1.0817734972768946e-08, | |
| "loss": 1.1523, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.9837781266352695, | |
| "grad_norm": 0.09623214092200802, | |
| "learning_rate": 8.022232055623913e-09, | |
| "loss": 1.1352, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.9863945578231292, | |
| "grad_norm": 0.09408768079361217, | |
| "learning_rate": 5.643557845518843e-09, | |
| "loss": 1.1799, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 0.989010989010989, | |
| "grad_norm": 0.09123794699673446, | |
| "learning_rate": 3.6819109610658486e-09, | |
| "loss": 1.1595, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.9916274201988488, | |
| "grad_norm": 0.09345060971759653, | |
| "learning_rate": 2.137455199215377e-09, | |
| "loss": 1.1038, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 0.9942438513867086, | |
| "grad_norm": 0.09109158309879825, | |
| "learning_rate": 1.0103195215788175e-09, | |
| "loss": 1.1526, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.9968602825745683, | |
| "grad_norm": 0.08848167389727797, | |
| "learning_rate": 3.005980436604494e-10, | |
| "loss": 1.1857, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.9994767137624281, | |
| "grad_norm": 0.09385894732886857, | |
| "learning_rate": 8.350027000392224e-12, | |
| "loss": 1.1728, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_runtime": 3.7017, | |
| "eval_samples_per_second": 2.701, | |
| "eval_steps_per_second": 0.81, | |
| "step": 1911 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 1911, | |
| "total_flos": 2.014665441804288e+16, | |
| "train_loss": 1.17021960310135, | |
| "train_runtime": 9201.8355, | |
| "train_samples_per_second": 3.321, | |
| "train_steps_per_second": 0.208 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 1911, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2.014665441804288e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |