| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0, | |
| "eval_steps": 500, | |
| "global_step": 4325, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.00023121387283236994, | |
| "grad_norm": 0.6965160472814916, | |
| "learning_rate": 4.6189376443418015e-07, | |
| "loss": 1.086, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.0011560693641618498, | |
| "grad_norm": 0.6835787858040867, | |
| "learning_rate": 2.309468822170901e-06, | |
| "loss": 1.128, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0023121387283236996, | |
| "grad_norm": 0.6184758767742636, | |
| "learning_rate": 4.618937644341802e-06, | |
| "loss": 1.1188, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.003468208092485549, | |
| "grad_norm": 0.6078620387437725, | |
| "learning_rate": 6.928406466512702e-06, | |
| "loss": 1.1243, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.004624277456647399, | |
| "grad_norm": 0.5222130979899404, | |
| "learning_rate": 9.237875288683604e-06, | |
| "loss": 1.1013, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.005780346820809248, | |
| "grad_norm": 0.4982345960164489, | |
| "learning_rate": 1.1547344110854504e-05, | |
| "loss": 1.0928, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.006936416184971098, | |
| "grad_norm": 0.40171891380185876, | |
| "learning_rate": 1.3856812933025404e-05, | |
| "loss": 1.0736, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.008092485549132947, | |
| "grad_norm": 0.41106693801534305, | |
| "learning_rate": 1.6166281755196306e-05, | |
| "loss": 1.0672, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.009248554913294798, | |
| "grad_norm": 0.4109497681037648, | |
| "learning_rate": 1.8475750577367208e-05, | |
| "loss": 1.0258, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.010404624277456647, | |
| "grad_norm": 0.379459396453402, | |
| "learning_rate": 2.0785219399538107e-05, | |
| "loss": 1.0102, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.011560693641618497, | |
| "grad_norm": 0.3721675211937845, | |
| "learning_rate": 2.309468822170901e-05, | |
| "loss": 1.0234, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.012716763005780347, | |
| "grad_norm": 0.373222147309766, | |
| "learning_rate": 2.540415704387991e-05, | |
| "loss": 0.992, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.013872832369942197, | |
| "grad_norm": 0.3702598509393352, | |
| "learning_rate": 2.771362586605081e-05, | |
| "loss": 1.0326, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.015028901734104046, | |
| "grad_norm": 0.35490433661120324, | |
| "learning_rate": 3.0023094688221707e-05, | |
| "loss": 1.0246, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.016184971098265895, | |
| "grad_norm": 0.34072939280119285, | |
| "learning_rate": 3.233256351039261e-05, | |
| "loss": 1.0115, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.017341040462427744, | |
| "grad_norm": 0.3831809542290894, | |
| "learning_rate": 3.464203233256351e-05, | |
| "loss": 1.0331, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.018497109826589597, | |
| "grad_norm": 0.3623596879709559, | |
| "learning_rate": 3.6951501154734416e-05, | |
| "loss": 1.0022, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.019653179190751446, | |
| "grad_norm": 0.3931577119185935, | |
| "learning_rate": 3.9260969976905315e-05, | |
| "loss": 0.9997, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.020809248554913295, | |
| "grad_norm": 0.4109841798218973, | |
| "learning_rate": 4.1570438799076213e-05, | |
| "loss": 0.9914, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.021965317919075144, | |
| "grad_norm": 0.40213543229926174, | |
| "learning_rate": 4.387990762124711e-05, | |
| "loss": 0.9643, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.023121387283236993, | |
| "grad_norm": 0.3657003831875091, | |
| "learning_rate": 4.618937644341802e-05, | |
| "loss": 0.9957, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.024277456647398842, | |
| "grad_norm": 0.38609844859658143, | |
| "learning_rate": 4.8498845265588916e-05, | |
| "loss": 1.0111, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.025433526011560695, | |
| "grad_norm": 0.3583592530014769, | |
| "learning_rate": 5.080831408775982e-05, | |
| "loss": 0.9963, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.026589595375722544, | |
| "grad_norm": 0.3554977804638662, | |
| "learning_rate": 5.311778290993071e-05, | |
| "loss": 0.9817, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.027745664739884393, | |
| "grad_norm": 0.3840657570276599, | |
| "learning_rate": 5.542725173210162e-05, | |
| "loss": 1.0232, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.028901734104046242, | |
| "grad_norm": 0.3774759292695808, | |
| "learning_rate": 5.7736720554272516e-05, | |
| "loss": 1.0139, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.03005780346820809, | |
| "grad_norm": 0.360754142577289, | |
| "learning_rate": 6.0046189376443415e-05, | |
| "loss": 0.9982, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.03121387283236994, | |
| "grad_norm": 0.3811963374124325, | |
| "learning_rate": 6.235565819861431e-05, | |
| "loss": 1.0088, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.03236994219653179, | |
| "grad_norm": 0.3667939507833288, | |
| "learning_rate": 6.466512702078523e-05, | |
| "loss": 0.9887, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.03352601156069364, | |
| "grad_norm": 0.3988058302280759, | |
| "learning_rate": 6.697459584295612e-05, | |
| "loss": 0.9513, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.03468208092485549, | |
| "grad_norm": 0.3603023427371918, | |
| "learning_rate": 6.928406466512702e-05, | |
| "loss": 1.0239, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.035838150289017344, | |
| "grad_norm": 0.321539455275297, | |
| "learning_rate": 7.159353348729792e-05, | |
| "loss": 0.9821, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.03699421965317919, | |
| "grad_norm": 0.3548185437448054, | |
| "learning_rate": 7.390300230946883e-05, | |
| "loss": 1.0028, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.03815028901734104, | |
| "grad_norm": 0.33174112797447736, | |
| "learning_rate": 7.621247113163973e-05, | |
| "loss": 0.9852, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.03930635838150289, | |
| "grad_norm": 0.3428952862407286, | |
| "learning_rate": 7.852193995381063e-05, | |
| "loss": 1.0035, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.04046242774566474, | |
| "grad_norm": 0.3164096194909282, | |
| "learning_rate": 8.083140877598153e-05, | |
| "loss": 0.9594, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.04161849710982659, | |
| "grad_norm": 0.3210772711947753, | |
| "learning_rate": 8.314087759815243e-05, | |
| "loss": 0.9751, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.04277456647398844, | |
| "grad_norm": 0.31913209612903376, | |
| "learning_rate": 8.545034642032334e-05, | |
| "loss": 0.9816, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.04393063583815029, | |
| "grad_norm": 0.32476583892299626, | |
| "learning_rate": 8.775981524249422e-05, | |
| "loss": 1.0382, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.04508670520231214, | |
| "grad_norm": 0.33804033601960015, | |
| "learning_rate": 9.006928406466512e-05, | |
| "loss": 1.0278, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.046242774566473986, | |
| "grad_norm": 0.3034185595284857, | |
| "learning_rate": 9.237875288683603e-05, | |
| "loss": 0.9209, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.047398843930635835, | |
| "grad_norm": 0.30966667195251285, | |
| "learning_rate": 9.468822170900693e-05, | |
| "loss": 0.9929, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.048554913294797684, | |
| "grad_norm": 0.3110271670167101, | |
| "learning_rate": 9.699769053117783e-05, | |
| "loss": 0.999, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.04971098265895954, | |
| "grad_norm": 0.3168339881753396, | |
| "learning_rate": 9.930715935334873e-05, | |
| "loss": 0.9801, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.05086705202312139, | |
| "grad_norm": 0.2980534767723837, | |
| "learning_rate": 0.00010161662817551964, | |
| "loss": 0.964, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.05202312138728324, | |
| "grad_norm": 0.29439975733255125, | |
| "learning_rate": 0.00010392609699769054, | |
| "loss": 1.0141, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.05317919075144509, | |
| "grad_norm": 0.2939300503984728, | |
| "learning_rate": 0.00010623556581986143, | |
| "loss": 1.029, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.05433526011560694, | |
| "grad_norm": 0.2921065506824694, | |
| "learning_rate": 0.00010854503464203234, | |
| "loss": 1.0008, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.055491329479768786, | |
| "grad_norm": 0.29281624873588324, | |
| "learning_rate": 0.00011085450346420324, | |
| "loss": 1.0286, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.056647398843930635, | |
| "grad_norm": 0.30008365055838865, | |
| "learning_rate": 0.00011316397228637415, | |
| "loss": 1.0009, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.057803468208092484, | |
| "grad_norm": 0.2821304520615669, | |
| "learning_rate": 0.00011547344110854503, | |
| "loss": 1.0172, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.058959537572254334, | |
| "grad_norm": 0.28925572801814625, | |
| "learning_rate": 0.00011778290993071594, | |
| "loss": 1.0106, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.06011560693641618, | |
| "grad_norm": 0.2979979142081887, | |
| "learning_rate": 0.00012009237875288683, | |
| "loss": 0.9851, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.06127167630057803, | |
| "grad_norm": 0.27608874909832487, | |
| "learning_rate": 0.00012240184757505776, | |
| "loss": 0.9386, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.06242774566473988, | |
| "grad_norm": 0.270653606190121, | |
| "learning_rate": 0.00012471131639722863, | |
| "loss": 0.9823, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.06358381502890173, | |
| "grad_norm": 0.2889956899438749, | |
| "learning_rate": 0.00012702078521939955, | |
| "loss": 1.0096, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.06473988439306358, | |
| "grad_norm": 0.30488051626437424, | |
| "learning_rate": 0.00012933025404157045, | |
| "loss": 0.9819, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.06589595375722543, | |
| "grad_norm": 0.2884447508894878, | |
| "learning_rate": 0.00013163972286374135, | |
| "loss": 0.9468, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.06705202312138728, | |
| "grad_norm": 0.2755861252646099, | |
| "learning_rate": 0.00013394919168591225, | |
| "loss": 0.9828, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.06820809248554913, | |
| "grad_norm": 0.3025742735794824, | |
| "learning_rate": 0.00013625866050808315, | |
| "loss": 1.0178, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.06936416184971098, | |
| "grad_norm": 0.28423295151616335, | |
| "learning_rate": 0.00013856812933025404, | |
| "loss": 0.9945, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.07052023121387284, | |
| "grad_norm": 0.27585086154702476, | |
| "learning_rate": 0.00014087759815242494, | |
| "loss": 0.9695, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.07167630057803469, | |
| "grad_norm": 0.3038442212974576, | |
| "learning_rate": 0.00014318706697459584, | |
| "loss": 0.9808, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.07283236994219654, | |
| "grad_norm": 0.27277935404650916, | |
| "learning_rate": 0.00014549653579676674, | |
| "loss": 0.9846, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 0.07398843930635839, | |
| "grad_norm": 0.29591421870168255, | |
| "learning_rate": 0.00014780600461893767, | |
| "loss": 0.9954, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.07514450867052024, | |
| "grad_norm": 0.2823015918956852, | |
| "learning_rate": 0.00015011547344110854, | |
| "loss": 0.9674, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 0.07630057803468208, | |
| "grad_norm": 0.27344790533462154, | |
| "learning_rate": 0.00015242494226327946, | |
| "loss": 1.0236, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.07745664739884393, | |
| "grad_norm": 0.2949347092864385, | |
| "learning_rate": 0.00015473441108545036, | |
| "loss": 0.9813, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 0.07861271676300578, | |
| "grad_norm": 0.272235850063355, | |
| "learning_rate": 0.00015704387990762126, | |
| "loss": 0.987, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.07976878612716763, | |
| "grad_norm": 0.28565982534338485, | |
| "learning_rate": 0.00015935334872979216, | |
| "loss": 0.9716, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 0.08092485549132948, | |
| "grad_norm": 0.28274111136822716, | |
| "learning_rate": 0.00016166281755196306, | |
| "loss": 0.9779, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.08208092485549133, | |
| "grad_norm": 0.2878620812403313, | |
| "learning_rate": 0.00016397228637413396, | |
| "loss": 1.0209, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 0.08323699421965318, | |
| "grad_norm": 0.2865629307216328, | |
| "learning_rate": 0.00016628175519630485, | |
| "loss": 0.9921, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.08439306358381503, | |
| "grad_norm": 0.2826816337246457, | |
| "learning_rate": 0.00016859122401847575, | |
| "loss": 0.9348, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 0.08554913294797688, | |
| "grad_norm": 0.2881907360895622, | |
| "learning_rate": 0.00017090069284064668, | |
| "loss": 1.0231, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.08670520231213873, | |
| "grad_norm": 0.27638308021239116, | |
| "learning_rate": 0.00017321016166281755, | |
| "loss": 1.0079, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 0.08786127167630058, | |
| "grad_norm": 0.28798387568151884, | |
| "learning_rate": 0.00017551963048498845, | |
| "loss": 0.9165, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.08901734104046242, | |
| "grad_norm": 0.2767437621962458, | |
| "learning_rate": 0.00017782909930715937, | |
| "loss": 0.9489, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 0.09017341040462427, | |
| "grad_norm": 0.2877748564441354, | |
| "learning_rate": 0.00018013856812933024, | |
| "loss": 1.0022, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.09132947976878612, | |
| "grad_norm": 0.28925322685576144, | |
| "learning_rate": 0.00018244803695150117, | |
| "loss": 1.004, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 0.09248554913294797, | |
| "grad_norm": 0.31070512225405156, | |
| "learning_rate": 0.00018475750577367207, | |
| "loss": 0.9753, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.09364161849710982, | |
| "grad_norm": 0.31735644814371816, | |
| "learning_rate": 0.00018706697459584297, | |
| "loss": 0.9781, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 0.09479768786127167, | |
| "grad_norm": 0.2974104968051762, | |
| "learning_rate": 0.00018937644341801387, | |
| "loss": 0.9659, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.09595375722543352, | |
| "grad_norm": 0.2701025540904289, | |
| "learning_rate": 0.00019168591224018476, | |
| "loss": 0.9294, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 0.09710982658959537, | |
| "grad_norm": 0.27428411358071536, | |
| "learning_rate": 0.00019399538106235566, | |
| "loss": 0.9513, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.09826589595375723, | |
| "grad_norm": 0.2745240121214777, | |
| "learning_rate": 0.0001963048498845266, | |
| "loss": 1.0152, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 0.09942196531791908, | |
| "grad_norm": 0.274405426645142, | |
| "learning_rate": 0.00019861431870669746, | |
| "loss": 0.9863, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.10057803468208093, | |
| "grad_norm": 0.9048729498402529, | |
| "learning_rate": 0.00019999986968812804, | |
| "loss": 0.9992, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 0.10173410404624278, | |
| "grad_norm": 0.29062903824052616, | |
| "learning_rate": 0.00019999840368346898, | |
| "loss": 1.0509, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.10289017341040463, | |
| "grad_norm": 0.28317998532870003, | |
| "learning_rate": 0.0001999953088082702, | |
| "loss": 0.9774, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 0.10404624277456648, | |
| "grad_norm": 0.3000365098515979, | |
| "learning_rate": 0.000199990585112944, | |
| "loss": 0.9822, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.10520231213872833, | |
| "grad_norm": 0.28508661433222776, | |
| "learning_rate": 0.00019998423267443454, | |
| "loss": 0.9991, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 0.10635838150289018, | |
| "grad_norm": 0.300390972643917, | |
| "learning_rate": 0.00019997625159621642, | |
| "loss": 0.9411, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.10751445086705202, | |
| "grad_norm": 0.3065580188577536, | |
| "learning_rate": 0.0001999666420082932, | |
| "loss": 1.0153, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 0.10867052023121387, | |
| "grad_norm": 0.30176770344761106, | |
| "learning_rate": 0.00019995540406719507, | |
| "loss": 0.9451, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.10982658959537572, | |
| "grad_norm": 0.28363462288539226, | |
| "learning_rate": 0.0001999425379559765, | |
| "loss": 1.0229, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 0.11098265895953757, | |
| "grad_norm": 0.2979648349669768, | |
| "learning_rate": 0.00019992804388421312, | |
| "loss": 0.9615, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.11213872832369942, | |
| "grad_norm": 0.3533699255705701, | |
| "learning_rate": 0.00019991192208799837, | |
| "loss": 0.9945, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 0.11329479768786127, | |
| "grad_norm": 0.29374690183301444, | |
| "learning_rate": 0.0001998941728299396, | |
| "loss": 0.9481, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.11445086705202312, | |
| "grad_norm": 0.27638233259638606, | |
| "learning_rate": 0.0001998747963991539, | |
| "loss": 0.976, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 0.11560693641618497, | |
| "grad_norm": 0.29591583372167063, | |
| "learning_rate": 0.00019985379311126327, | |
| "loss": 0.9776, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.11676300578034682, | |
| "grad_norm": 0.28596466519406494, | |
| "learning_rate": 0.00019983116330838955, | |
| "loss": 1.0003, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 0.11791907514450867, | |
| "grad_norm": 0.27570121782043344, | |
| "learning_rate": 0.00019980690735914877, | |
| "loss": 0.9797, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.11907514450867052, | |
| "grad_norm": 0.30038501431153675, | |
| "learning_rate": 0.0001997810256586453, | |
| "loss": 0.9865, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 0.12023121387283237, | |
| "grad_norm": 0.31132305182282943, | |
| "learning_rate": 0.00019975351862846523, | |
| "loss": 1.0071, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.12138728323699421, | |
| "grad_norm": 0.28621069496048757, | |
| "learning_rate": 0.00019972438671666967, | |
| "loss": 0.9877, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 0.12254335260115606, | |
| "grad_norm": 0.3030051142967184, | |
| "learning_rate": 0.00019969363039778728, | |
| "loss": 0.9894, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.12369942196531791, | |
| "grad_norm": 0.2866405334142299, | |
| "learning_rate": 0.0001996612501728067, | |
| "loss": 0.9893, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 0.12485549132947976, | |
| "grad_norm": 0.2867249309375898, | |
| "learning_rate": 0.00019962724656916826, | |
| "loss": 0.9765, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.1260115606936416, | |
| "grad_norm": 0.31798324367678865, | |
| "learning_rate": 0.00019959162014075553, | |
| "loss": 0.9465, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 0.12716763005780346, | |
| "grad_norm": 0.30253196958089823, | |
| "learning_rate": 0.0001995543714678861, | |
| "loss": 0.9635, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.1283236994219653, | |
| "grad_norm": 0.28954658171708875, | |
| "learning_rate": 0.00019951550115730244, | |
| "loss": 1.0041, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 0.12947976878612716, | |
| "grad_norm": 0.28359238422516453, | |
| "learning_rate": 0.00019947500984216157, | |
| "loss": 0.9837, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.130635838150289, | |
| "grad_norm": 0.29624734780777734, | |
| "learning_rate": 0.00019943289818202519, | |
| "loss": 0.9375, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 0.13179190751445086, | |
| "grad_norm": 0.291198302886137, | |
| "learning_rate": 0.0001993891668628486, | |
| "loss": 0.9665, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.1329479768786127, | |
| "grad_norm": 0.31056892991094237, | |
| "learning_rate": 0.00019934381659696989, | |
| "loss": 0.9414, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 0.13410404624277455, | |
| "grad_norm": 0.3136680287460065, | |
| "learning_rate": 0.0001992968481230978, | |
| "loss": 1.0442, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.1352601156069364, | |
| "grad_norm": 0.30128304774644027, | |
| "learning_rate": 0.0001992482622063003, | |
| "loss": 0.9916, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 0.13641618497109825, | |
| "grad_norm": 0.29097714851626455, | |
| "learning_rate": 0.00019919805963799166, | |
| "loss": 0.9947, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.1375722543352601, | |
| "grad_norm": 0.280417582339227, | |
| "learning_rate": 0.0001991462412359198, | |
| "loss": 0.9825, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 0.13872832369942195, | |
| "grad_norm": 0.28842456996684646, | |
| "learning_rate": 0.00019909280784415287, | |
| "loss": 1.0237, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.13988439306358383, | |
| "grad_norm": 0.2874190385868597, | |
| "learning_rate": 0.00019903776033306555, | |
| "loss": 0.9611, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 0.14104046242774568, | |
| "grad_norm": 0.2854599682323198, | |
| "learning_rate": 0.00019898109959932478, | |
| "loss": 0.9879, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.14219653179190753, | |
| "grad_norm": 0.2986976522432421, | |
| "learning_rate": 0.0001989228265658754, | |
| "loss": 0.9911, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 0.14335260115606938, | |
| "grad_norm": 0.329057703574734, | |
| "learning_rate": 0.00019886294218192477, | |
| "loss": 0.9714, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.14450867052023122, | |
| "grad_norm": 0.29538740567126964, | |
| "learning_rate": 0.00019880144742292753, | |
| "loss": 0.955, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 0.14566473988439307, | |
| "grad_norm": 0.32615973261215037, | |
| "learning_rate": 0.00019873834329056975, | |
| "loss": 0.9789, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.14682080924855492, | |
| "grad_norm": 0.26819671057976713, | |
| "learning_rate": 0.00019867363081275242, | |
| "loss": 0.9471, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 0.14797687861271677, | |
| "grad_norm": 0.28488339516835476, | |
| "learning_rate": 0.00019860731104357485, | |
| "loss": 0.9779, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.14913294797687862, | |
| "grad_norm": 0.27998012828783303, | |
| "learning_rate": 0.00019853938506331749, | |
| "loss": 0.951, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 0.15028901734104047, | |
| "grad_norm": 0.29388395642325527, | |
| "learning_rate": 0.00019846985397842427, | |
| "loss": 0.9554, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.15144508670520232, | |
| "grad_norm": 0.2896808457234832, | |
| "learning_rate": 0.0001983987189214846, | |
| "loss": 0.9651, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 0.15260115606936417, | |
| "grad_norm": 0.30032384976542736, | |
| "learning_rate": 0.000198325981051215, | |
| "loss": 0.943, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.15375722543352602, | |
| "grad_norm": 0.2977312487124492, | |
| "learning_rate": 0.00019825164155244012, | |
| "loss": 0.9887, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 0.15491329479768787, | |
| "grad_norm": 0.341662364901083, | |
| "learning_rate": 0.00019817570163607347, | |
| "loss": 1.0059, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.15606936416184972, | |
| "grad_norm": 0.2996891760902693, | |
| "learning_rate": 0.00019809816253909773, | |
| "loss": 0.9673, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 0.15722543352601157, | |
| "grad_norm": 0.2882952073089595, | |
| "learning_rate": 0.00019801902552454454, | |
| "loss": 0.9558, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.15838150289017341, | |
| "grad_norm": 0.3203389676438109, | |
| "learning_rate": 0.00019793829188147406, | |
| "loss": 1.0122, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 0.15953757225433526, | |
| "grad_norm": 0.30270844487268483, | |
| "learning_rate": 0.00019785596292495376, | |
| "loss": 0.9822, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.1606936416184971, | |
| "grad_norm": 0.27999850496563145, | |
| "learning_rate": 0.00019777203999603717, | |
| "loss": 0.9841, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 0.16184971098265896, | |
| "grad_norm": 0.29383524324706056, | |
| "learning_rate": 0.000197686524461742, | |
| "loss": 1.0269, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.1630057803468208, | |
| "grad_norm": 0.28066838492867907, | |
| "learning_rate": 0.0001975994177150278, | |
| "loss": 0.9927, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 0.16416184971098266, | |
| "grad_norm": 0.31122106212206363, | |
| "learning_rate": 0.0001975107211747734, | |
| "loss": 0.9632, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.1653179190751445, | |
| "grad_norm": 0.30120413493928255, | |
| "learning_rate": 0.00019742043628575364, | |
| "loss": 0.9739, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 0.16647398843930636, | |
| "grad_norm": 0.30660646504263267, | |
| "learning_rate": 0.00019732856451861594, | |
| "loss": 0.976, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.1676300578034682, | |
| "grad_norm": 0.2856850255835453, | |
| "learning_rate": 0.0001972351073698564, | |
| "loss": 0.9848, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 0.16878612716763006, | |
| "grad_norm": 0.3147493299962656, | |
| "learning_rate": 0.0001971400663617952, | |
| "loss": 0.9921, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.1699421965317919, | |
| "grad_norm": 0.29251745238857035, | |
| "learning_rate": 0.0001970434430425521, | |
| "loss": 0.9398, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 0.17109826589595376, | |
| "grad_norm": 0.2938440274420457, | |
| "learning_rate": 0.000196945238986021, | |
| "loss": 1.0153, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.1722543352601156, | |
| "grad_norm": 0.2898363733115959, | |
| "learning_rate": 0.00019684545579184433, | |
| "loss": 0.9576, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 0.17341040462427745, | |
| "grad_norm": 0.3088888037796155, | |
| "learning_rate": 0.00019674409508538718, | |
| "loss": 0.9696, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.1745664739884393, | |
| "grad_norm": 0.2942346284280998, | |
| "learning_rate": 0.0001966411585177105, | |
| "loss": 1.0203, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 0.17572254335260115, | |
| "grad_norm": 0.2852125727236794, | |
| "learning_rate": 0.00019653664776554455, | |
| "loss": 0.9556, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.176878612716763, | |
| "grad_norm": 0.3016262447783914, | |
| "learning_rate": 0.0001964305645312613, | |
| "loss": 0.9896, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 0.17803468208092485, | |
| "grad_norm": 0.30255168675186533, | |
| "learning_rate": 0.00019632291054284693, | |
| "loss": 0.9839, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.1791907514450867, | |
| "grad_norm": 0.2806238558760804, | |
| "learning_rate": 0.0001962136875538735, | |
| "loss": 0.9748, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 0.18034682080924855, | |
| "grad_norm": 0.2945672906624874, | |
| "learning_rate": 0.00019610289734347053, | |
| "loss": 0.9479, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.1815028901734104, | |
| "grad_norm": 0.3012247410303452, | |
| "learning_rate": 0.00019599054171629595, | |
| "loss": 1.0132, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 0.18265895953757225, | |
| "grad_norm": 0.2944499117709193, | |
| "learning_rate": 0.0001958766225025066, | |
| "loss": 0.9336, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.1838150289017341, | |
| "grad_norm": 0.2989322109974369, | |
| "learning_rate": 0.0001957611415577287, | |
| "loss": 0.9857, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 0.18497109826589594, | |
| "grad_norm": 0.2859334477998094, | |
| "learning_rate": 0.0001956441007630273, | |
| "loss": 0.9831, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.1861271676300578, | |
| "grad_norm": 0.27928427673168016, | |
| "learning_rate": 0.0001955255020248759, | |
| "loss": 0.9582, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 0.18728323699421964, | |
| "grad_norm": 0.3044716011707441, | |
| "learning_rate": 0.00019540534727512522, | |
| "loss": 1.0061, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.1884393063583815, | |
| "grad_norm": 0.30123009109430604, | |
| "learning_rate": 0.00019528363847097185, | |
| "loss": 1.0015, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 0.18959537572254334, | |
| "grad_norm": 0.2983431588431708, | |
| "learning_rate": 0.00019516037759492627, | |
| "loss": 0.9917, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.1907514450867052, | |
| "grad_norm": 0.3045474200889231, | |
| "learning_rate": 0.00019503556665478067, | |
| "loss": 0.9924, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 0.19190751445086704, | |
| "grad_norm": 0.3023598788495744, | |
| "learning_rate": 0.00019490920768357607, | |
| "loss": 0.9824, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.1930635838150289, | |
| "grad_norm": 0.288004788186244, | |
| "learning_rate": 0.00019478130273956943, | |
| "loss": 0.9756, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 0.19421965317919074, | |
| "grad_norm": 0.28774556008482255, | |
| "learning_rate": 0.00019465185390619996, | |
| "loss": 0.9292, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.19537572254335261, | |
| "grad_norm": 0.29766392431820693, | |
| "learning_rate": 0.00019452086329205522, | |
| "loss": 1.008, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 0.19653179190751446, | |
| "grad_norm": 0.2968523784854959, | |
| "learning_rate": 0.00019438833303083678, | |
| "loss": 0.9469, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.1976878612716763, | |
| "grad_norm": 0.32050710908212127, | |
| "learning_rate": 0.00019425426528132546, | |
| "loss": 0.9584, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 0.19884393063583816, | |
| "grad_norm": 0.30039833468190924, | |
| "learning_rate": 0.00019411866222734627, | |
| "loss": 0.9345, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.3003418210003542, | |
| "learning_rate": 0.00019398152607773264, | |
| "loss": 1.0149, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 0.20115606936416186, | |
| "grad_norm": 0.28740303119481114, | |
| "learning_rate": 0.00019384285906629055, | |
| "loss": 0.9453, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.2023121387283237, | |
| "grad_norm": 0.3022020472256075, | |
| "learning_rate": 0.00019370266345176214, | |
| "loss": 0.9964, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 0.20346820809248556, | |
| "grad_norm": 0.2887054624910223, | |
| "learning_rate": 0.00019356094151778895, | |
| "loss": 0.9857, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.2046242774566474, | |
| "grad_norm": 0.38190567119224494, | |
| "learning_rate": 0.00019341769557287467, | |
| "loss": 0.9554, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 0.20578034682080926, | |
| "grad_norm": 0.29810678010523667, | |
| "learning_rate": 0.00019327292795034753, | |
| "loss": 0.9138, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.2069364161849711, | |
| "grad_norm": 0.30694177853911697, | |
| "learning_rate": 0.00019312664100832233, | |
| "loss": 0.9504, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 0.20809248554913296, | |
| "grad_norm": 0.31819590085942273, | |
| "learning_rate": 0.00019297883712966204, | |
| "loss": 0.9969, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.2092485549132948, | |
| "grad_norm": 0.2915158769948663, | |
| "learning_rate": 0.00019282951872193885, | |
| "loss": 1.0207, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 0.21040462427745665, | |
| "grad_norm": 0.3020525391391881, | |
| "learning_rate": 0.0001926786882173952, | |
| "loss": 0.9424, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.2115606936416185, | |
| "grad_norm": 0.3095938351879466, | |
| "learning_rate": 0.000192526348072904, | |
| "loss": 0.9762, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 0.21271676300578035, | |
| "grad_norm": 0.32067080864793646, | |
| "learning_rate": 0.0001923725007699285, | |
| "loss": 0.9423, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.2138728323699422, | |
| "grad_norm": 0.3144763229931657, | |
| "learning_rate": 0.00019221714881448217, | |
| "loss": 0.9824, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 0.21502890173410405, | |
| "grad_norm": 0.3179959968229521, | |
| "learning_rate": 0.0001920602947370876, | |
| "loss": 0.9829, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.2161849710982659, | |
| "grad_norm": 0.30113560950220297, | |
| "learning_rate": 0.00019190194109273544, | |
| "loss": 0.9463, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 0.21734104046242775, | |
| "grad_norm": 0.2896899865821755, | |
| "learning_rate": 0.00019174209046084276, | |
| "loss": 0.9786, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.2184971098265896, | |
| "grad_norm": 0.2969332481357098, | |
| "learning_rate": 0.00019158074544521094, | |
| "loss": 1.0183, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 0.21965317919075145, | |
| "grad_norm": 0.29984068010125464, | |
| "learning_rate": 0.0001914179086739834, | |
| "loss": 0.9688, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.2208092485549133, | |
| "grad_norm": 0.3179298483108417, | |
| "learning_rate": 0.0001912535827996026, | |
| "loss": 1.0213, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 0.22196531791907514, | |
| "grad_norm": 0.30071667875773894, | |
| "learning_rate": 0.0001910877704987671, | |
| "loss": 0.9947, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.223121387283237, | |
| "grad_norm": 0.3162017726279279, | |
| "learning_rate": 0.00019092047447238773, | |
| "loss": 0.9765, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 0.22427745664739884, | |
| "grad_norm": 0.31999805840625895, | |
| "learning_rate": 0.0001907516974455436, | |
| "loss": 0.9956, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.2254335260115607, | |
| "grad_norm": 0.2931503664377131, | |
| "learning_rate": 0.00019058144216743797, | |
| "loss": 1.002, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 0.22658959537572254, | |
| "grad_norm": 0.30298932952701096, | |
| "learning_rate": 0.0001904097114113531, | |
| "loss": 1.0159, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.2277456647398844, | |
| "grad_norm": 0.29804869616385626, | |
| "learning_rate": 0.0001902365079746054, | |
| "loss": 0.9771, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 0.22890173410404624, | |
| "grad_norm": 0.30761841093550274, | |
| "learning_rate": 0.00019006183467849957, | |
| "loss": 1.0157, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.2300578034682081, | |
| "grad_norm": 0.3080044180069082, | |
| "learning_rate": 0.000189885694368283, | |
| "loss": 0.9463, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 0.23121387283236994, | |
| "grad_norm": 0.29852967122969754, | |
| "learning_rate": 0.00018970808991309904, | |
| "loss": 1.0021, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.2323699421965318, | |
| "grad_norm": 0.2987990831868652, | |
| "learning_rate": 0.00018952902420594058, | |
| "loss": 0.9492, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 0.23352601156069364, | |
| "grad_norm": 0.30446826555699585, | |
| "learning_rate": 0.0001893485001636026, | |
| "loss": 1.0019, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.23468208092485549, | |
| "grad_norm": 0.30998357590060016, | |
| "learning_rate": 0.00018916652072663515, | |
| "loss": 0.9478, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 0.23583815028901733, | |
| "grad_norm": 0.3109448921090665, | |
| "learning_rate": 0.0001889830888592949, | |
| "loss": 0.9797, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.23699421965317918, | |
| "grad_norm": 0.30658444282489306, | |
| "learning_rate": 0.00018879820754949718, | |
| "loss": 0.9976, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 0.23815028901734103, | |
| "grad_norm": 0.32337481964353393, | |
| "learning_rate": 0.0001886118798087673, | |
| "loss": 0.9622, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.23930635838150288, | |
| "grad_norm": 0.30934913326885294, | |
| "learning_rate": 0.00018842410867219136, | |
| "loss": 1.0095, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 0.24046242774566473, | |
| "grad_norm": 0.32554022935815935, | |
| "learning_rate": 0.0001882348971983669, | |
| "loss": 1.0082, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.24161849710982658, | |
| "grad_norm": 0.28387403686918444, | |
| "learning_rate": 0.0001880442484693531, | |
| "loss": 0.9433, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 0.24277456647398843, | |
| "grad_norm": 0.2889713456813008, | |
| "learning_rate": 0.0001878521655906205, | |
| "loss": 0.994, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.24393063583815028, | |
| "grad_norm": 0.3038933468668014, | |
| "learning_rate": 0.00018765865169100048, | |
| "loss": 0.966, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 0.24508670520231213, | |
| "grad_norm": 0.294597892145829, | |
| "learning_rate": 0.00018746370992263423, | |
| "loss": 0.9501, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.24624277456647398, | |
| "grad_norm": 0.30895343503048994, | |
| "learning_rate": 0.00018726734346092148, | |
| "loss": 0.9663, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 0.24739884393063583, | |
| "grad_norm": 0.32867090123543136, | |
| "learning_rate": 0.00018706955550446878, | |
| "loss": 1.0, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.24855491329479767, | |
| "grad_norm": 0.2860732876796574, | |
| "learning_rate": 0.00018687034927503728, | |
| "loss": 0.9282, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 0.24971098265895952, | |
| "grad_norm": 0.3622792683930793, | |
| "learning_rate": 0.00018666972801749035, | |
| "loss": 0.9534, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.2508670520231214, | |
| "grad_norm": 0.3215073797028641, | |
| "learning_rate": 0.00018646769499974076, | |
| "loss": 1.0177, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 0.2520231213872832, | |
| "grad_norm": 0.30391881014243827, | |
| "learning_rate": 0.00018626425351269733, | |
| "loss": 1.0213, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.25317919075144507, | |
| "grad_norm": 0.2959182253635083, | |
| "learning_rate": 0.00018605940687021133, | |
| "loss": 0.9265, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 0.2543352601156069, | |
| "grad_norm": 0.3054358776400607, | |
| "learning_rate": 0.00018585315840902275, | |
| "loss": 0.9566, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.25549132947976877, | |
| "grad_norm": 0.3038159760611022, | |
| "learning_rate": 0.00018564551148870563, | |
| "loss": 0.9728, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 0.2566473988439306, | |
| "grad_norm": 0.295408221908172, | |
| "learning_rate": 0.0001854364694916134, | |
| "loss": 0.9769, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.25780346820809247, | |
| "grad_norm": 0.3040191717007843, | |
| "learning_rate": 0.00018522603582282396, | |
| "loss": 0.9745, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 0.2589595375722543, | |
| "grad_norm": 0.3276795968564327, | |
| "learning_rate": 0.0001850142139100841, | |
| "loss": 0.9843, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.26011560693641617, | |
| "grad_norm": 0.30521432519770036, | |
| "learning_rate": 0.0001848010072037536, | |
| "loss": 0.9958, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 0.261271676300578, | |
| "grad_norm": 0.3074723851239691, | |
| "learning_rate": 0.0001845864191767491, | |
| "loss": 1.0061, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.26242774566473986, | |
| "grad_norm": 0.38542353780888683, | |
| "learning_rate": 0.0001843704533244876, | |
| "loss": 0.9527, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 0.2635838150289017, | |
| "grad_norm": 0.31655887417225065, | |
| "learning_rate": 0.00018415311316482934, | |
| "loss": 0.9677, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.26473988439306356, | |
| "grad_norm": 0.296622948449929, | |
| "learning_rate": 0.00018393440223802077, | |
| "loss": 1.0277, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 0.2658959537572254, | |
| "grad_norm": 0.30339302168936083, | |
| "learning_rate": 0.0001837143241066365, | |
| "loss": 0.9996, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.26705202312138726, | |
| "grad_norm": 0.321536892508796, | |
| "learning_rate": 0.00018349288235552168, | |
| "loss": 0.9599, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 0.2682080924855491, | |
| "grad_norm": 0.3220388229874098, | |
| "learning_rate": 0.0001832700805917333, | |
| "loss": 0.9957, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.26936416184971096, | |
| "grad_norm": 0.2951235948754925, | |
| "learning_rate": 0.00018304592244448156, | |
| "loss": 1.0261, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 0.2705202312138728, | |
| "grad_norm": 0.3044130581874429, | |
| "learning_rate": 0.0001828204115650708, | |
| "loss": 0.9871, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.27167630057803466, | |
| "grad_norm": 0.3009718363326464, | |
| "learning_rate": 0.00018259355162684, | |
| "loss": 0.9963, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 0.2728323699421965, | |
| "grad_norm": 0.28905525049692726, | |
| "learning_rate": 0.00018236534632510277, | |
| "loss": 0.9848, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.27398843930635836, | |
| "grad_norm": 0.35621478263055945, | |
| "learning_rate": 0.00018213579937708735, | |
| "loss": 0.9529, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 0.2751445086705202, | |
| "grad_norm": 0.3185821009153887, | |
| "learning_rate": 0.00018190491452187613, | |
| "loss": 1.0045, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.27630057803468205, | |
| "grad_norm": 0.3174337839488537, | |
| "learning_rate": 0.00018167269552034446, | |
| "loss": 0.9549, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 0.2774566473988439, | |
| "grad_norm": 0.2929697771476232, | |
| "learning_rate": 0.00018143914615509967, | |
| "loss": 0.9463, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.2786127167630058, | |
| "grad_norm": 0.3218921944038643, | |
| "learning_rate": 0.00018120427023041925, | |
| "loss": 1.0432, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 0.27976878612716766, | |
| "grad_norm": 0.29522034854514784, | |
| "learning_rate": 0.00018096807157218909, | |
| "loss": 0.9354, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.2809248554913295, | |
| "grad_norm": 0.31103148068930697, | |
| "learning_rate": 0.0001807305540278409, | |
| "loss": 0.9275, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 0.28208092485549136, | |
| "grad_norm": 0.3268759634134707, | |
| "learning_rate": 0.00018049172146628975, | |
| "loss": 1.0122, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.2832369942196532, | |
| "grad_norm": 0.2992667041917287, | |
| "learning_rate": 0.00018025157777787102, | |
| "loss": 0.9422, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 0.28439306358381505, | |
| "grad_norm": 0.31617502406730474, | |
| "learning_rate": 0.00018001012687427688, | |
| "loss": 0.9909, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.2855491329479769, | |
| "grad_norm": 0.31251457847228237, | |
| "learning_rate": 0.0001797673726884928, | |
| "loss": 0.9885, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 0.28670520231213875, | |
| "grad_norm": 0.32021979333618866, | |
| "learning_rate": 0.00017952331917473336, | |
| "loss": 0.9396, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.2878612716763006, | |
| "grad_norm": 0.34841403249467495, | |
| "learning_rate": 0.00017927797030837768, | |
| "loss": 0.9188, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 0.28901734104046245, | |
| "grad_norm": 0.40965133837398776, | |
| "learning_rate": 0.0001790313300859051, | |
| "loss": 0.9582, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.2901734104046243, | |
| "grad_norm": 0.3128265434514317, | |
| "learning_rate": 0.00017878340252482956, | |
| "loss": 0.9891, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 0.29132947976878615, | |
| "grad_norm": 0.3762731976913158, | |
| "learning_rate": 0.00017853419166363458, | |
| "loss": 0.973, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.292485549132948, | |
| "grad_norm": 0.3150565577316399, | |
| "learning_rate": 0.00017828370156170727, | |
| "loss": 0.9777, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 0.29364161849710985, | |
| "grad_norm": 0.3124944800404496, | |
| "learning_rate": 0.00017803193629927223, | |
| "loss": 0.984, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.2947976878612717, | |
| "grad_norm": 0.3456192268358289, | |
| "learning_rate": 0.0001777788999773251, | |
| "loss": 0.9881, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 0.29595375722543354, | |
| "grad_norm": 0.309939339205246, | |
| "learning_rate": 0.0001775245967175658, | |
| "loss": 0.9483, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.2971098265895954, | |
| "grad_norm": 0.29097211271370754, | |
| "learning_rate": 0.00017726903066233134, | |
| "loss": 0.9438, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 0.29826589595375724, | |
| "grad_norm": 0.29586840025056343, | |
| "learning_rate": 0.00017701220597452833, | |
| "loss": 0.9754, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.2994219653179191, | |
| "grad_norm": 0.3017342795593241, | |
| "learning_rate": 0.0001767541268375652, | |
| "loss": 0.9805, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 0.30057803468208094, | |
| "grad_norm": 0.32061970031794484, | |
| "learning_rate": 0.00017649479745528417, | |
| "loss": 0.9818, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.3017341040462428, | |
| "grad_norm": 0.3011478634925178, | |
| "learning_rate": 0.00017623422205189252, | |
| "loss": 0.9815, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 0.30289017341040464, | |
| "grad_norm": 0.3086640164396148, | |
| "learning_rate": 0.000175972404871894, | |
| "loss": 0.9756, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.3040462427745665, | |
| "grad_norm": 0.31272520758866784, | |
| "learning_rate": 0.0001757093501800196, | |
| "loss": 0.9993, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 0.30520231213872834, | |
| "grad_norm": 0.3574223169536535, | |
| "learning_rate": 0.0001754450622611581, | |
| "loss": 0.9468, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.3063583815028902, | |
| "grad_norm": 0.3147796469526059, | |
| "learning_rate": 0.0001751795454202863, | |
| "loss": 0.9848, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 0.30751445086705204, | |
| "grad_norm": 0.3118480322453381, | |
| "learning_rate": 0.0001749128039823988, | |
| "loss": 1.0249, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.3086705202312139, | |
| "grad_norm": 0.3023964221368843, | |
| "learning_rate": 0.00017464484229243768, | |
| "loss": 0.9232, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 0.30982658959537573, | |
| "grad_norm": 0.3310385612654826, | |
| "learning_rate": 0.0001743756647152216, | |
| "loss": 0.9975, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.3109826589595376, | |
| "grad_norm": 0.3318179078521207, | |
| "learning_rate": 0.00017410527563537488, | |
| "loss": 0.9776, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 0.31213872832369943, | |
| "grad_norm": 0.31582505608794464, | |
| "learning_rate": 0.00017383367945725584, | |
| "loss": 0.9191, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.3132947976878613, | |
| "grad_norm": 0.30919652895574773, | |
| "learning_rate": 0.00017356088060488525, | |
| "loss": 0.9813, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 0.31445086705202313, | |
| "grad_norm": 0.30484218478884034, | |
| "learning_rate": 0.00017328688352187416, | |
| "loss": 0.9791, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.315606936416185, | |
| "grad_norm": 0.3016535900655947, | |
| "learning_rate": 0.00017301169267135163, | |
| "loss": 0.9918, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 0.31676300578034683, | |
| "grad_norm": 0.2828981498735541, | |
| "learning_rate": 0.00017273531253589187, | |
| "loss": 0.9266, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.3179190751445087, | |
| "grad_norm": 0.3077602888091621, | |
| "learning_rate": 0.00017245774761744134, | |
| "loss": 0.9674, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 0.3190751445086705, | |
| "grad_norm": 0.3099506136784924, | |
| "learning_rate": 0.00017217900243724543, | |
| "loss": 0.9836, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.3202312138728324, | |
| "grad_norm": 0.31771107668630955, | |
| "learning_rate": 0.00017189908153577473, | |
| "loss": 0.9387, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 0.3213872832369942, | |
| "grad_norm": 0.3121605825107148, | |
| "learning_rate": 0.0001716179894726511, | |
| "loss": 1.0108, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.3225433526011561, | |
| "grad_norm": 0.32262006217730343, | |
| "learning_rate": 0.0001713357308265735, | |
| "loss": 1.0374, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 0.3236994219653179, | |
| "grad_norm": 0.3209655049858561, | |
| "learning_rate": 0.0001710523101952432, | |
| "loss": 0.9936, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.3248554913294798, | |
| "grad_norm": 0.32077514063057966, | |
| "learning_rate": 0.00017076773219528905, | |
| "loss": 0.9704, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 0.3260115606936416, | |
| "grad_norm": 0.30932663652684755, | |
| "learning_rate": 0.0001704820014621923, | |
| "loss": 1.003, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.32716763005780347, | |
| "grad_norm": 0.32273320904553404, | |
| "learning_rate": 0.00017019512265021097, | |
| "loss": 1.0388, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 0.3283236994219653, | |
| "grad_norm": 0.31422034130330717, | |
| "learning_rate": 0.00016990710043230406, | |
| "loss": 0.9556, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.32947976878612717, | |
| "grad_norm": 0.31168769800585655, | |
| "learning_rate": 0.00016961793950005558, | |
| "loss": 0.9746, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 0.330635838150289, | |
| "grad_norm": 0.3052533915264079, | |
| "learning_rate": 0.00016932764456359793, | |
| "loss": 0.9542, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.33179190751445087, | |
| "grad_norm": 0.2905803213236271, | |
| "learning_rate": 0.0001690362203515353, | |
| "loss": 0.97, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 0.3329479768786127, | |
| "grad_norm": 0.3350090558904066, | |
| "learning_rate": 0.00016874367161086662, | |
| "loss": 1.0443, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.33410404624277457, | |
| "grad_norm": 0.3060737171676248, | |
| "learning_rate": 0.00016845000310690815, | |
| "loss": 1.043, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 0.3352601156069364, | |
| "grad_norm": 0.3129317388012455, | |
| "learning_rate": 0.00016815521962321604, | |
| "loss": 0.9288, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.33641618497109826, | |
| "grad_norm": 0.3197265364198292, | |
| "learning_rate": 0.00016785932596150827, | |
| "loss": 0.984, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 0.3375722543352601, | |
| "grad_norm": 0.29604487460555096, | |
| "learning_rate": 0.0001675623269415864, | |
| "loss": 1.0016, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.33872832369942196, | |
| "grad_norm": 0.30262162872771553, | |
| "learning_rate": 0.00016726422740125728, | |
| "loss": 0.9856, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 0.3398843930635838, | |
| "grad_norm": 0.3087826779620012, | |
| "learning_rate": 0.000166965032196254, | |
| "loss": 0.9957, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.34104046242774566, | |
| "grad_norm": 0.3024417730810271, | |
| "learning_rate": 0.00016666474620015686, | |
| "loss": 0.939, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 0.3421965317919075, | |
| "grad_norm": 0.3138788978371571, | |
| "learning_rate": 0.0001663633743043141, | |
| "loss": 0.9363, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.34335260115606936, | |
| "grad_norm": 0.30022727252981385, | |
| "learning_rate": 0.0001660609214177621, | |
| "loss": 1.0356, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 0.3445086705202312, | |
| "grad_norm": 0.3554930404416277, | |
| "learning_rate": 0.00016575739246714547, | |
| "loss": 0.9741, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.34566473988439306, | |
| "grad_norm": 0.29207690202141207, | |
| "learning_rate": 0.00016545279239663682, | |
| "loss": 0.9914, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 0.3468208092485549, | |
| "grad_norm": 0.3029574756031942, | |
| "learning_rate": 0.00016514712616785612, | |
| "loss": 0.9421, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.34797687861271676, | |
| "grad_norm": 0.3003131478682294, | |
| "learning_rate": 0.00016484039875979005, | |
| "loss": 0.9536, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 0.3491329479768786, | |
| "grad_norm": 0.31513567646148855, | |
| "learning_rate": 0.00016453261516871068, | |
| "loss": 0.9426, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.35028901734104045, | |
| "grad_norm": 0.2996325149723125, | |
| "learning_rate": 0.00016422378040809437, | |
| "loss": 1.0104, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 0.3514450867052023, | |
| "grad_norm": 0.31000327701469227, | |
| "learning_rate": 0.00016391389950853977, | |
| "loss": 0.9899, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.35260115606936415, | |
| "grad_norm": 0.3077861834938761, | |
| "learning_rate": 0.0001636029775176862, | |
| "loss": 0.9865, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 0.353757225433526, | |
| "grad_norm": 0.3065445301834393, | |
| "learning_rate": 0.00016329101950013122, | |
| "loss": 0.9833, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.35491329479768785, | |
| "grad_norm": 0.3317987511497053, | |
| "learning_rate": 0.00016297803053734816, | |
| "loss": 0.9549, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 0.3560693641618497, | |
| "grad_norm": 0.3196645968421778, | |
| "learning_rate": 0.0001626640157276034, | |
| "loss": 0.9675, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.35722543352601155, | |
| "grad_norm": 0.3134584834564055, | |
| "learning_rate": 0.00016234898018587337, | |
| "loss": 0.9516, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 0.3583815028901734, | |
| "grad_norm": 0.3117834490142485, | |
| "learning_rate": 0.00016203292904376105, | |
| "loss": 0.9846, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.35953757225433525, | |
| "grad_norm": 0.3217830266213364, | |
| "learning_rate": 0.00016171586744941264, | |
| "loss": 0.967, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 0.3606936416184971, | |
| "grad_norm": 0.30934961569422764, | |
| "learning_rate": 0.00016139780056743342, | |
| "loss": 0.9649, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.36184971098265895, | |
| "grad_norm": 0.3292849028870402, | |
| "learning_rate": 0.00016107873357880384, | |
| "loss": 1.0175, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 0.3630057803468208, | |
| "grad_norm": 0.31062236776235347, | |
| "learning_rate": 0.00016075867168079507, | |
| "loss": 0.9696, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.36416184971098264, | |
| "grad_norm": 0.318894494073892, | |
| "learning_rate": 0.00016043762008688433, | |
| "loss": 0.9286, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 0.3653179190751445, | |
| "grad_norm": 0.3029890310755476, | |
| "learning_rate": 0.00016011558402666983, | |
| "loss": 0.9594, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.36647398843930634, | |
| "grad_norm": 0.3081821304700694, | |
| "learning_rate": 0.00015979256874578594, | |
| "loss": 0.987, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 0.3676300578034682, | |
| "grad_norm": 0.29222596605397133, | |
| "learning_rate": 0.00015946857950581734, | |
| "loss": 0.9919, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.36878612716763004, | |
| "grad_norm": 0.2955787813149893, | |
| "learning_rate": 0.0001591436215842135, | |
| "loss": 0.9653, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 0.3699421965317919, | |
| "grad_norm": 0.2991599664341822, | |
| "learning_rate": 0.0001588177002742029, | |
| "loss": 0.9874, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.37109826589595374, | |
| "grad_norm": 0.44600654437638615, | |
| "learning_rate": 0.00015849082088470638, | |
| "loss": 0.9504, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 0.3722543352601156, | |
| "grad_norm": 0.33315550683583905, | |
| "learning_rate": 0.00015816298874025102, | |
| "loss": 1.0328, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.37341040462427744, | |
| "grad_norm": 0.292061245143086, | |
| "learning_rate": 0.00015783420918088337, | |
| "loss": 0.9762, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 0.3745664739884393, | |
| "grad_norm": 0.31976882089395187, | |
| "learning_rate": 0.0001575044875620822, | |
| "loss": 1.026, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.37572254335260113, | |
| "grad_norm": 0.3238783922087859, | |
| "learning_rate": 0.0001571738292546716, | |
| "loss": 0.9496, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 0.376878612716763, | |
| "grad_norm": 0.2959033885199569, | |
| "learning_rate": 0.00015684223964473337, | |
| "loss": 0.9656, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.37803468208092483, | |
| "grad_norm": 0.3043686833702477, | |
| "learning_rate": 0.0001565097241335191, | |
| "loss": 0.954, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 0.3791907514450867, | |
| "grad_norm": 0.3015192978319062, | |
| "learning_rate": 0.00015617628813736247, | |
| "loss": 0.9908, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.38034682080924853, | |
| "grad_norm": 0.3082482576635595, | |
| "learning_rate": 0.00015584193708759094, | |
| "loss": 0.9477, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 0.3815028901734104, | |
| "grad_norm": 0.30690197747994147, | |
| "learning_rate": 0.00015550667643043716, | |
| "loss": 0.9547, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.38265895953757223, | |
| "grad_norm": 0.32095813086552044, | |
| "learning_rate": 0.0001551705116269504, | |
| "loss": 0.9946, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 0.3838150289017341, | |
| "grad_norm": 0.3102088056503803, | |
| "learning_rate": 0.0001548334481529075, | |
| "loss": 0.9755, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.38497109826589593, | |
| "grad_norm": 0.31740424344750273, | |
| "learning_rate": 0.00015449549149872376, | |
| "loss": 0.986, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 0.3861271676300578, | |
| "grad_norm": 0.29760230294640583, | |
| "learning_rate": 0.00015415664716936345, | |
| "loss": 0.9736, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.3872832369942196, | |
| "grad_norm": 0.33696332230509884, | |
| "learning_rate": 0.00015381692068425004, | |
| "loss": 0.9833, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 0.3884393063583815, | |
| "grad_norm": 0.2971442840267895, | |
| "learning_rate": 0.0001534763175771766, | |
| "loss": 0.9787, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.3895953757225434, | |
| "grad_norm": 0.2985513113352265, | |
| "learning_rate": 0.00015313484339621534, | |
| "loss": 0.9586, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 0.39075144508670523, | |
| "grad_norm": 0.3070566835370781, | |
| "learning_rate": 0.00015279250370362735, | |
| "loss": 0.9878, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.3919075144508671, | |
| "grad_norm": 0.3277524257534511, | |
| "learning_rate": 0.00015244930407577205, | |
| "loss": 1.0016, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 0.3930635838150289, | |
| "grad_norm": 0.30050526813256595, | |
| "learning_rate": 0.00015210525010301638, | |
| "loss": 0.9553, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.3942196531791908, | |
| "grad_norm": 0.30810810818757023, | |
| "learning_rate": 0.0001517603473896435, | |
| "loss": 0.9559, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 0.3953757225433526, | |
| "grad_norm": 0.299950087627466, | |
| "learning_rate": 0.00015141460155376182, | |
| "loss": 0.9609, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.3965317919075145, | |
| "grad_norm": 0.2925418247045739, | |
| "learning_rate": 0.00015106801822721338, | |
| "loss": 0.9763, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 0.3976878612716763, | |
| "grad_norm": 0.30196182567060115, | |
| "learning_rate": 0.00015072060305548187, | |
| "loss": 0.9959, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.3988439306358382, | |
| "grad_norm": 0.30989108180452857, | |
| "learning_rate": 0.0001503723616976011, | |
| "loss": 1.0003, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.3039991146331331, | |
| "learning_rate": 0.00015002329982606255, | |
| "loss": 1.0345, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.40115606936416187, | |
| "grad_norm": 0.3211973783178471, | |
| "learning_rate": 0.00014967342312672283, | |
| "loss": 0.9384, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 0.4023121387283237, | |
| "grad_norm": 0.29388068969488124, | |
| "learning_rate": 0.00014932273729871152, | |
| "loss": 0.9051, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.40346820809248557, | |
| "grad_norm": 0.3265810444042218, | |
| "learning_rate": 0.0001489712480543379, | |
| "loss": 0.9835, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 0.4046242774566474, | |
| "grad_norm": 0.303317950576793, | |
| "learning_rate": 0.0001486189611189981, | |
| "loss": 0.9446, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.40578034682080927, | |
| "grad_norm": 0.2967360524243329, | |
| "learning_rate": 0.00014826588223108185, | |
| "loss": 0.9908, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 0.4069364161849711, | |
| "grad_norm": 0.30908569033672595, | |
| "learning_rate": 0.00014791201714187897, | |
| "loss": 0.9118, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.40809248554913297, | |
| "grad_norm": 0.2921540136185523, | |
| "learning_rate": 0.0001475573716154856, | |
| "loss": 1.0177, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 0.4092485549132948, | |
| "grad_norm": 0.30891513255558445, | |
| "learning_rate": 0.00014720195142871054, | |
| "loss": 0.9528, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.41040462427745666, | |
| "grad_norm": 0.3116110416347837, | |
| "learning_rate": 0.00014684576237098082, | |
| "loss": 1.0153, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 0.4115606936416185, | |
| "grad_norm": 0.29853304778547163, | |
| "learning_rate": 0.00014648881024424774, | |
| "loss": 0.9607, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.41271676300578036, | |
| "grad_norm": 0.29621019258375375, | |
| "learning_rate": 0.00014613110086289218, | |
| "loss": 1.0178, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 0.4138728323699422, | |
| "grad_norm": 0.32202847876721696, | |
| "learning_rate": 0.00014577264005362985, | |
| "loss": 0.9274, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.41502890173410406, | |
| "grad_norm": 0.30619120168198916, | |
| "learning_rate": 0.00014541343365541645, | |
| "loss": 0.9435, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 0.4161849710982659, | |
| "grad_norm": 0.3001554889949122, | |
| "learning_rate": 0.00014505348751935263, | |
| "loss": 0.9738, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.41734104046242776, | |
| "grad_norm": 0.3065569332725715, | |
| "learning_rate": 0.00014469280750858854, | |
| "loss": 0.9627, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 0.4184971098265896, | |
| "grad_norm": 0.3346803326675102, | |
| "learning_rate": 0.00014433139949822837, | |
| "loss": 1.0008, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.41965317919075146, | |
| "grad_norm": 0.30285946590074797, | |
| "learning_rate": 0.00014396926937523477, | |
| "loss": 0.9681, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 0.4208092485549133, | |
| "grad_norm": 0.323926429665197, | |
| "learning_rate": 0.0001436064230383327, | |
| "loss": 0.9883, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.42196531791907516, | |
| "grad_norm": 0.30822208654391275, | |
| "learning_rate": 0.00014324286639791367, | |
| "loss": 0.9471, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 0.423121387283237, | |
| "grad_norm": 0.3043728994137006, | |
| "learning_rate": 0.00014287860537593917, | |
| "loss": 0.9837, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.42427745664739885, | |
| "grad_norm": 0.3042147218697011, | |
| "learning_rate": 0.00014251364590584444, | |
| "loss": 0.9576, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 0.4254335260115607, | |
| "grad_norm": 0.32773524321463, | |
| "learning_rate": 0.00014214799393244166, | |
| "loss": 0.9356, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.42658959537572255, | |
| "grad_norm": 0.30439292824288355, | |
| "learning_rate": 0.00014178165541182312, | |
| "loss": 0.9421, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 0.4277456647398844, | |
| "grad_norm": 0.321382456218326, | |
| "learning_rate": 0.00014141463631126442, | |
| "loss": 0.9515, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.42890173410404625, | |
| "grad_norm": 0.3130786437336031, | |
| "learning_rate": 0.0001410469426091269, | |
| "loss": 0.9715, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 0.4300578034682081, | |
| "grad_norm": 0.3135399317680074, | |
| "learning_rate": 0.00014067858029476063, | |
| "loss": 0.9474, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.43121387283236995, | |
| "grad_norm": 0.31368812931362966, | |
| "learning_rate": 0.00014030955536840656, | |
| "loss": 1.0225, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 0.4323699421965318, | |
| "grad_norm": 0.3332336708705887, | |
| "learning_rate": 0.00013993987384109898, | |
| "loss": 1.0098, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.43352601156069365, | |
| "grad_norm": 0.2950594206550405, | |
| "learning_rate": 0.00013956954173456747, | |
| "loss": 0.9846, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 0.4346820809248555, | |
| "grad_norm": 0.3001574947847011, | |
| "learning_rate": 0.000139198565081139, | |
| "loss": 0.9853, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.43583815028901735, | |
| "grad_norm": 0.29311121484460284, | |
| "learning_rate": 0.00013882694992363936, | |
| "loss": 1.0175, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 0.4369942196531792, | |
| "grad_norm": 0.3286518580874968, | |
| "learning_rate": 0.00013845470231529502, | |
| "loss": 0.9845, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.43815028901734104, | |
| "grad_norm": 0.31669776510548286, | |
| "learning_rate": 0.00013808182831963442, | |
| "loss": 1.0096, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 0.4393063583815029, | |
| "grad_norm": 0.3230500170692119, | |
| "learning_rate": 0.00013770833401038912, | |
| "loss": 0.9652, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.44046242774566474, | |
| "grad_norm": 0.314774115101565, | |
| "learning_rate": 0.0001373342254713951, | |
| "loss": 0.9884, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 0.4416184971098266, | |
| "grad_norm": 0.309673258526753, | |
| "learning_rate": 0.00013695950879649338, | |
| "loss": 0.9617, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.44277456647398844, | |
| "grad_norm": 0.31688401964004287, | |
| "learning_rate": 0.00013658419008943088, | |
| "loss": 1.0007, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 0.4439306358381503, | |
| "grad_norm": 0.3115831394799577, | |
| "learning_rate": 0.00013620827546376112, | |
| "loss": 0.9837, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.44508670520231214, | |
| "grad_norm": 0.3134574472279371, | |
| "learning_rate": 0.00013583177104274435, | |
| "loss": 0.9748, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 0.446242774566474, | |
| "grad_norm": 0.3388852510725773, | |
| "learning_rate": 0.00013545468295924812, | |
| "loss": 0.9825, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.44739884393063584, | |
| "grad_norm": 0.33068272114069625, | |
| "learning_rate": 0.00013507701735564716, | |
| "loss": 0.9552, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 0.4485549132947977, | |
| "grad_norm": 0.3070498417777377, | |
| "learning_rate": 0.00013469878038372348, | |
| "loss": 0.9842, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.44971098265895953, | |
| "grad_norm": 0.32088136083821, | |
| "learning_rate": 0.00013431997820456592, | |
| "loss": 0.9635, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 0.4508670520231214, | |
| "grad_norm": 0.317988833081837, | |
| "learning_rate": 0.00013394061698847022, | |
| "loss": 0.9922, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.45202312138728323, | |
| "grad_norm": 0.29873515381181037, | |
| "learning_rate": 0.000133560702914838, | |
| "loss": 0.9808, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 0.4531791907514451, | |
| "grad_norm": 0.334307242594275, | |
| "learning_rate": 0.00013318024217207652, | |
| "loss": 0.9285, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.45433526011560693, | |
| "grad_norm": 0.3161167520009514, | |
| "learning_rate": 0.00013279924095749768, | |
| "loss": 0.9721, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 0.4554913294797688, | |
| "grad_norm": 0.3219859959358273, | |
| "learning_rate": 0.00013241770547721703, | |
| "loss": 1.008, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.45664739884393063, | |
| "grad_norm": 0.27935961874420406, | |
| "learning_rate": 0.00013203564194605284, | |
| "loss": 0.9502, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 0.4578034682080925, | |
| "grad_norm": 0.29913347538052254, | |
| "learning_rate": 0.0001316530565874248, | |
| "loss": 0.9791, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.45895953757225433, | |
| "grad_norm": 0.30294269390330414, | |
| "learning_rate": 0.00013126995563325254, | |
| "loss": 0.9763, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 0.4601156069364162, | |
| "grad_norm": 0.32096224239126736, | |
| "learning_rate": 0.00013088634532385424, | |
| "loss": 0.9238, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.461271676300578, | |
| "grad_norm": 0.2960586714061201, | |
| "learning_rate": 0.000130502231907845, | |
| "loss": 0.9533, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 0.4624277456647399, | |
| "grad_norm": 0.29803737045431256, | |
| "learning_rate": 0.000130117621642035, | |
| "loss": 0.9526, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.4635838150289017, | |
| "grad_norm": 0.31720967984437226, | |
| "learning_rate": 0.00012973252079132749, | |
| "loss": 0.9566, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 0.4647398843930636, | |
| "grad_norm": 0.31204560106706253, | |
| "learning_rate": 0.00012934693562861692, | |
| "loss": 0.9821, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.4658959537572254, | |
| "grad_norm": 0.3452586478382497, | |
| "learning_rate": 0.00012896087243468673, | |
| "loss": 0.9866, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 0.46705202312138727, | |
| "grad_norm": 0.30419391343270963, | |
| "learning_rate": 0.00012857433749810691, | |
| "loss": 0.9465, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.4682080924855491, | |
| "grad_norm": 0.302216494494177, | |
| "learning_rate": 0.00012818733711513164, | |
| "loss": 0.9928, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 0.46936416184971097, | |
| "grad_norm": 0.29660145267520094, | |
| "learning_rate": 0.00012779987758959683, | |
| "loss": 0.9714, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.4705202312138728, | |
| "grad_norm": 0.3375993332751583, | |
| "learning_rate": 0.00012741196523281728, | |
| "loss": 1.004, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 0.47167630057803467, | |
| "grad_norm": 0.31685124172490736, | |
| "learning_rate": 0.0001270236063634839, | |
| "loss": 0.9686, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.4728323699421965, | |
| "grad_norm": 0.30517277761996336, | |
| "learning_rate": 0.00012663480730756095, | |
| "loss": 0.97, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 0.47398843930635837, | |
| "grad_norm": 0.3075134986191579, | |
| "learning_rate": 0.00012624557439818275, | |
| "loss": 0.9535, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.4751445086705202, | |
| "grad_norm": 0.2914116111037525, | |
| "learning_rate": 0.00012585591397555078, | |
| "loss": 0.9549, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 0.47630057803468207, | |
| "grad_norm": 0.3065733883077486, | |
| "learning_rate": 0.00012546583238683015, | |
| "loss": 0.9694, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.4774566473988439, | |
| "grad_norm": 0.30076466916700556, | |
| "learning_rate": 0.00012507533598604632, | |
| "loss": 0.9802, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 0.47861271676300576, | |
| "grad_norm": 0.29670240314259055, | |
| "learning_rate": 0.00012468443113398175, | |
| "loss": 0.9366, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.4797687861271676, | |
| "grad_norm": 0.31183074363125884, | |
| "learning_rate": 0.00012429312419807198, | |
| "loss": 0.966, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 0.48092485549132946, | |
| "grad_norm": 0.31278790481596425, | |
| "learning_rate": 0.00012390142155230217, | |
| "loss": 0.9893, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.4820809248554913, | |
| "grad_norm": 0.28207826631174193, | |
| "learning_rate": 0.0001235093295771032, | |
| "loss": 0.9472, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 0.48323699421965316, | |
| "grad_norm": 0.32635523820738965, | |
| "learning_rate": 0.00012311685465924774, | |
| "loss": 0.9089, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.484393063583815, | |
| "grad_norm": 0.2977916871662523, | |
| "learning_rate": 0.00012272400319174607, | |
| "loss": 0.9834, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 0.48554913294797686, | |
| "grad_norm": 0.31990291106992935, | |
| "learning_rate": 0.00012233078157374217, | |
| "loss": 0.9312, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.4867052023121387, | |
| "grad_norm": 0.33073575866363214, | |
| "learning_rate": 0.00012193719621040942, | |
| "loss": 0.9795, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 0.48786127167630056, | |
| "grad_norm": 0.2915838062263623, | |
| "learning_rate": 0.00012154325351284618, | |
| "loss": 0.9789, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.4890173410404624, | |
| "grad_norm": 0.33822747094942934, | |
| "learning_rate": 0.00012114895989797144, | |
| "loss": 0.9304, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 0.49017341040462425, | |
| "grad_norm": 0.32370602693562334, | |
| "learning_rate": 0.00012075432178842021, | |
| "loss": 0.9428, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.4913294797687861, | |
| "grad_norm": 0.32375527811459415, | |
| "learning_rate": 0.00012035934561243905, | |
| "loss": 0.9718, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 0.49248554913294795, | |
| "grad_norm": 0.2943351586407433, | |
| "learning_rate": 0.00011996403780378123, | |
| "loss": 0.9712, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.4936416184971098, | |
| "grad_norm": 0.3223304229208655, | |
| "learning_rate": 0.00011956840480160194, | |
| "loss": 1.0046, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 0.49479768786127165, | |
| "grad_norm": 0.3032804365412004, | |
| "learning_rate": 0.00011917245305035354, | |
| "loss": 0.9596, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.4959537572254335, | |
| "grad_norm": 0.3086608217360584, | |
| "learning_rate": 0.00011877618899968037, | |
| "loss": 0.9473, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 0.49710982658959535, | |
| "grad_norm": 0.3131187277547376, | |
| "learning_rate": 0.00011837961910431383, | |
| "loss": 1.0065, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.4982658959537572, | |
| "grad_norm": 0.2975605931413944, | |
| "learning_rate": 0.00011798274982396726, | |
| "loss": 0.9481, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 0.49942196531791905, | |
| "grad_norm": 0.29498275298057963, | |
| "learning_rate": 0.00011758558762323067, | |
| "loss": 0.9884, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.500578034682081, | |
| "grad_norm": 0.2906355713880922, | |
| "learning_rate": 0.00011718813897146535, | |
| "loss": 0.9643, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 0.5017341040462427, | |
| "grad_norm": 0.2879177716708955, | |
| "learning_rate": 0.00011679041034269869, | |
| "loss": 0.9496, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.5028901734104047, | |
| "grad_norm": 0.3107002014222183, | |
| "learning_rate": 0.00011639240821551858, | |
| "loss": 0.9489, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 0.5040462427745664, | |
| "grad_norm": 0.30854297451303886, | |
| "learning_rate": 0.00011599413907296785, | |
| "loss": 0.9887, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.5052023121387283, | |
| "grad_norm": 0.3250596343211611, | |
| "learning_rate": 0.00011559560940243888, | |
| "loss": 0.9421, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 0.5063583815028901, | |
| "grad_norm": 0.303124134082483, | |
| "learning_rate": 0.00011519682569556758, | |
| "loss": 0.967, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.507514450867052, | |
| "grad_norm": 0.29292499319855175, | |
| "learning_rate": 0.00011479779444812808, | |
| "loss": 0.9679, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 0.5086705202312138, | |
| "grad_norm": 0.30291810874235703, | |
| "learning_rate": 0.00011439852215992647, | |
| "loss": 0.997, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.5098265895953757, | |
| "grad_norm": 0.3234308605182878, | |
| "learning_rate": 0.0001139990153346953, | |
| "loss": 0.9876, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 0.5109826589595375, | |
| "grad_norm": 0.3137214941805028, | |
| "learning_rate": 0.00011359928047998744, | |
| "loss": 1.0407, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.5121387283236994, | |
| "grad_norm": 0.32171251618436913, | |
| "learning_rate": 0.0001131993241070701, | |
| "loss": 0.9783, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 0.5132947976878612, | |
| "grad_norm": 0.2971713793214721, | |
| "learning_rate": 0.00011279915273081876, | |
| "loss": 0.9678, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.5144508670520231, | |
| "grad_norm": 0.30876666041737444, | |
| "learning_rate": 0.00011239877286961122, | |
| "loss": 0.9717, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 0.5156069364161849, | |
| "grad_norm": 0.31611543232380335, | |
| "learning_rate": 0.00011199819104522114, | |
| "loss": 0.9611, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.5167630057803468, | |
| "grad_norm": 0.31365800007794736, | |
| "learning_rate": 0.000111597413782712, | |
| "loss": 0.986, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 0.5179190751445086, | |
| "grad_norm": 0.3387876838248837, | |
| "learning_rate": 0.00011119644761033078, | |
| "loss": 0.9865, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.5190751445086705, | |
| "grad_norm": 0.3090392049931908, | |
| "learning_rate": 0.00011079529905940163, | |
| "loss": 0.9264, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 0.5202312138728323, | |
| "grad_norm": 0.30547601371038785, | |
| "learning_rate": 0.0001103939746642194, | |
| "loss": 0.9293, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.5213872832369942, | |
| "grad_norm": 0.30920860300711217, | |
| "learning_rate": 0.00010999248096194326, | |
| "loss": 0.9759, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 0.522543352601156, | |
| "grad_norm": 0.30207153503695156, | |
| "learning_rate": 0.00010959082449249026, | |
| "loss": 0.9557, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.5236994219653179, | |
| "grad_norm": 0.29504681849985004, | |
| "learning_rate": 0.00010918901179842877, | |
| "loss": 0.9686, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 0.5248554913294797, | |
| "grad_norm": 0.29286119267320176, | |
| "learning_rate": 0.00010878704942487183, | |
| "loss": 1.0042, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.5260115606936416, | |
| "grad_norm": 0.3062243965654378, | |
| "learning_rate": 0.00010838494391937064, | |
| "loss": 0.9784, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 0.5271676300578034, | |
| "grad_norm": 0.3079900592590067, | |
| "learning_rate": 0.00010798270183180794, | |
| "loss": 0.9503, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.5283236994219653, | |
| "grad_norm": 0.3154259872984876, | |
| "learning_rate": 0.0001075803297142911, | |
| "loss": 0.9509, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 0.5294797687861271, | |
| "grad_norm": 0.29229323286742326, | |
| "learning_rate": 0.00010717783412104568, | |
| "loss": 0.9557, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.530635838150289, | |
| "grad_norm": 0.31601404726519217, | |
| "learning_rate": 0.00010677522160830848, | |
| "loss": 0.9042, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 0.5317919075144508, | |
| "grad_norm": 0.28505318471127816, | |
| "learning_rate": 0.00010637249873422077, | |
| "loss": 0.9692, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.5329479768786127, | |
| "grad_norm": 0.32485305181812835, | |
| "learning_rate": 0.00010596967205872154, | |
| "loss": 1.0065, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 0.5341040462427745, | |
| "grad_norm": 0.3078039739362063, | |
| "learning_rate": 0.00010556674814344059, | |
| "loss": 0.9284, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.5352601156069364, | |
| "grad_norm": 0.3208549342399588, | |
| "learning_rate": 0.00010516373355159159, | |
| "loss": 0.9477, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 0.5364161849710982, | |
| "grad_norm": 0.30748525424656054, | |
| "learning_rate": 0.00010476063484786535, | |
| "loss": 0.9629, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.5375722543352601, | |
| "grad_norm": 0.35275688331568944, | |
| "learning_rate": 0.0001043574585983227, | |
| "loss": 1.0113, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 0.5387283236994219, | |
| "grad_norm": 0.3253346439794717, | |
| "learning_rate": 0.00010395421137028761, | |
| "loss": 1.0346, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.5398843930635838, | |
| "grad_norm": 0.3095697055293057, | |
| "learning_rate": 0.00010355089973224026, | |
| "loss": 0.9546, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 0.5410404624277456, | |
| "grad_norm": 0.307599147895419, | |
| "learning_rate": 0.00010314753025370991, | |
| "loss": 0.9836, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.5421965317919075, | |
| "grad_norm": 0.31630009518917934, | |
| "learning_rate": 0.00010274410950516815, | |
| "loss": 1.0071, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 0.5433526011560693, | |
| "grad_norm": 0.2834312244927633, | |
| "learning_rate": 0.00010234064405792154, | |
| "loss": 0.9489, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.5445086705202312, | |
| "grad_norm": 0.2970129068571585, | |
| "learning_rate": 0.0001019371404840048, | |
| "loss": 0.9351, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 0.545664739884393, | |
| "grad_norm": 0.3025660920447834, | |
| "learning_rate": 0.0001015336053560737, | |
| "loss": 0.99, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.5468208092485549, | |
| "grad_norm": 0.30633618369872967, | |
| "learning_rate": 0.00010113004524729799, | |
| "loss": 0.954, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 0.5479768786127167, | |
| "grad_norm": 0.28629344395996326, | |
| "learning_rate": 0.00010072646673125432, | |
| "loss": 0.9783, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.5491329479768786, | |
| "grad_norm": 0.3097159968894246, | |
| "learning_rate": 0.00010032287638181919, | |
| "loss": 1.0303, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 0.5502890173410404, | |
| "grad_norm": 0.30322928513172714, | |
| "learning_rate": 9.991928077306183e-05, | |
| "loss": 1.0, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.5514450867052023, | |
| "grad_norm": 0.4644501347952281, | |
| "learning_rate": 9.951568647913718e-05, | |
| "loss": 0.9294, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 0.5526011560693641, | |
| "grad_norm": 0.3124185608435605, | |
| "learning_rate": 9.911210007417869e-05, | |
| "loss": 0.9847, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.553757225433526, | |
| "grad_norm": 0.32991063517383223, | |
| "learning_rate": 9.870852813219143e-05, | |
| "loss": 0.9755, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 0.5549132947976878, | |
| "grad_norm": 0.31527757887354024, | |
| "learning_rate": 9.830497722694478e-05, | |
| "loss": 0.9819, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.5560693641618497, | |
| "grad_norm": 0.3094199650313883, | |
| "learning_rate": 9.790145393186541e-05, | |
| "loss": 0.9409, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 0.5572254335260116, | |
| "grad_norm": 0.3055144703983288, | |
| "learning_rate": 9.749796481993042e-05, | |
| "loss": 0.9674, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.5583815028901734, | |
| "grad_norm": 0.31607857296755776, | |
| "learning_rate": 9.709451646355996e-05, | |
| "loss": 1.0174, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 0.5595375722543353, | |
| "grad_norm": 0.2975421743209585, | |
| "learning_rate": 9.669111543451033e-05, | |
| "loss": 0.9683, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.5606936416184971, | |
| "grad_norm": 0.30435556233851097, | |
| "learning_rate": 9.628776830376698e-05, | |
| "loss": 0.9539, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 0.561849710982659, | |
| "grad_norm": 0.31164699307502913, | |
| "learning_rate": 9.588448164143739e-05, | |
| "loss": 0.9716, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.5630057803468208, | |
| "grad_norm": 0.3029837051614153, | |
| "learning_rate": 9.548126201664398e-05, | |
| "loss": 0.9488, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 0.5641618497109827, | |
| "grad_norm": 0.30375216989550236, | |
| "learning_rate": 9.507811599741735e-05, | |
| "loss": 0.9241, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.5653179190751445, | |
| "grad_norm": 0.3044100811963364, | |
| "learning_rate": 9.467505015058901e-05, | |
| "loss": 1.0212, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 0.5664739884393064, | |
| "grad_norm": 0.3079385379059537, | |
| "learning_rate": 9.427207104168467e-05, | |
| "loss": 0.9341, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.5676300578034682, | |
| "grad_norm": 0.3012241533223519, | |
| "learning_rate": 9.386918523481709e-05, | |
| "loss": 0.9533, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 0.5687861271676301, | |
| "grad_norm": 0.32392283471513106, | |
| "learning_rate": 9.346639929257916e-05, | |
| "loss": 0.9888, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.5699421965317919, | |
| "grad_norm": 0.28835183242553575, | |
| "learning_rate": 9.306371977593726e-05, | |
| "loss": 0.9847, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 0.5710982658959538, | |
| "grad_norm": 0.29592049768177997, | |
| "learning_rate": 9.26611532441241e-05, | |
| "loss": 0.9075, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.5722543352601156, | |
| "grad_norm": 0.30510640464990846, | |
| "learning_rate": 9.225870625453192e-05, | |
| "loss": 0.9276, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 0.5734104046242775, | |
| "grad_norm": 0.32217649840985024, | |
| "learning_rate": 9.18563853626059e-05, | |
| "loss": 0.9936, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.5745664739884393, | |
| "grad_norm": 0.2945778811908368, | |
| "learning_rate": 9.145419712173713e-05, | |
| "loss": 0.9775, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 0.5757225433526012, | |
| "grad_norm": 0.2937161972776823, | |
| "learning_rate": 9.105214808315588e-05, | |
| "loss": 0.9293, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.576878612716763, | |
| "grad_norm": 0.34400282833354867, | |
| "learning_rate": 9.065024479582513e-05, | |
| "loss": 1.0045, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 0.5780346820809249, | |
| "grad_norm": 0.30233509314710183, | |
| "learning_rate": 9.024849380633359e-05, | |
| "loss": 0.9786, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.5791907514450867, | |
| "grad_norm": 0.29454389871868664, | |
| "learning_rate": 8.984690165878921e-05, | |
| "loss": 0.9584, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 0.5803468208092486, | |
| "grad_norm": 0.30071228762049557, | |
| "learning_rate": 8.944547489471265e-05, | |
| "loss": 0.955, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.5815028901734104, | |
| "grad_norm": 0.3169803125653554, | |
| "learning_rate": 8.904422005293052e-05, | |
| "loss": 1.0198, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 0.5826589595375723, | |
| "grad_norm": 0.31831872317303483, | |
| "learning_rate": 8.864314366946913e-05, | |
| "loss": 0.9781, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.5838150289017341, | |
| "grad_norm": 0.30646509617401063, | |
| "learning_rate": 8.824225227744782e-05, | |
| "loss": 0.9556, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 0.584971098265896, | |
| "grad_norm": 0.28426966152836436, | |
| "learning_rate": 8.784155240697254e-05, | |
| "loss": 0.9811, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.5861271676300578, | |
| "grad_norm": 0.29076631427987554, | |
| "learning_rate": 8.74410505850297e-05, | |
| "loss": 0.9653, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 0.5872832369942197, | |
| "grad_norm": 0.2993968457362386, | |
| "learning_rate": 8.704075333537963e-05, | |
| "loss": 0.9267, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.5884393063583815, | |
| "grad_norm": 0.28060562517633875, | |
| "learning_rate": 8.66406671784503e-05, | |
| "loss": 0.9767, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 0.5895953757225434, | |
| "grad_norm": 0.29394163379907895, | |
| "learning_rate": 8.624079863123135e-05, | |
| "loss": 0.9692, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.5907514450867052, | |
| "grad_norm": 0.29752047867631554, | |
| "learning_rate": 8.584115420716777e-05, | |
| "loss": 1.0218, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 0.5919075144508671, | |
| "grad_norm": 0.2960489971126818, | |
| "learning_rate": 8.544174041605363e-05, | |
| "loss": 0.9386, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.5930635838150289, | |
| "grad_norm": 0.2991003815095411, | |
| "learning_rate": 8.504256376392647e-05, | |
| "loss": 0.951, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 0.5942196531791908, | |
| "grad_norm": 0.2875329275503883, | |
| "learning_rate": 8.464363075296095e-05, | |
| "loss": 0.9595, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.5953757225433526, | |
| "grad_norm": 0.29629340379738117, | |
| "learning_rate": 8.424494788136303e-05, | |
| "loss": 0.946, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 0.5965317919075145, | |
| "grad_norm": 0.3039710490176798, | |
| "learning_rate": 8.384652164326432e-05, | |
| "loss": 0.9297, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.5976878612716763, | |
| "grad_norm": 0.2867197601664578, | |
| "learning_rate": 8.344835852861595e-05, | |
| "loss": 0.9655, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 0.5988439306358382, | |
| "grad_norm": 0.3021141932586307, | |
| "learning_rate": 8.305046502308319e-05, | |
| "loss": 0.9388, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 0.32608595836134247, | |
| "learning_rate": 8.265284760793957e-05, | |
| "loss": 0.948, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 0.6011560693641619, | |
| "grad_norm": 0.3077259465134535, | |
| "learning_rate": 8.225551275996138e-05, | |
| "loss": 1.0123, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.6023121387283237, | |
| "grad_norm": 0.28311149702035393, | |
| "learning_rate": 8.185846695132227e-05, | |
| "loss": 0.9456, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 0.6034682080924856, | |
| "grad_norm": 0.31276946079278556, | |
| "learning_rate": 8.146171664948769e-05, | |
| "loss": 0.9755, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.6046242774566474, | |
| "grad_norm": 0.3264513747288462, | |
| "learning_rate": 8.10652683171095e-05, | |
| "loss": 0.9619, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 0.6057803468208093, | |
| "grad_norm": 0.30285428517053464, | |
| "learning_rate": 8.066912841192099e-05, | |
| "loss": 0.9344, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.6069364161849711, | |
| "grad_norm": 0.3644501014383341, | |
| "learning_rate": 8.027330338663132e-05, | |
| "loss": 0.9794, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 0.608092485549133, | |
| "grad_norm": 0.28904372054279964, | |
| "learning_rate": 7.987779968882061e-05, | |
| "loss": 0.941, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.6092485549132948, | |
| "grad_norm": 0.2895705859655337, | |
| "learning_rate": 7.9482623760835e-05, | |
| "loss": 0.951, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 0.6104046242774567, | |
| "grad_norm": 0.29749813183970186, | |
| "learning_rate": 7.908778203968146e-05, | |
| "loss": 0.9244, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.6115606936416185, | |
| "grad_norm": 0.31971185227728377, | |
| "learning_rate": 7.869328095692312e-05, | |
| "loss": 0.9645, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 0.6127167630057804, | |
| "grad_norm": 0.2873929264676909, | |
| "learning_rate": 7.829912693857454e-05, | |
| "loss": 0.9739, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.6138728323699422, | |
| "grad_norm": 0.2962385037449908, | |
| "learning_rate": 7.79053264049968e-05, | |
| "loss": 1.0025, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 0.6150289017341041, | |
| "grad_norm": 0.2889472954071748, | |
| "learning_rate": 7.751188577079327e-05, | |
| "loss": 0.9764, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.6161849710982659, | |
| "grad_norm": 0.2914115227852295, | |
| "learning_rate": 7.711881144470481e-05, | |
| "loss": 0.9575, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 0.6173410404624278, | |
| "grad_norm": 0.31913799827615985, | |
| "learning_rate": 7.672610982950546e-05, | |
| "loss": 0.9684, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.6184971098265896, | |
| "grad_norm": 0.3005365331976947, | |
| "learning_rate": 7.633378732189833e-05, | |
| "loss": 0.941, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 0.6196531791907515, | |
| "grad_norm": 0.29855395413135466, | |
| "learning_rate": 7.594185031241115e-05, | |
| "loss": 0.9504, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.6208092485549133, | |
| "grad_norm": 0.30079166608986607, | |
| "learning_rate": 7.555030518529227e-05, | |
| "loss": 0.9489, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 0.6219653179190752, | |
| "grad_norm": 0.2835216761875895, | |
| "learning_rate": 7.515915831840682e-05, | |
| "loss": 1.0283, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.623121387283237, | |
| "grad_norm": 0.296763030995749, | |
| "learning_rate": 7.476841608313253e-05, | |
| "loss": 0.9495, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 0.6242774566473989, | |
| "grad_norm": 0.2965441757959662, | |
| "learning_rate": 7.437808484425614e-05, | |
| "loss": 0.9207, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.6254335260115607, | |
| "grad_norm": 0.28563226791333124, | |
| "learning_rate": 7.398817095986978e-05, | |
| "loss": 0.9529, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 0.6265895953757226, | |
| "grad_norm": 0.2958876716229884, | |
| "learning_rate": 7.359868078126714e-05, | |
| "loss": 0.9415, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.6277456647398844, | |
| "grad_norm": 0.29947263183760775, | |
| "learning_rate": 7.320962065284032e-05, | |
| "loss": 0.9153, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 0.6289017341040463, | |
| "grad_norm": 0.2921294166805471, | |
| "learning_rate": 7.282099691197632e-05, | |
| "loss": 1.0061, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.630057803468208, | |
| "grad_norm": 0.3065806154787742, | |
| "learning_rate": 7.243281588895374e-05, | |
| "loss": 0.9713, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 0.63121387283237, | |
| "grad_norm": 0.29670093707685785, | |
| "learning_rate": 7.204508390683991e-05, | |
| "loss": 0.9152, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.6323699421965318, | |
| "grad_norm": 0.28334644508416545, | |
| "learning_rate": 7.165780728138769e-05, | |
| "loss": 0.9216, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 0.6335260115606937, | |
| "grad_norm": 0.2983458223496864, | |
| "learning_rate": 7.127099232093252e-05, | |
| "loss": 0.9684, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.6346820809248555, | |
| "grad_norm": 0.297707639389454, | |
| "learning_rate": 7.08846453262901e-05, | |
| "loss": 0.9677, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 0.6358381502890174, | |
| "grad_norm": 0.3030445642630547, | |
| "learning_rate": 7.049877259065312e-05, | |
| "loss": 0.991, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.6369942196531792, | |
| "grad_norm": 0.2967929219140181, | |
| "learning_rate": 7.011338039948925e-05, | |
| "loss": 0.9331, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 0.638150289017341, | |
| "grad_norm": 0.30678542010472665, | |
| "learning_rate": 6.972847503043864e-05, | |
| "loss": 0.9952, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.6393063583815028, | |
| "grad_norm": 0.2869522581890146, | |
| "learning_rate": 6.934406275321147e-05, | |
| "loss": 0.9817, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 0.6404624277456648, | |
| "grad_norm": 0.30227078806254565, | |
| "learning_rate": 6.896014982948602e-05, | |
| "loss": 0.9713, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.6416184971098265, | |
| "grad_norm": 0.29642633769940174, | |
| "learning_rate": 6.857674251280671e-05, | |
| "loss": 0.9495, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 0.6427745664739885, | |
| "grad_norm": 0.3324961966488245, | |
| "learning_rate": 6.819384704848199e-05, | |
| "loss": 1.0679, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.6439306358381502, | |
| "grad_norm": 0.2888690447316259, | |
| "learning_rate": 6.781146967348284e-05, | |
| "loss": 0.941, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 0.6450867052023121, | |
| "grad_norm": 0.29579888471073373, | |
| "learning_rate": 6.742961661634115e-05, | |
| "loss": 1.0323, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.6462427745664739, | |
| "grad_norm": 0.2989300466555907, | |
| "learning_rate": 6.704829409704809e-05, | |
| "loss": 0.9723, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 0.6473988439306358, | |
| "grad_norm": 0.27150342274150757, | |
| "learning_rate": 6.666750832695306e-05, | |
| "loss": 0.9397, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.6485549132947976, | |
| "grad_norm": 0.29998837345134693, | |
| "learning_rate": 6.628726550866227e-05, | |
| "loss": 0.9835, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 0.6497109826589595, | |
| "grad_norm": 0.28399979161419353, | |
| "learning_rate": 6.59075718359378e-05, | |
| "loss": 0.9509, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.6508670520231213, | |
| "grad_norm": 0.30370562141401386, | |
| "learning_rate": 6.552843349359688e-05, | |
| "loss": 0.967, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 0.6520231213872832, | |
| "grad_norm": 0.3363515271195273, | |
| "learning_rate": 6.514985665741073e-05, | |
| "loss": 1.0054, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.653179190751445, | |
| "grad_norm": 0.2941850764424752, | |
| "learning_rate": 6.477184749400438e-05, | |
| "loss": 0.9958, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 0.6543352601156069, | |
| "grad_norm": 0.2845955247940358, | |
| "learning_rate": 6.439441216075605e-05, | |
| "loss": 0.9199, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.6554913294797687, | |
| "grad_norm": 0.30264607648319775, | |
| "learning_rate": 6.401755680569683e-05, | |
| "loss": 0.9974, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 0.6566473988439306, | |
| "grad_norm": 0.3049902419448789, | |
| "learning_rate": 6.36412875674105e-05, | |
| "loss": 0.9412, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.6578034682080924, | |
| "grad_norm": 0.28427694016822674, | |
| "learning_rate": 6.326561057493376e-05, | |
| "loss": 0.9618, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 0.6589595375722543, | |
| "grad_norm": 0.2923557343391648, | |
| "learning_rate": 6.28905319476561e-05, | |
| "loss": 0.9738, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.6601156069364161, | |
| "grad_norm": 0.3011922931000182, | |
| "learning_rate": 6.251605779522032e-05, | |
| "loss": 0.9547, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 0.661271676300578, | |
| "grad_norm": 0.2778141274186397, | |
| "learning_rate": 6.214219421742295e-05, | |
| "loss": 1.0062, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.6624277456647398, | |
| "grad_norm": 0.28425833137988876, | |
| "learning_rate": 6.176894730411483e-05, | |
| "loss": 0.9529, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 0.6635838150289017, | |
| "grad_norm": 0.2914968347637112, | |
| "learning_rate": 6.139632313510212e-05, | |
| "loss": 0.944, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.6647398843930635, | |
| "grad_norm": 0.2910244571300189, | |
| "learning_rate": 6.1024327780046944e-05, | |
| "loss": 1.0063, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 0.6658959537572254, | |
| "grad_norm": 0.28730856037506486, | |
| "learning_rate": 6.065296729836879e-05, | |
| "loss": 0.978, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.6670520231213873, | |
| "grad_norm": 0.26692983480937005, | |
| "learning_rate": 6.028224773914575e-05, | |
| "loss": 0.953, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 0.6682080924855491, | |
| "grad_norm": 0.28871757324400055, | |
| "learning_rate": 5.991217514101586e-05, | |
| "loss": 0.9275, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.669364161849711, | |
| "grad_norm": 0.29493246673699086, | |
| "learning_rate": 5.9542755532078856e-05, | |
| "loss": 1.0001, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 0.6705202312138728, | |
| "grad_norm": 0.3100505532341507, | |
| "learning_rate": 5.917399492979805e-05, | |
| "loss": 0.9716, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.6716763005780347, | |
| "grad_norm": 0.296746629997667, | |
| "learning_rate": 5.880589934090206e-05, | |
| "loss": 0.987, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 0.6728323699421965, | |
| "grad_norm": 0.3175494758298602, | |
| "learning_rate": 5.843847476128722e-05, | |
| "loss": 0.9643, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.6739884393063584, | |
| "grad_norm": 0.310484738297623, | |
| "learning_rate": 5.807172717591984e-05, | |
| "loss": 0.9322, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 0.6751445086705202, | |
| "grad_norm": 0.29571778911697455, | |
| "learning_rate": 5.770566255873866e-05, | |
| "loss": 0.929, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.6763005780346821, | |
| "grad_norm": 0.2876079644684902, | |
| "learning_rate": 5.734028687255751e-05, | |
| "loss": 0.9644, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 0.6774566473988439, | |
| "grad_norm": 0.3002944071515893, | |
| "learning_rate": 5.697560606896839e-05, | |
| "loss": 0.9987, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.6786127167630058, | |
| "grad_norm": 0.2931767870621979, | |
| "learning_rate": 5.6611626088244194e-05, | |
| "loss": 0.9474, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 0.6797687861271676, | |
| "grad_norm": 0.28720352206351696, | |
| "learning_rate": 5.6248352859242314e-05, | |
| "loss": 1.0355, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.6809248554913295, | |
| "grad_norm": 0.3191718472968375, | |
| "learning_rate": 5.588579229930784e-05, | |
| "loss": 0.9699, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 0.6820809248554913, | |
| "grad_norm": 0.30266420054049115, | |
| "learning_rate": 5.552395031417712e-05, | |
| "loss": 0.9358, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.6832369942196532, | |
| "grad_norm": 0.30130721873082394, | |
| "learning_rate": 5.516283279788183e-05, | |
| "loss": 0.9496, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 0.684393063583815, | |
| "grad_norm": 0.2929120042219513, | |
| "learning_rate": 5.4802445632652634e-05, | |
| "loss": 0.9248, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.6855491329479769, | |
| "grad_norm": 0.31437058476192303, | |
| "learning_rate": 5.444279468882358e-05, | |
| "loss": 0.9745, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 0.6867052023121387, | |
| "grad_norm": 0.2878711246279315, | |
| "learning_rate": 5.408388582473651e-05, | |
| "loss": 1.0008, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.6878612716763006, | |
| "grad_norm": 0.2808303288729568, | |
| "learning_rate": 5.3725724886645526e-05, | |
| "loss": 0.9381, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 0.6890173410404624, | |
| "grad_norm": 0.3006592351672636, | |
| "learning_rate": 5.3368317708621674e-05, | |
| "loss": 1.0076, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.6901734104046243, | |
| "grad_norm": 0.30870133430305197, | |
| "learning_rate": 5.3011670112458224e-05, | |
| "loss": 0.9616, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 0.6913294797687861, | |
| "grad_norm": 0.29489554268370943, | |
| "learning_rate": 5.2655787907575436e-05, | |
| "loss": 0.908, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.692485549132948, | |
| "grad_norm": 0.30595244192322063, | |
| "learning_rate": 5.230067689092629e-05, | |
| "loss": 0.9902, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 0.6936416184971098, | |
| "grad_norm": 0.3007538157511167, | |
| "learning_rate": 5.19463428469019e-05, | |
| "loss": 1.0259, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.6947976878612717, | |
| "grad_norm": 0.28815906202526853, | |
| "learning_rate": 5.159279154723715e-05, | |
| "loss": 0.9497, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 0.6959537572254335, | |
| "grad_norm": 0.2752889596438775, | |
| "learning_rate": 5.124002875091704e-05, | |
| "loss": 0.9586, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.6971098265895954, | |
| "grad_norm": 0.2850467130558078, | |
| "learning_rate": 5.088806020408252e-05, | |
| "loss": 0.9049, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 0.6982658959537572, | |
| "grad_norm": 0.2750258968098293, | |
| "learning_rate": 5.053689163993703e-05, | |
| "loss": 0.937, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.6994219653179191, | |
| "grad_norm": 0.29519726643012756, | |
| "learning_rate": 5.018652877865322e-05, | |
| "loss": 0.9325, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 0.7005780346820809, | |
| "grad_norm": 0.31869143160605945, | |
| "learning_rate": 4.983697732727964e-05, | |
| "loss": 0.9484, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.7017341040462428, | |
| "grad_norm": 0.30604777371620057, | |
| "learning_rate": 4.948824297964774e-05, | |
| "loss": 0.9497, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 0.7028901734104046, | |
| "grad_norm": 0.2781484111225009, | |
| "learning_rate": 4.914033141627931e-05, | |
| "loss": 0.9732, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.7040462427745665, | |
| "grad_norm": 0.287979860552131, | |
| "learning_rate": 4.87932483042937e-05, | |
| "loss": 0.9513, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 0.7052023121387283, | |
| "grad_norm": 0.28553452094386605, | |
| "learning_rate": 4.8446999297315764e-05, | |
| "loss": 0.9528, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.7063583815028902, | |
| "grad_norm": 0.26876115159064556, | |
| "learning_rate": 4.810159003538365e-05, | |
| "loss": 0.9513, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 0.707514450867052, | |
| "grad_norm": 0.29148448420091455, | |
| "learning_rate": 4.775702614485678e-05, | |
| "loss": 0.9732, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.7086705202312139, | |
| "grad_norm": 0.2858967419780267, | |
| "learning_rate": 4.7413313238324556e-05, | |
| "loss": 0.9874, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 0.7098265895953757, | |
| "grad_norm": 0.2853440571230289, | |
| "learning_rate": 4.707045691451456e-05, | |
| "loss": 0.9365, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.7109826589595376, | |
| "grad_norm": 0.2756362587234919, | |
| "learning_rate": 4.6728462758201574e-05, | |
| "loss": 0.918, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 0.7121387283236994, | |
| "grad_norm": 0.2933985662205089, | |
| "learning_rate": 4.638733634011663e-05, | |
| "loss": 0.9652, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.7132947976878613, | |
| "grad_norm": 0.3088785959772121, | |
| "learning_rate": 4.604708321685618e-05, | |
| "loss": 0.9468, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 0.7144508670520231, | |
| "grad_norm": 0.2897414636494522, | |
| "learning_rate": 4.5707708930791514e-05, | |
| "loss": 0.9136, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.715606936416185, | |
| "grad_norm": 0.2977453761954692, | |
| "learning_rate": 4.536921900997872e-05, | |
| "loss": 0.9684, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 0.7167630057803468, | |
| "grad_norm": 0.31568624217549895, | |
| "learning_rate": 4.5031618968068325e-05, | |
| "loss": 0.9804, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.7179190751445087, | |
| "grad_norm": 0.2937121832237494, | |
| "learning_rate": 4.4694914304215796e-05, | |
| "loss": 0.9923, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 0.7190751445086705, | |
| "grad_norm": 0.31223322912252777, | |
| "learning_rate": 4.4359110502991773e-05, | |
| "loss": 0.9493, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.7202312138728324, | |
| "grad_norm": 0.2948338805474285, | |
| "learning_rate": 4.402421303429274e-05, | |
| "loss": 1.0058, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 0.7213872832369942, | |
| "grad_norm": 0.2998471688331991, | |
| "learning_rate": 4.3690227353251944e-05, | |
| "loss": 0.9793, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.7225433526011561, | |
| "grad_norm": 0.30105760378227486, | |
| "learning_rate": 4.335715890015067e-05, | |
| "loss": 0.9636, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 0.7236994219653179, | |
| "grad_norm": 0.298969154483284, | |
| "learning_rate": 4.302501310032937e-05, | |
| "loss": 0.9849, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.7248554913294798, | |
| "grad_norm": 0.30686646202523543, | |
| "learning_rate": 4.26937953640995e-05, | |
| "loss": 0.943, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 0.7260115606936416, | |
| "grad_norm": 0.2906008606262198, | |
| "learning_rate": 4.236351108665537e-05, | |
| "loss": 0.9584, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.7271676300578035, | |
| "grad_norm": 0.2965620261809724, | |
| "learning_rate": 4.203416564798608e-05, | |
| "loss": 0.9376, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 0.7283236994219653, | |
| "grad_norm": 0.29569547507554944, | |
| "learning_rate": 4.170576441278815e-05, | |
| "loss": 1.0158, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.7294797687861272, | |
| "grad_norm": 0.2873361360138618, | |
| "learning_rate": 4.137831273037793e-05, | |
| "loss": 0.956, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 0.730635838150289, | |
| "grad_norm": 0.28863435820628536, | |
| "learning_rate": 4.1051815934604465e-05, | |
| "loss": 0.9456, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.7317919075144509, | |
| "grad_norm": 0.30820662139709337, | |
| "learning_rate": 4.072627934376292e-05, | |
| "loss": 0.9627, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 0.7329479768786127, | |
| "grad_norm": 0.2733290689658499, | |
| "learning_rate": 4.0401708260507495e-05, | |
| "loss": 0.9896, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.7341040462427746, | |
| "grad_norm": 0.29448552437974784, | |
| "learning_rate": 4.00781079717653e-05, | |
| "loss": 0.9817, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 0.7352601156069364, | |
| "grad_norm": 0.31792876220744604, | |
| "learning_rate": 3.975548374865034e-05, | |
| "loss": 0.98, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.7364161849710983, | |
| "grad_norm": 0.2867123384035597, | |
| "learning_rate": 3.943384084637732e-05, | |
| "loss": 0.9845, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 0.7375722543352601, | |
| "grad_norm": 0.29809867628254705, | |
| "learning_rate": 3.9113184504176426e-05, | |
| "loss": 0.9589, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.738728323699422, | |
| "grad_norm": 0.2962633412794635, | |
| "learning_rate": 3.879351994520774e-05, | |
| "loss": 0.9644, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 0.7398843930635838, | |
| "grad_norm": 0.3173965169375924, | |
| "learning_rate": 3.847485237647614e-05, | |
| "loss": 0.9243, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.7410404624277457, | |
| "grad_norm": 0.2802160296625643, | |
| "learning_rate": 3.815718698874672e-05, | |
| "loss": 0.9627, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 0.7421965317919075, | |
| "grad_norm": 0.30606251064274426, | |
| "learning_rate": 3.7840528956459956e-05, | |
| "loss": 1.0133, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.7433526011560694, | |
| "grad_norm": 0.29567079239489624, | |
| "learning_rate": 3.752488343764751e-05, | |
| "loss": 1.0064, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 0.7445086705202312, | |
| "grad_norm": 0.2908365758289181, | |
| "learning_rate": 3.721025557384845e-05, | |
| "loss": 0.939, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.7456647398843931, | |
| "grad_norm": 0.2821772422176111, | |
| "learning_rate": 3.689665049002513e-05, | |
| "loss": 0.9176, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 0.7468208092485549, | |
| "grad_norm": 0.28323988627298013, | |
| "learning_rate": 3.658407329447986e-05, | |
| "loss": 0.9504, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.7479768786127168, | |
| "grad_norm": 0.3052577077686943, | |
| "learning_rate": 3.627252907877184e-05, | |
| "loss": 0.963, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 0.7491329479768786, | |
| "grad_norm": 0.29287895420892424, | |
| "learning_rate": 3.5962022917633976e-05, | |
| "loss": 0.9758, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.7502890173410405, | |
| "grad_norm": 0.26219424592081725, | |
| "learning_rate": 3.56525598688904e-05, | |
| "loss": 0.9501, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 0.7514450867052023, | |
| "grad_norm": 0.2999575076011054, | |
| "learning_rate": 3.534414497337406e-05, | |
| "loss": 0.9979, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.7526011560693642, | |
| "grad_norm": 0.30401739845864295, | |
| "learning_rate": 3.503678325484448e-05, | |
| "loss": 0.96, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 0.753757225433526, | |
| "grad_norm": 0.2861448274685923, | |
| "learning_rate": 3.473047971990605e-05, | |
| "loss": 0.9956, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.7549132947976879, | |
| "grad_norm": 0.2749463140020902, | |
| "learning_rate": 3.442523935792651e-05, | |
| "loss": 1.0207, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 0.7560693641618497, | |
| "grad_norm": 0.2987193247584773, | |
| "learning_rate": 3.4121067140955455e-05, | |
| "loss": 0.9469, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.7572254335260116, | |
| "grad_norm": 0.29389883736035816, | |
| "learning_rate": 3.3817968023643766e-05, | |
| "loss": 0.9934, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 0.7583815028901734, | |
| "grad_norm": 0.28653966462371155, | |
| "learning_rate": 3.351594694316239e-05, | |
| "loss": 0.9634, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.7595375722543353, | |
| "grad_norm": 0.29816187415165, | |
| "learning_rate": 3.321500881912225e-05, | |
| "loss": 0.9398, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 0.7606936416184971, | |
| "grad_norm": 0.289319657394509, | |
| "learning_rate": 3.29151585534941e-05, | |
| "loss": 0.8928, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.761849710982659, | |
| "grad_norm": 0.29521868021236264, | |
| "learning_rate": 3.261640103052849e-05, | |
| "loss": 0.9653, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 0.7630057803468208, | |
| "grad_norm": 0.29934248191757035, | |
| "learning_rate": 3.23187411166764e-05, | |
| "loss": 0.97, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.7641618497109827, | |
| "grad_norm": 0.32270679766275234, | |
| "learning_rate": 3.2022183660509916e-05, | |
| "loss": 0.9495, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 0.7653179190751445, | |
| "grad_norm": 0.30814076544629976, | |
| "learning_rate": 3.172673349264316e-05, | |
| "loss": 0.9897, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.7664739884393064, | |
| "grad_norm": 0.2731527495101352, | |
| "learning_rate": 3.143239542565365e-05, | |
| "loss": 0.9922, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 0.7676300578034682, | |
| "grad_norm": 0.29822125577085, | |
| "learning_rate": 3.113917425400406e-05, | |
| "loss": 0.974, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.7687861271676301, | |
| "grad_norm": 0.27831179956286994, | |
| "learning_rate": 3.084707475396385e-05, | |
| "loss": 0.9309, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 0.7699421965317919, | |
| "grad_norm": 0.30061057019675924, | |
| "learning_rate": 3.05561016835317e-05, | |
| "loss": 0.9889, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.7710982658959538, | |
| "grad_norm": 0.2839150865870422, | |
| "learning_rate": 3.026625978235793e-05, | |
| "loss": 0.9279, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 0.7722543352601156, | |
| "grad_norm": 0.28191298234209533, | |
| "learning_rate": 2.9977553771667178e-05, | |
| "loss": 0.9043, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.7734104046242775, | |
| "grad_norm": 0.2839643289974997, | |
| "learning_rate": 2.968998835418174e-05, | |
| "loss": 0.958, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 0.7745664739884393, | |
| "grad_norm": 0.2727626163667981, | |
| "learning_rate": 2.9403568214044687e-05, | |
| "loss": 1.0045, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.7757225433526012, | |
| "grad_norm": 0.2829586617421094, | |
| "learning_rate": 2.9118298016743815e-05, | |
| "loss": 0.9477, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 0.776878612716763, | |
| "grad_norm": 0.2863725325553883, | |
| "learning_rate": 2.8834182409035527e-05, | |
| "loss": 1.0095, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.7780346820809249, | |
| "grad_norm": 0.2981971073978521, | |
| "learning_rate": 2.8551226018869105e-05, | |
| "loss": 1.0128, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 0.7791907514450868, | |
| "grad_norm": 0.2783020472744863, | |
| "learning_rate": 2.8269433455311378e-05, | |
| "loss": 0.9581, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.7803468208092486, | |
| "grad_norm": 0.29194953070328256, | |
| "learning_rate": 2.798880930847173e-05, | |
| "loss": 0.9402, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 0.7815028901734105, | |
| "grad_norm": 0.31785104915649315, | |
| "learning_rate": 2.7709358149427113e-05, | |
| "loss": 0.9912, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.7826589595375723, | |
| "grad_norm": 0.27416031521109446, | |
| "learning_rate": 2.7431084530147834e-05, | |
| "loss": 0.9413, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 0.7838150289017342, | |
| "grad_norm": 0.3023857012875686, | |
| "learning_rate": 2.7153992983423283e-05, | |
| "loss": 1.0101, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.784971098265896, | |
| "grad_norm": 0.27883119977909965, | |
| "learning_rate": 2.687808802278805e-05, | |
| "loss": 0.9699, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 0.7861271676300579, | |
| "grad_norm": 0.28656925239378356, | |
| "learning_rate": 2.6603374142448467e-05, | |
| "loss": 0.9459, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.7872832369942196, | |
| "grad_norm": 0.2805856931300408, | |
| "learning_rate": 2.632985581720947e-05, | |
| "loss": 0.9347, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 0.7884393063583816, | |
| "grad_norm": 0.28925155839879374, | |
| "learning_rate": 2.6057537502401598e-05, | |
| "loss": 0.9549, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.7895953757225433, | |
| "grad_norm": 0.30326180151228377, | |
| "learning_rate": 2.5786423633808487e-05, | |
| "loss": 0.9954, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 0.7907514450867053, | |
| "grad_norm": 0.310861152386562, | |
| "learning_rate": 2.5516518627594542e-05, | |
| "loss": 0.981, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.791907514450867, | |
| "grad_norm": 0.29730609595809665, | |
| "learning_rate": 2.524782688023305e-05, | |
| "loss": 0.9933, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 0.793063583815029, | |
| "grad_norm": 0.29011083826669604, | |
| "learning_rate": 2.4980352768434643e-05, | |
| "loss": 1.0396, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.7942196531791907, | |
| "grad_norm": 0.2941491077374515, | |
| "learning_rate": 2.4714100649075833e-05, | |
| "loss": 0.9676, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 0.7953757225433526, | |
| "grad_norm": 0.28795285220192574, | |
| "learning_rate": 2.4449074859128197e-05, | |
| "loss": 0.9545, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.7965317919075144, | |
| "grad_norm": 0.2973713924271482, | |
| "learning_rate": 2.4185279715587704e-05, | |
| "loss": 0.925, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 0.7976878612716763, | |
| "grad_norm": 0.28132269243331093, | |
| "learning_rate": 2.39227195154043e-05, | |
| "loss": 0.974, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.7988439306358381, | |
| "grad_norm": 0.3130593570967486, | |
| "learning_rate": 2.366139853541197e-05, | |
| "loss": 0.9429, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 0.28725553924621783, | |
| "learning_rate": 2.340132103225916e-05, | |
| "loss": 0.9609, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.8011560693641618, | |
| "grad_norm": 0.2927313493050735, | |
| "learning_rate": 2.3142491242339338e-05, | |
| "loss": 0.9801, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 0.8023121387283237, | |
| "grad_norm": 0.2959751791171696, | |
| "learning_rate": 2.288491338172196e-05, | |
| "loss": 0.9983, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.8034682080924855, | |
| "grad_norm": 0.29009100429119816, | |
| "learning_rate": 2.262859164608393e-05, | |
| "loss": 0.9435, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 0.8046242774566474, | |
| "grad_norm": 0.28065545174572615, | |
| "learning_rate": 2.2373530210641103e-05, | |
| "loss": 0.9664, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.8057803468208092, | |
| "grad_norm": 0.3079082233043266, | |
| "learning_rate": 2.2119733230080408e-05, | |
| "loss": 0.9339, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 0.8069364161849711, | |
| "grad_norm": 0.2994856214599522, | |
| "learning_rate": 2.186720483849206e-05, | |
| "loss": 0.9642, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.8080924855491329, | |
| "grad_norm": 0.3150825598363239, | |
| "learning_rate": 2.1615949149302305e-05, | |
| "loss": 0.9391, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 0.8092485549132948, | |
| "grad_norm": 0.28663989647188765, | |
| "learning_rate": 2.1365970255206402e-05, | |
| "loss": 0.9533, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.8104046242774566, | |
| "grad_norm": 0.3064963810102998, | |
| "learning_rate": 2.1117272228101902e-05, | |
| "loss": 0.9269, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 0.8115606936416185, | |
| "grad_norm": 0.2888662994868131, | |
| "learning_rate": 2.0869859119022328e-05, | |
| "loss": 0.9684, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.8127167630057803, | |
| "grad_norm": 0.29449459192953004, | |
| "learning_rate": 2.0623734958071296e-05, | |
| "loss": 0.9588, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 0.8138728323699422, | |
| "grad_norm": 0.28505307744008485, | |
| "learning_rate": 2.037890375435677e-05, | |
| "loss": 0.969, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.815028901734104, | |
| "grad_norm": 0.264553046638561, | |
| "learning_rate": 2.0135369495925714e-05, | |
| "loss": 0.8865, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 0.8161849710982659, | |
| "grad_norm": 0.2854615552898905, | |
| "learning_rate": 1.9893136149699287e-05, | |
| "loss": 0.9606, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.8173410404624277, | |
| "grad_norm": 0.30132043472669484, | |
| "learning_rate": 1.9652207661408073e-05, | |
| "loss": 0.9783, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 0.8184971098265896, | |
| "grad_norm": 0.28158707938088035, | |
| "learning_rate": 1.941258795552785e-05, | |
| "loss": 0.9702, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.8196531791907514, | |
| "grad_norm": 0.3117678798282903, | |
| "learning_rate": 1.917428093521576e-05, | |
| "loss": 0.9381, | |
| "step": 3545 | |
| }, | |
| { | |
| "epoch": 0.8208092485549133, | |
| "grad_norm": 0.2721446216347946, | |
| "learning_rate": 1.8937290482246606e-05, | |
| "loss": 0.9294, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.8219653179190751, | |
| "grad_norm": 0.274947743569076, | |
| "learning_rate": 1.870162045694971e-05, | |
| "loss": 0.9874, | |
| "step": 3555 | |
| }, | |
| { | |
| "epoch": 0.823121387283237, | |
| "grad_norm": 0.29164661845373113, | |
| "learning_rate": 1.8467274698145942e-05, | |
| "loss": 0.9641, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.8242774566473988, | |
| "grad_norm": 0.2857254550756751, | |
| "learning_rate": 1.8234257023085234e-05, | |
| "loss": 0.9629, | |
| "step": 3565 | |
| }, | |
| { | |
| "epoch": 0.8254335260115607, | |
| "grad_norm": 0.28344620289201494, | |
| "learning_rate": 1.8002571227384467e-05, | |
| "loss": 0.9213, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.8265895953757225, | |
| "grad_norm": 0.2940307676845195, | |
| "learning_rate": 1.777222108496558e-05, | |
| "loss": 0.9567, | |
| "step": 3575 | |
| }, | |
| { | |
| "epoch": 0.8277456647398844, | |
| "grad_norm": 0.29951372831586726, | |
| "learning_rate": 1.7543210347994022e-05, | |
| "loss": 0.9813, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.8289017341040462, | |
| "grad_norm": 0.2878040658280256, | |
| "learning_rate": 1.7315542746817825e-05, | |
| "loss": 0.9152, | |
| "step": 3585 | |
| }, | |
| { | |
| "epoch": 0.8300578034682081, | |
| "grad_norm": 0.30016211006512195, | |
| "learning_rate": 1.7089221989906633e-05, | |
| "loss": 0.9656, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.8312138728323699, | |
| "grad_norm": 0.2744245197404456, | |
| "learning_rate": 1.6864251763791428e-05, | |
| "loss": 1.0234, | |
| "step": 3595 | |
| }, | |
| { | |
| "epoch": 0.8323699421965318, | |
| "grad_norm": 0.28064881967639976, | |
| "learning_rate": 1.664063573300446e-05, | |
| "loss": 1.0006, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.8335260115606936, | |
| "grad_norm": 0.28500326120194347, | |
| "learning_rate": 1.6418377540019536e-05, | |
| "loss": 0.9442, | |
| "step": 3605 | |
| }, | |
| { | |
| "epoch": 0.8346820809248555, | |
| "grad_norm": 0.27741047130235014, | |
| "learning_rate": 1.6197480805192634e-05, | |
| "loss": 0.9173, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.8358381502890173, | |
| "grad_norm": 0.2952544695479881, | |
| "learning_rate": 1.5977949126703084e-05, | |
| "loss": 0.9546, | |
| "step": 3615 | |
| }, | |
| { | |
| "epoch": 0.8369942196531792, | |
| "grad_norm": 0.2753923526465453, | |
| "learning_rate": 1.5759786080494743e-05, | |
| "loss": 0.9922, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.838150289017341, | |
| "grad_norm": 0.2713440737201456, | |
| "learning_rate": 1.554299522021796e-05, | |
| "loss": 0.9215, | |
| "step": 3625 | |
| }, | |
| { | |
| "epoch": 0.8393063583815029, | |
| "grad_norm": 0.2924976737824655, | |
| "learning_rate": 1.5327580077171587e-05, | |
| "loss": 0.9553, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.8404624277456647, | |
| "grad_norm": 0.28640397728398614, | |
| "learning_rate": 1.5113544160245397e-05, | |
| "loss": 0.9395, | |
| "step": 3635 | |
| }, | |
| { | |
| "epoch": 0.8416184971098266, | |
| "grad_norm": 0.2859545519050961, | |
| "learning_rate": 1.4900890955863067e-05, | |
| "loss": 0.9711, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.8427745664739884, | |
| "grad_norm": 0.3089713704921713, | |
| "learning_rate": 1.4689623927925289e-05, | |
| "loss": 1.0092, | |
| "step": 3645 | |
| }, | |
| { | |
| "epoch": 0.8439306358381503, | |
| "grad_norm": 0.2837926097401027, | |
| "learning_rate": 1.4479746517753335e-05, | |
| "loss": 1.0052, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.8450867052023121, | |
| "grad_norm": 0.30261005244224143, | |
| "learning_rate": 1.4271262144033116e-05, | |
| "loss": 0.9991, | |
| "step": 3655 | |
| }, | |
| { | |
| "epoch": 0.846242774566474, | |
| "grad_norm": 0.2674381336270983, | |
| "learning_rate": 1.4064174202759407e-05, | |
| "loss": 0.9552, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.8473988439306358, | |
| "grad_norm": 0.2767756118448426, | |
| "learning_rate": 1.3858486067180465e-05, | |
| "loss": 0.9573, | |
| "step": 3665 | |
| }, | |
| { | |
| "epoch": 0.8485549132947977, | |
| "grad_norm": 0.2836915452259556, | |
| "learning_rate": 1.3654201087743279e-05, | |
| "loss": 0.9466, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.8497109826589595, | |
| "grad_norm": 0.30898236268087276, | |
| "learning_rate": 1.3451322592038774e-05, | |
| "loss": 0.9379, | |
| "step": 3675 | |
| }, | |
| { | |
| "epoch": 0.8508670520231214, | |
| "grad_norm": 0.30463841544406706, | |
| "learning_rate": 1.3249853884747753e-05, | |
| "loss": 0.9506, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.8520231213872832, | |
| "grad_norm": 0.29760105531487135, | |
| "learning_rate": 1.3049798247587064e-05, | |
| "loss": 0.9603, | |
| "step": 3685 | |
| }, | |
| { | |
| "epoch": 0.8531791907514451, | |
| "grad_norm": 0.2910112980743013, | |
| "learning_rate": 1.2851158939256002e-05, | |
| "loss": 0.9903, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.8543352601156069, | |
| "grad_norm": 0.28051123703308095, | |
| "learning_rate": 1.2653939195383446e-05, | |
| "loss": 0.9552, | |
| "step": 3695 | |
| }, | |
| { | |
| "epoch": 0.8554913294797688, | |
| "grad_norm": 0.32807409572083474, | |
| "learning_rate": 1.2458142228474967e-05, | |
| "loss": 0.9391, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.8566473988439306, | |
| "grad_norm": 0.2857511643393787, | |
| "learning_rate": 1.2263771227860555e-05, | |
| "loss": 0.9249, | |
| "step": 3705 | |
| }, | |
| { | |
| "epoch": 0.8578034682080925, | |
| "grad_norm": 0.3130878799401169, | |
| "learning_rate": 1.2070829359642743e-05, | |
| "loss": 0.9801, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.8589595375722543, | |
| "grad_norm": 0.2857834049135047, | |
| "learning_rate": 1.1879319766644969e-05, | |
| "loss": 0.9881, | |
| "step": 3715 | |
| }, | |
| { | |
| "epoch": 0.8601156069364162, | |
| "grad_norm": 0.28295250626185514, | |
| "learning_rate": 1.168924556836034e-05, | |
| "loss": 0.9212, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.861271676300578, | |
| "grad_norm": 0.270720741833164, | |
| "learning_rate": 1.1500609860900934e-05, | |
| "loss": 0.9208, | |
| "step": 3725 | |
| }, | |
| { | |
| "epoch": 0.8624277456647399, | |
| "grad_norm": 0.3310546829847928, | |
| "learning_rate": 1.131341571694724e-05, | |
| "loss": 0.9134, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.8635838150289017, | |
| "grad_norm": 0.2906866125100536, | |
| "learning_rate": 1.1127666185698183e-05, | |
| "loss": 0.9292, | |
| "step": 3735 | |
| }, | |
| { | |
| "epoch": 0.8647398843930636, | |
| "grad_norm": 0.2857402985603114, | |
| "learning_rate": 1.0943364292821478e-05, | |
| "loss": 0.9782, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.8658959537572254, | |
| "grad_norm": 0.30621748628878964, | |
| "learning_rate": 1.0760513040404275e-05, | |
| "loss": 0.9538, | |
| "step": 3745 | |
| }, | |
| { | |
| "epoch": 0.8670520231213873, | |
| "grad_norm": 0.3076325870742748, | |
| "learning_rate": 1.0579115406904327e-05, | |
| "loss": 0.9814, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.8682080924855491, | |
| "grad_norm": 0.2875679417313269, | |
| "learning_rate": 1.0399174347101404e-05, | |
| "loss": 0.9682, | |
| "step": 3755 | |
| }, | |
| { | |
| "epoch": 0.869364161849711, | |
| "grad_norm": 0.2827349143068363, | |
| "learning_rate": 1.0220692792049169e-05, | |
| "loss": 0.9715, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.8705202312138728, | |
| "grad_norm": 0.2898950516764476, | |
| "learning_rate": 1.0043673649027518e-05, | |
| "loss": 0.9809, | |
| "step": 3765 | |
| }, | |
| { | |
| "epoch": 0.8716763005780347, | |
| "grad_norm": 0.29127323235770597, | |
| "learning_rate": 9.86811980149519e-06, | |
| "loss": 0.9738, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.8728323699421965, | |
| "grad_norm": 0.289489961872677, | |
| "learning_rate": 9.694034109042694e-06, | |
| "loss": 1.0206, | |
| "step": 3775 | |
| }, | |
| { | |
| "epoch": 0.8739884393063584, | |
| "grad_norm": 0.2820602963126248, | |
| "learning_rate": 9.521419407345878e-06, | |
| "loss": 0.8967, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.8751445086705202, | |
| "grad_norm": 0.31061259437004146, | |
| "learning_rate": 9.350278508119636e-06, | |
| "loss": 0.9786, | |
| "step": 3785 | |
| }, | |
| { | |
| "epoch": 0.8763005780346821, | |
| "grad_norm": 0.2815496774647879, | |
| "learning_rate": 9.180614199072146e-06, | |
| "loss": 0.9485, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.8774566473988439, | |
| "grad_norm": 0.3518473115040322, | |
| "learning_rate": 9.012429243859487e-06, | |
| "loss": 0.9903, | |
| "step": 3795 | |
| }, | |
| { | |
| "epoch": 0.8786127167630058, | |
| "grad_norm": 0.2916712847778055, | |
| "learning_rate": 8.845726382040597e-06, | |
| "loss": 0.9397, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.8797687861271676, | |
| "grad_norm": 0.29392008724784624, | |
| "learning_rate": 8.680508329032589e-06, | |
| "loss": 0.9629, | |
| "step": 3805 | |
| }, | |
| { | |
| "epoch": 0.8809248554913295, | |
| "grad_norm": 0.28601993585254176, | |
| "learning_rate": 8.516777776066643e-06, | |
| "loss": 0.982, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.8820809248554913, | |
| "grad_norm": 0.28346541614728327, | |
| "learning_rate": 8.354537390144057e-06, | |
| "loss": 0.9446, | |
| "step": 3815 | |
| }, | |
| { | |
| "epoch": 0.8832369942196532, | |
| "grad_norm": 0.289284064978837, | |
| "learning_rate": 8.19378981399287e-06, | |
| "loss": 0.9647, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.884393063583815, | |
| "grad_norm": 0.2867377684057541, | |
| "learning_rate": 8.034537666024822e-06, | |
| "loss": 0.9455, | |
| "step": 3825 | |
| }, | |
| { | |
| "epoch": 0.8855491329479769, | |
| "grad_norm": 0.2756548624705791, | |
| "learning_rate": 7.876783540292599e-06, | |
| "loss": 1.0273, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.8867052023121387, | |
| "grad_norm": 0.2936449029949148, | |
| "learning_rate": 7.720530006447736e-06, | |
| "loss": 0.9742, | |
| "step": 3835 | |
| }, | |
| { | |
| "epoch": 0.8878612716763006, | |
| "grad_norm": 0.297967509970897, | |
| "learning_rate": 7.565779609698631e-06, | |
| "loss": 0.9329, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.8890173410404625, | |
| "grad_norm": 0.2942735618729235, | |
| "learning_rate": 7.412534870769116e-06, | |
| "loss": 0.9581, | |
| "step": 3845 | |
| }, | |
| { | |
| "epoch": 0.8901734104046243, | |
| "grad_norm": 0.28806737344976613, | |
| "learning_rate": 7.260798285857484e-06, | |
| "loss": 1.0075, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.8913294797687862, | |
| "grad_norm": 0.2831951777045236, | |
| "learning_rate": 7.110572326595711e-06, | |
| "loss": 0.9971, | |
| "step": 3855 | |
| }, | |
| { | |
| "epoch": 0.892485549132948, | |
| "grad_norm": 0.28520928967193043, | |
| "learning_rate": 6.961859440009233e-06, | |
| "loss": 1.0001, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.8936416184971099, | |
| "grad_norm": 0.27689057228854386, | |
| "learning_rate": 6.8146620484771495e-06, | |
| "loss": 0.9889, | |
| "step": 3865 | |
| }, | |
| { | |
| "epoch": 0.8947976878612717, | |
| "grad_norm": 0.285699685858991, | |
| "learning_rate": 6.668982549692649e-06, | |
| "loss": 0.9478, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.8959537572254336, | |
| "grad_norm": 0.2754348125812895, | |
| "learning_rate": 6.524823316624063e-06, | |
| "loss": 0.9343, | |
| "step": 3875 | |
| }, | |
| { | |
| "epoch": 0.8971098265895954, | |
| "grad_norm": 0.29989809168873266, | |
| "learning_rate": 6.382186697476167e-06, | |
| "loss": 0.9789, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.8982658959537573, | |
| "grad_norm": 0.2828138414197748, | |
| "learning_rate": 6.2410750156518985e-06, | |
| "loss": 1.0062, | |
| "step": 3885 | |
| }, | |
| { | |
| "epoch": 0.8994219653179191, | |
| "grad_norm": 0.2889085637993367, | |
| "learning_rate": 6.101490569714574e-06, | |
| "loss": 0.9696, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 0.900578034682081, | |
| "grad_norm": 0.29960325302848234, | |
| "learning_rate": 5.963435633350412e-06, | |
| "loss": 0.9162, | |
| "step": 3895 | |
| }, | |
| { | |
| "epoch": 0.9017341040462428, | |
| "grad_norm": 0.3005589289503204, | |
| "learning_rate": 5.826912455331468e-06, | |
| "loss": 0.9903, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.9028901734104047, | |
| "grad_norm": 0.2928261089403607, | |
| "learning_rate": 5.691923259479093e-06, | |
| "loss": 0.9736, | |
| "step": 3905 | |
| }, | |
| { | |
| "epoch": 0.9040462427745665, | |
| "grad_norm": 0.28429998893881614, | |
| "learning_rate": 5.558470244627634e-06, | |
| "loss": 0.9711, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 0.9052023121387284, | |
| "grad_norm": 0.2784468816364227, | |
| "learning_rate": 5.4265555845886215e-06, | |
| "loss": 0.9811, | |
| "step": 3915 | |
| }, | |
| { | |
| "epoch": 0.9063583815028902, | |
| "grad_norm": 0.2872050965542848, | |
| "learning_rate": 5.29618142811541e-06, | |
| "loss": 0.954, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.9075144508670521, | |
| "grad_norm": 0.2948554706499198, | |
| "learning_rate": 5.16734989886809e-06, | |
| "loss": 0.9811, | |
| "step": 3925 | |
| }, | |
| { | |
| "epoch": 0.9086705202312139, | |
| "grad_norm": 0.28399320997388067, | |
| "learning_rate": 5.040063095379011e-06, | |
| "loss": 0.9175, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.9098265895953758, | |
| "grad_norm": 0.2855808882463355, | |
| "learning_rate": 4.914323091018535e-06, | |
| "loss": 0.9901, | |
| "step": 3935 | |
| }, | |
| { | |
| "epoch": 0.9109826589595376, | |
| "grad_norm": 0.3021329131025513, | |
| "learning_rate": 4.790131933961206e-06, | |
| "loss": 0.9613, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.9121387283236995, | |
| "grad_norm": 0.2862589716791116, | |
| "learning_rate": 4.6674916471524995e-06, | |
| "loss": 0.9961, | |
| "step": 3945 | |
| }, | |
| { | |
| "epoch": 0.9132947976878613, | |
| "grad_norm": 0.2998472712073513, | |
| "learning_rate": 4.546404228275824e-06, | |
| "loss": 0.9845, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.9144508670520232, | |
| "grad_norm": 0.28603652316011496, | |
| "learning_rate": 4.426871649719932e-06, | |
| "loss": 0.9765, | |
| "step": 3955 | |
| }, | |
| { | |
| "epoch": 0.915606936416185, | |
| "grad_norm": 0.2872761079408887, | |
| "learning_rate": 4.3088958585468686e-06, | |
| "loss": 0.9997, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.9167630057803469, | |
| "grad_norm": 0.2873199733628149, | |
| "learning_rate": 4.192478776460229e-06, | |
| "loss": 1.0064, | |
| "step": 3965 | |
| }, | |
| { | |
| "epoch": 0.9179190751445087, | |
| "grad_norm": 0.27515435582137365, | |
| "learning_rate": 4.077622299773831e-06, | |
| "loss": 0.9189, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 0.9190751445086706, | |
| "grad_norm": 0.27999010046883294, | |
| "learning_rate": 3.96432829938086e-06, | |
| "loss": 0.9885, | |
| "step": 3975 | |
| }, | |
| { | |
| "epoch": 0.9202312138728324, | |
| "grad_norm": 0.2635891865916151, | |
| "learning_rate": 3.8525986207233465e-06, | |
| "loss": 0.9352, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.9213872832369943, | |
| "grad_norm": 0.28633688125843376, | |
| "learning_rate": 3.742435083762186e-06, | |
| "loss": 0.9599, | |
| "step": 3985 | |
| }, | |
| { | |
| "epoch": 0.922543352601156, | |
| "grad_norm": 0.2826637970887437, | |
| "learning_rate": 3.633839482947421e-06, | |
| "loss": 0.9948, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.923699421965318, | |
| "grad_norm": 0.30153590546133957, | |
| "learning_rate": 3.526813587189026e-06, | |
| "loss": 0.9804, | |
| "step": 3995 | |
| }, | |
| { | |
| "epoch": 0.9248554913294798, | |
| "grad_norm": 0.2858269823618572, | |
| "learning_rate": 3.4213591398281175e-06, | |
| "loss": 0.974, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.9260115606936417, | |
| "grad_norm": 0.2820688133929364, | |
| "learning_rate": 3.3174778586085643e-06, | |
| "loss": 0.9566, | |
| "step": 4005 | |
| }, | |
| { | |
| "epoch": 0.9271676300578034, | |
| "grad_norm": 0.3000844480451751, | |
| "learning_rate": 3.2151714356489225e-06, | |
| "loss": 1.0118, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 0.9283236994219654, | |
| "grad_norm": 0.2790666100533931, | |
| "learning_rate": 3.114441537415014e-06, | |
| "loss": 0.9533, | |
| "step": 4015 | |
| }, | |
| { | |
| "epoch": 0.9294797687861271, | |
| "grad_norm": 0.29749566810305794, | |
| "learning_rate": 3.0152898046926557e-06, | |
| "loss": 0.9569, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.930635838150289, | |
| "grad_norm": 0.28219907290129675, | |
| "learning_rate": 2.917717852560997e-06, | |
| "loss": 0.9512, | |
| "step": 4025 | |
| }, | |
| { | |
| "epoch": 0.9317919075144508, | |
| "grad_norm": 0.29888043306041123, | |
| "learning_rate": 2.8217272703661923e-06, | |
| "loss": 1.0231, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 0.9329479768786128, | |
| "grad_norm": 0.26919643142805166, | |
| "learning_rate": 2.727319621695501e-06, | |
| "loss": 0.9873, | |
| "step": 4035 | |
| }, | |
| { | |
| "epoch": 0.9341040462427745, | |
| "grad_norm": 0.29380224096591, | |
| "learning_rate": 2.6344964443518526e-06, | |
| "loss": 0.9385, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.9352601156069364, | |
| "grad_norm": 0.28299962979190113, | |
| "learning_rate": 2.5432592503288e-06, | |
| "loss": 0.9292, | |
| "step": 4045 | |
| }, | |
| { | |
| "epoch": 0.9364161849710982, | |
| "grad_norm": 0.29962326178158855, | |
| "learning_rate": 2.453609525785816e-06, | |
| "loss": 0.9773, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.9375722543352601, | |
| "grad_norm": 0.32360192327986553, | |
| "learning_rate": 2.3655487310241585e-06, | |
| "loss": 0.9529, | |
| "step": 4055 | |
| }, | |
| { | |
| "epoch": 0.9387283236994219, | |
| "grad_norm": 0.30410924441522585, | |
| "learning_rate": 2.279078300463089e-06, | |
| "loss": 1.0106, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.9398843930635838, | |
| "grad_norm": 0.30431377413637145, | |
| "learning_rate": 2.1941996426164344e-06, | |
| "loss": 0.9735, | |
| "step": 4065 | |
| }, | |
| { | |
| "epoch": 0.9410404624277456, | |
| "grad_norm": 0.2903468628517127, | |
| "learning_rate": 2.1109141400697418e-06, | |
| "loss": 0.972, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 0.9421965317919075, | |
| "grad_norm": 0.2954528106164249, | |
| "learning_rate": 2.029223149457682e-06, | |
| "loss": 0.9662, | |
| "step": 4075 | |
| }, | |
| { | |
| "epoch": 0.9433526011560693, | |
| "grad_norm": 0.27439734246785746, | |
| "learning_rate": 1.949128001441969e-06, | |
| "loss": 0.9509, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.9445086705202312, | |
| "grad_norm": 0.27172305484407105, | |
| "learning_rate": 1.8706300006896882e-06, | |
| "loss": 0.8628, | |
| "step": 4085 | |
| }, | |
| { | |
| "epoch": 0.945664739884393, | |
| "grad_norm": 0.27216092212065524, | |
| "learning_rate": 1.7937304258520692e-06, | |
| "loss": 0.9785, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 0.9468208092485549, | |
| "grad_norm": 0.2918352349773759, | |
| "learning_rate": 1.718430529543613e-06, | |
| "loss": 0.9571, | |
| "step": 4095 | |
| }, | |
| { | |
| "epoch": 0.9479768786127167, | |
| "grad_norm": 0.272999376680257, | |
| "learning_rate": 1.6447315383217643e-06, | |
| "loss": 0.918, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.9491329479768786, | |
| "grad_norm": 0.27901699608664005, | |
| "learning_rate": 1.5726346526668156e-06, | |
| "loss": 0.9528, | |
| "step": 4105 | |
| }, | |
| { | |
| "epoch": 0.9502890173410404, | |
| "grad_norm": 0.28289047639247433, | |
| "learning_rate": 1.5021410469624465e-06, | |
| "loss": 0.9067, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.9514450867052023, | |
| "grad_norm": 0.2825322282067992, | |
| "learning_rate": 1.4332518694765707e-06, | |
| "loss": 0.8735, | |
| "step": 4115 | |
| }, | |
| { | |
| "epoch": 0.9526011560693641, | |
| "grad_norm": 0.28837647678016226, | |
| "learning_rate": 1.3659682423425968e-06, | |
| "loss": 0.9432, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.953757225433526, | |
| "grad_norm": 0.2876129817030531, | |
| "learning_rate": 1.30029126154122e-06, | |
| "loss": 0.9268, | |
| "step": 4125 | |
| }, | |
| { | |
| "epoch": 0.9549132947976878, | |
| "grad_norm": 0.2792371887058164, | |
| "learning_rate": 1.236221996882514e-06, | |
| "loss": 0.9726, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 0.9560693641618497, | |
| "grad_norm": 0.29099718253951457, | |
| "learning_rate": 1.1737614919885008e-06, | |
| "loss": 0.9209, | |
| "step": 4135 | |
| }, | |
| { | |
| "epoch": 0.9572254335260115, | |
| "grad_norm": 0.2995617326772327, | |
| "learning_rate": 1.1129107642761872e-06, | |
| "loss": 0.9187, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.9583815028901734, | |
| "grad_norm": 0.2762400721384854, | |
| "learning_rate": 1.053670804940987e-06, | |
| "loss": 0.9142, | |
| "step": 4145 | |
| }, | |
| { | |
| "epoch": 0.9595375722543352, | |
| "grad_norm": 0.2931863784593866, | |
| "learning_rate": 9.960425789405258e-07, | |
| "loss": 0.9084, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.9606936416184971, | |
| "grad_norm": 0.2880567882354411, | |
| "learning_rate": 9.400270249789955e-07, | |
| "loss": 0.9868, | |
| "step": 4155 | |
| }, | |
| { | |
| "epoch": 0.9618497109826589, | |
| "grad_norm": 0.2871636561882668, | |
| "learning_rate": 8.856250554918344e-07, | |
| "loss": 0.9152, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.9630057803468208, | |
| "grad_norm": 0.28517050924588744, | |
| "learning_rate": 8.328375566308166e-07, | |
| "loss": 0.9757, | |
| "step": 4165 | |
| }, | |
| { | |
| "epoch": 0.9641618497109826, | |
| "grad_norm": 0.28587044682235146, | |
| "learning_rate": 7.816653882496971e-07, | |
| "loss": 0.9537, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.9653179190751445, | |
| "grad_norm": 0.29706875055091925, | |
| "learning_rate": 7.321093838901449e-07, | |
| "loss": 0.9247, | |
| "step": 4175 | |
| }, | |
| { | |
| "epoch": 0.9664739884393063, | |
| "grad_norm": 0.2861635930847044, | |
| "learning_rate": 6.841703507682206e-07, | |
| "loss": 0.8905, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.9676300578034682, | |
| "grad_norm": 0.2813627882340819, | |
| "learning_rate": 6.378490697611761e-07, | |
| "loss": 0.9749, | |
| "step": 4185 | |
| }, | |
| { | |
| "epoch": 0.96878612716763, | |
| "grad_norm": 0.2865848575770194, | |
| "learning_rate": 5.93146295394742e-07, | |
| "loss": 0.9443, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 0.9699421965317919, | |
| "grad_norm": 0.27882077648739195, | |
| "learning_rate": 5.500627558308713e-07, | |
| "loss": 0.9398, | |
| "step": 4195 | |
| }, | |
| { | |
| "epoch": 0.9710982658959537, | |
| "grad_norm": 0.3065313177380281, | |
| "learning_rate": 5.085991528558487e-07, | |
| "loss": 0.9236, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.9722543352601156, | |
| "grad_norm": 0.2871839420940439, | |
| "learning_rate": 4.687561618688663e-07, | |
| "loss": 0.9676, | |
| "step": 4205 | |
| }, | |
| { | |
| "epoch": 0.9734104046242774, | |
| "grad_norm": 0.2637381920338272, | |
| "learning_rate": 4.3053443187103247e-07, | |
| "loss": 0.8815, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 0.9745664739884393, | |
| "grad_norm": 0.285755856830912, | |
| "learning_rate": 3.9393458545479157e-07, | |
| "loss": 0.9926, | |
| "step": 4215 | |
| }, | |
| { | |
| "epoch": 0.9757225433526011, | |
| "grad_norm": 0.30359812477643444, | |
| "learning_rate": 3.589572187937651e-07, | |
| "loss": 0.9604, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.976878612716763, | |
| "grad_norm": 0.29778116804881444, | |
| "learning_rate": 3.2560290163307083e-07, | |
| "loss": 0.9788, | |
| "step": 4225 | |
| }, | |
| { | |
| "epoch": 0.9780346820809248, | |
| "grad_norm": 0.2794224102916479, | |
| "learning_rate": 2.93872177280019e-07, | |
| "loss": 0.9427, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.9791907514450867, | |
| "grad_norm": 0.3008883017130643, | |
| "learning_rate": 2.637655625952973e-07, | |
| "loss": 0.9475, | |
| "step": 4235 | |
| }, | |
| { | |
| "epoch": 0.9803468208092485, | |
| "grad_norm": 0.28456190197348324, | |
| "learning_rate": 2.3528354798451058e-07, | |
| "loss": 0.9461, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.9815028901734104, | |
| "grad_norm": 0.2829229588306782, | |
| "learning_rate": 2.0842659739019887e-07, | |
| "loss": 0.9524, | |
| "step": 4245 | |
| }, | |
| { | |
| "epoch": 0.9826589595375722, | |
| "grad_norm": 0.283105235302632, | |
| "learning_rate": 1.8319514828430973e-07, | |
| "loss": 0.9474, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.9838150289017341, | |
| "grad_norm": 0.27690208166745106, | |
| "learning_rate": 1.5958961166104847e-07, | |
| "loss": 0.9351, | |
| "step": 4255 | |
| }, | |
| { | |
| "epoch": 0.9849710982658959, | |
| "grad_norm": 0.25712842285219684, | |
| "learning_rate": 1.3761037203017245e-07, | |
| "loss": 0.914, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.9861271676300578, | |
| "grad_norm": 0.2735837511043363, | |
| "learning_rate": 1.1725778741076276e-07, | |
| "loss": 0.8924, | |
| "step": 4265 | |
| }, | |
| { | |
| "epoch": 0.9872832369942196, | |
| "grad_norm": 0.2820464817023967, | |
| "learning_rate": 9.853218932536212e-08, | |
| "loss": 0.9359, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 0.9884393063583815, | |
| "grad_norm": 0.27653871874823904, | |
| "learning_rate": 8.143388279460151e-08, | |
| "loss": 0.9704, | |
| "step": 4275 | |
| }, | |
| { | |
| "epoch": 0.9895953757225433, | |
| "grad_norm": 0.2836276601553615, | |
| "learning_rate": 6.596314633219303e-08, | |
| "loss": 0.9966, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.9907514450867052, | |
| "grad_norm": 0.27880904921583405, | |
| "learning_rate": 5.2120231940433474e-08, | |
| "loss": 0.9227, | |
| "step": 4285 | |
| }, | |
| { | |
| "epoch": 0.991907514450867, | |
| "grad_norm": 0.2657505634085022, | |
| "learning_rate": 3.990536510608544e-08, | |
| "loss": 0.9657, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.9930635838150289, | |
| "grad_norm": 0.2908473414800192, | |
| "learning_rate": 2.9318744796669183e-08, | |
| "loss": 0.9206, | |
| "step": 4295 | |
| }, | |
| { | |
| "epoch": 0.9942196531791907, | |
| "grad_norm": 0.2957217704373017, | |
| "learning_rate": 2.036054345729843e-08, | |
| "loss": 0.9528, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.9953757225433526, | |
| "grad_norm": 0.2910406474246415, | |
| "learning_rate": 1.3030907007793857e-08, | |
| "loss": 1.0005, | |
| "step": 4305 | |
| }, | |
| { | |
| "epoch": 0.9965317919075144, | |
| "grad_norm": 0.2796254503042765, | |
| "learning_rate": 7.329954840362696e-09, | |
| "loss": 1.0259, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 0.9976878612716763, | |
| "grad_norm": 0.2844197423108749, | |
| "learning_rate": 3.257779817600337e-09, | |
| "loss": 0.9363, | |
| "step": 4315 | |
| }, | |
| { | |
| "epoch": 0.9988439306358381, | |
| "grad_norm": 0.3388694141182201, | |
| "learning_rate": 8.144482710248369e-10, | |
| "loss": 1.0117, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.29406054557823574, | |
| "learning_rate": 0.0, | |
| "loss": 0.9703, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_runtime": 3.3948, | |
| "eval_samples_per_second": 2.946, | |
| "eval_steps_per_second": 0.884, | |
| "step": 4325 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "step": 4325, | |
| "total_flos": 1.71674607550464e+16, | |
| "train_loss": 0.9719910388461427, | |
| "train_runtime": 20047.9922, | |
| "train_samples_per_second": 3.452, | |
| "train_steps_per_second": 0.216 | |
| } | |
| ], | |
| "logging_steps": 5, | |
| "max_steps": 4325, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 1, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.71674607550464e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |