{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.10710221567708682, "eval_steps": 200.0, "global_step": 2400, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 4.4625923198786177e-05, "grad_norm": 0.38751721382141113, "learning_rate": 2.677376171352075e-09, "loss": 5.556500434875488, "step": 1 }, { "epoch": 0.00022312961599393088, "grad_norm": 0.37884023785591125, "learning_rate": 1.3386880856760375e-08, "loss": 5.5404791831970215, "step": 5 }, { "epoch": 0.00044625923198786177, "grad_norm": 0.38282614946365356, "learning_rate": 2.677376171352075e-08, "loss": 5.552686309814453, "step": 10 }, { "epoch": 0.0006693888479817927, "grad_norm": 0.3849342167377472, "learning_rate": 4.016064257028113e-08, "loss": 5.491200637817383, "step": 15 }, { "epoch": 0.0008925184639757235, "grad_norm": 0.3901270925998688, "learning_rate": 5.35475234270415e-08, "loss": 5.454954147338867, "step": 20 }, { "epoch": 0.0011156480799696543, "grad_norm": 0.38169801235198975, "learning_rate": 6.693440428380188e-08, "loss": 5.510653686523438, "step": 25 }, { "epoch": 0.0013387776959635854, "grad_norm": 0.37212520837783813, "learning_rate": 8.032128514056226e-08, "loss": 5.510681915283203, "step": 30 }, { "epoch": 0.0015619073119575162, "grad_norm": 0.37682151794433594, "learning_rate": 9.370816599732262e-08, "loss": 5.651955413818359, "step": 35 }, { "epoch": 0.001785036927951447, "grad_norm": 0.4001241624355316, "learning_rate": 1.07095046854083e-07, "loss": 5.452754211425781, "step": 40 }, { "epoch": 0.002008166543945378, "grad_norm": 0.405489057302475, "learning_rate": 1.2048192771084337e-07, "loss": 5.520745849609375, "step": 45 }, { "epoch": 0.0022312961599393086, "grad_norm": 0.39249682426452637, "learning_rate": 1.3386880856760375e-07, "loss": 5.529154205322266, "step": 50 }, { "epoch": 0.0024544257759332396, "grad_norm": 0.3897538483142853, "learning_rate": 1.4725568942436413e-07, "loss": 5.533833312988281, "step": 55 }, { "epoch": 0.0026775553919271707, "grad_norm": 0.3601232171058655, "learning_rate": 1.606425702811245e-07, "loss": 5.433098220825196, "step": 60 }, { "epoch": 0.0029006850079211013, "grad_norm": 0.3851630985736847, "learning_rate": 1.7402945113788486e-07, "loss": 5.608365631103515, "step": 65 }, { "epoch": 0.0031238146239150324, "grad_norm": 0.37249451875686646, "learning_rate": 1.8741633199464525e-07, "loss": 5.5082344055175785, "step": 70 }, { "epoch": 0.003346944239908963, "grad_norm": 0.37606358528137207, "learning_rate": 2.0080321285140563e-07, "loss": 5.465315246582032, "step": 75 }, { "epoch": 0.003570073855902894, "grad_norm": 0.3777235448360443, "learning_rate": 2.14190093708166e-07, "loss": 5.494733428955078, "step": 80 }, { "epoch": 0.0037932034718968248, "grad_norm": 0.3441930115222931, "learning_rate": 2.2757697456492636e-07, "loss": 5.509701538085937, "step": 85 }, { "epoch": 0.004016333087890756, "grad_norm": 0.41749539971351624, "learning_rate": 2.4096385542168674e-07, "loss": 5.484718704223633, "step": 90 }, { "epoch": 0.004239462703884687, "grad_norm": 0.426409512758255, "learning_rate": 2.543507362784471e-07, "loss": 5.485423278808594, "step": 95 }, { "epoch": 0.004462592319878617, "grad_norm": 0.3890352249145508, "learning_rate": 2.677376171352075e-07, "loss": 5.415613174438477, "step": 100 }, { "epoch": 0.004685721935872548, "grad_norm": 0.3934517502784729, "learning_rate": 2.811244979919679e-07, "loss": 5.502951049804688, "step": 105 }, { "epoch": 0.004908851551866479, "grad_norm": 0.418043315410614, "learning_rate": 2.9451137884872826e-07, "loss": 5.533256149291992, "step": 110 }, { "epoch": 0.00513198116786041, "grad_norm": 0.38715749979019165, "learning_rate": 3.078982597054886e-07, "loss": 5.538555908203125, "step": 115 }, { "epoch": 0.005355110783854341, "grad_norm": 0.38706645369529724, "learning_rate": 3.21285140562249e-07, "loss": 5.495933151245117, "step": 120 }, { "epoch": 0.005578240399848272, "grad_norm": 0.37968552112579346, "learning_rate": 3.346720214190094e-07, "loss": 5.507209014892578, "step": 125 }, { "epoch": 0.005801370015842203, "grad_norm": 0.37994489073753357, "learning_rate": 3.4805890227576973e-07, "loss": 5.505420684814453, "step": 130 }, { "epoch": 0.006024499631836134, "grad_norm": 0.37164536118507385, "learning_rate": 3.614457831325301e-07, "loss": 5.491921615600586, "step": 135 }, { "epoch": 0.006247629247830065, "grad_norm": 0.3722690939903259, "learning_rate": 3.748326639892905e-07, "loss": 5.4893798828125, "step": 140 }, { "epoch": 0.006470758863823995, "grad_norm": 0.37924614548683167, "learning_rate": 3.882195448460509e-07, "loss": 5.498228073120117, "step": 145 }, { "epoch": 0.006693888479817926, "grad_norm": 0.4063311517238617, "learning_rate": 4.0160642570281125e-07, "loss": 5.492959213256836, "step": 150 }, { "epoch": 0.006917018095811857, "grad_norm": 0.40316569805145264, "learning_rate": 4.149933065595716e-07, "loss": 5.330196762084961, "step": 155 }, { "epoch": 0.007140147711805788, "grad_norm": 0.38558679819107056, "learning_rate": 4.28380187416332e-07, "loss": 5.483318328857422, "step": 160 }, { "epoch": 0.0073632773277997185, "grad_norm": 0.3990952968597412, "learning_rate": 4.417670682730924e-07, "loss": 5.4153087615966795, "step": 165 }, { "epoch": 0.0075864069437936495, "grad_norm": 0.41763484477996826, "learning_rate": 4.551539491298527e-07, "loss": 5.46466178894043, "step": 170 }, { "epoch": 0.007809536559787581, "grad_norm": 0.41940048336982727, "learning_rate": 4.6854082998661315e-07, "loss": 5.480670166015625, "step": 175 }, { "epoch": 0.008032666175781512, "grad_norm": 0.4029385447502136, "learning_rate": 4.819277108433735e-07, "loss": 5.44397201538086, "step": 180 }, { "epoch": 0.008255795791775442, "grad_norm": 0.4002819061279297, "learning_rate": 4.953145917001339e-07, "loss": 5.355976867675781, "step": 185 }, { "epoch": 0.008478925407769374, "grad_norm": 0.41026008129119873, "learning_rate": 5.087014725568942e-07, "loss": 5.4827728271484375, "step": 190 }, { "epoch": 0.008702055023763304, "grad_norm": 0.4057953655719757, "learning_rate": 5.220883534136546e-07, "loss": 5.4723457336425785, "step": 195 }, { "epoch": 0.008925184639757234, "grad_norm": 0.39414113759994507, "learning_rate": 5.35475234270415e-07, "loss": 5.458871841430664, "step": 200 }, { "epoch": 0.009148314255751166, "grad_norm": 0.40346527099609375, "learning_rate": 5.488621151271754e-07, "loss": 5.332700729370117, "step": 205 }, { "epoch": 0.009371443871745096, "grad_norm": 0.4204488694667816, "learning_rate": 5.622489959839358e-07, "loss": 5.45002326965332, "step": 210 }, { "epoch": 0.009594573487739028, "grad_norm": 0.4042298495769501, "learning_rate": 5.756358768406961e-07, "loss": 5.469032287597656, "step": 215 }, { "epoch": 0.009817703103732959, "grad_norm": 0.42059728503227234, "learning_rate": 5.890227576974565e-07, "loss": 5.416038513183594, "step": 220 }, { "epoch": 0.010040832719726889, "grad_norm": 0.4257930517196655, "learning_rate": 6.024096385542169e-07, "loss": 5.365918731689453, "step": 225 }, { "epoch": 0.01026396233572082, "grad_norm": 0.4297005534172058, "learning_rate": 6.157965194109772e-07, "loss": 5.386648941040039, "step": 230 }, { "epoch": 0.010487091951714751, "grad_norm": 0.44034355878829956, "learning_rate": 6.291834002677377e-07, "loss": 5.4175865173339846, "step": 235 }, { "epoch": 0.010710221567708683, "grad_norm": 0.4148600995540619, "learning_rate": 6.42570281124498e-07, "loss": 5.456137084960938, "step": 240 }, { "epoch": 0.010933351183702613, "grad_norm": 0.3927938938140869, "learning_rate": 6.559571619812583e-07, "loss": 5.421617889404297, "step": 245 }, { "epoch": 0.011156480799696543, "grad_norm": 0.46364349126815796, "learning_rate": 6.693440428380188e-07, "loss": 5.263519287109375, "step": 250 }, { "epoch": 0.011379610415690475, "grad_norm": 0.4380209147930145, "learning_rate": 6.827309236947791e-07, "loss": 5.4163257598876955, "step": 255 }, { "epoch": 0.011602740031684405, "grad_norm": 0.44802144169807434, "learning_rate": 6.961178045515395e-07, "loss": 5.307229232788086, "step": 260 }, { "epoch": 0.011825869647678336, "grad_norm": 0.48721882700920105, "learning_rate": 7.095046854082999e-07, "loss": 5.361153030395508, "step": 265 }, { "epoch": 0.012048999263672268, "grad_norm": 0.4634721875190735, "learning_rate": 7.228915662650602e-07, "loss": 5.335968017578125, "step": 270 }, { "epoch": 0.012272128879666198, "grad_norm": 0.44632282853126526, "learning_rate": 7.362784471218206e-07, "loss": 5.377736282348633, "step": 275 }, { "epoch": 0.01249525849566013, "grad_norm": 0.46110817790031433, "learning_rate": 7.49665327978581e-07, "loss": 5.379433059692383, "step": 280 }, { "epoch": 0.01271838811165406, "grad_norm": 0.4558963477611542, "learning_rate": 7.630522088353414e-07, "loss": 5.252085876464844, "step": 285 }, { "epoch": 0.01294151772764799, "grad_norm": 0.4590746760368347, "learning_rate": 7.764390896921018e-07, "loss": 5.320807647705078, "step": 290 }, { "epoch": 0.013164647343641922, "grad_norm": 0.458065927028656, "learning_rate": 7.898259705488621e-07, "loss": 5.342620086669922, "step": 295 }, { "epoch": 0.013387776959635852, "grad_norm": 0.47713929414749146, "learning_rate": 8.032128514056225e-07, "loss": 5.309723281860352, "step": 300 }, { "epoch": 0.013610906575629784, "grad_norm": 0.4549182653427124, "learning_rate": 8.16599732262383e-07, "loss": 5.263495635986328, "step": 305 }, { "epoch": 0.013834036191623714, "grad_norm": 0.45380067825317383, "learning_rate": 8.299866131191432e-07, "loss": 5.270233535766602, "step": 310 }, { "epoch": 0.014057165807617645, "grad_norm": 0.4339616894721985, "learning_rate": 8.433734939759036e-07, "loss": 5.273446655273437, "step": 315 }, { "epoch": 0.014280295423611577, "grad_norm": 0.44467630982398987, "learning_rate": 8.56760374832664e-07, "loss": 5.210543823242188, "step": 320 }, { "epoch": 0.014503425039605507, "grad_norm": 0.4964953064918518, "learning_rate": 8.701472556894243e-07, "loss": 5.216452026367188, "step": 325 }, { "epoch": 0.014726554655599437, "grad_norm": 0.4745877683162689, "learning_rate": 8.835341365461848e-07, "loss": 5.237066650390625, "step": 330 }, { "epoch": 0.014949684271593369, "grad_norm": 0.4923485815525055, "learning_rate": 8.969210174029452e-07, "loss": 5.16167106628418, "step": 335 }, { "epoch": 0.015172813887587299, "grad_norm": 0.45548155903816223, "learning_rate": 9.103078982597054e-07, "loss": 5.184793090820312, "step": 340 }, { "epoch": 0.015395943503581231, "grad_norm": 0.5323435068130493, "learning_rate": 9.236947791164659e-07, "loss": 5.1162841796875, "step": 345 }, { "epoch": 0.015619073119575161, "grad_norm": 0.5220831036567688, "learning_rate": 9.370816599732263e-07, "loss": 5.135643005371094, "step": 350 }, { "epoch": 0.01584220273556909, "grad_norm": 0.5357551574707031, "learning_rate": 9.504685408299866e-07, "loss": 5.001235580444336, "step": 355 }, { "epoch": 0.016065332351563023, "grad_norm": 0.49958014488220215, "learning_rate": 9.63855421686747e-07, "loss": 5.096822357177734, "step": 360 }, { "epoch": 0.016288461967556955, "grad_norm": 0.53453129529953, "learning_rate": 9.772423025435074e-07, "loss": 5.057343673706055, "step": 365 }, { "epoch": 0.016511591583550884, "grad_norm": 0.49969372153282166, "learning_rate": 9.906291834002677e-07, "loss": 5.097013092041015, "step": 370 }, { "epoch": 0.016734721199544816, "grad_norm": 0.477762907743454, "learning_rate": 1.0040160642570282e-06, "loss": 5.05908317565918, "step": 375 }, { "epoch": 0.016957850815538748, "grad_norm": 0.5632966756820679, "learning_rate": 1.0174029451137885e-06, "loss": 4.880051040649414, "step": 380 }, { "epoch": 0.017180980431532676, "grad_norm": 0.4806019961833954, "learning_rate": 1.0307898259705488e-06, "loss": 4.938255310058594, "step": 385 }, { "epoch": 0.017404110047526608, "grad_norm": 0.5882935523986816, "learning_rate": 1.0441767068273092e-06, "loss": 4.972189331054688, "step": 390 }, { "epoch": 0.01762723966352054, "grad_norm": 0.583991289138794, "learning_rate": 1.0575635876840697e-06, "loss": 4.8074909210205075, "step": 395 }, { "epoch": 0.01785036927951447, "grad_norm": 0.6130750179290771, "learning_rate": 1.07095046854083e-06, "loss": 4.851003265380859, "step": 400 }, { "epoch": 0.0180734988955084, "grad_norm": 0.630670428276062, "learning_rate": 1.0843373493975905e-06, "loss": 4.742978286743164, "step": 405 }, { "epoch": 0.018296628511502332, "grad_norm": 0.6021527647972107, "learning_rate": 1.0977242302543508e-06, "loss": 4.795417022705078, "step": 410 }, { "epoch": 0.018519758127496264, "grad_norm": 0.6027496457099915, "learning_rate": 1.111111111111111e-06, "loss": 4.799094009399414, "step": 415 }, { "epoch": 0.018742887743490193, "grad_norm": 0.62542325258255, "learning_rate": 1.1244979919678715e-06, "loss": 4.734822845458984, "step": 420 }, { "epoch": 0.018966017359484125, "grad_norm": 0.6744217872619629, "learning_rate": 1.137884872824632e-06, "loss": 4.470468902587891, "step": 425 }, { "epoch": 0.019189146975478057, "grad_norm": 0.6994063854217529, "learning_rate": 1.1512717536813923e-06, "loss": 4.7334716796875, "step": 430 }, { "epoch": 0.019412276591471985, "grad_norm": 0.6043664813041687, "learning_rate": 1.1646586345381526e-06, "loss": 4.6335186004638675, "step": 435 }, { "epoch": 0.019635406207465917, "grad_norm": 0.600983738899231, "learning_rate": 1.178045515394913e-06, "loss": 4.521899795532226, "step": 440 }, { "epoch": 0.01985853582345985, "grad_norm": 0.632734477519989, "learning_rate": 1.1914323962516733e-06, "loss": 4.455641555786133, "step": 445 }, { "epoch": 0.020081665439453777, "grad_norm": 0.6841812133789062, "learning_rate": 1.2048192771084338e-06, "loss": 4.552269363403321, "step": 450 }, { "epoch": 0.02030479505544771, "grad_norm": 0.7014154195785522, "learning_rate": 1.2182061579651943e-06, "loss": 4.4612682342529295, "step": 455 }, { "epoch": 0.02052792467144164, "grad_norm": 0.7957334518432617, "learning_rate": 1.2315930388219544e-06, "loss": 4.16539421081543, "step": 460 }, { "epoch": 0.02075105428743557, "grad_norm": 0.6679075360298157, "learning_rate": 1.2449799196787148e-06, "loss": 4.380799102783203, "step": 465 }, { "epoch": 0.020974183903429502, "grad_norm": 0.6911934614181519, "learning_rate": 1.2583668005354753e-06, "loss": 4.188001251220703, "step": 470 }, { "epoch": 0.021197313519423434, "grad_norm": 0.6941277384757996, "learning_rate": 1.2717536813922356e-06, "loss": 4.30323600769043, "step": 475 }, { "epoch": 0.021420443135417366, "grad_norm": 0.6411834359169006, "learning_rate": 1.285140562248996e-06, "loss": 4.210566711425781, "step": 480 }, { "epoch": 0.021643572751411294, "grad_norm": 0.6447572112083435, "learning_rate": 1.2985274431057564e-06, "loss": 4.145897674560547, "step": 485 }, { "epoch": 0.021866702367405226, "grad_norm": 0.6956306099891663, "learning_rate": 1.3119143239625166e-06, "loss": 4.1259113311767575, "step": 490 }, { "epoch": 0.022089831983399158, "grad_norm": 0.6369218826293945, "learning_rate": 1.3253012048192771e-06, "loss": 4.050044250488281, "step": 495 }, { "epoch": 0.022312961599393086, "grad_norm": 0.6199172735214233, "learning_rate": 1.3386880856760376e-06, "loss": 3.9885379791259767, "step": 500 }, { "epoch": 0.02253609121538702, "grad_norm": 0.6004146933555603, "learning_rate": 1.3520749665327979e-06, "loss": 3.9876686096191407, "step": 505 }, { "epoch": 0.02275922083138095, "grad_norm": 0.6573876142501831, "learning_rate": 1.3654618473895582e-06, "loss": 3.860630416870117, "step": 510 }, { "epoch": 0.02298235044737488, "grad_norm": 0.6570749878883362, "learning_rate": 1.3788487282463186e-06, "loss": 3.987852096557617, "step": 515 }, { "epoch": 0.02320548006336881, "grad_norm": 0.6052077412605286, "learning_rate": 1.392235609103079e-06, "loss": 3.773631286621094, "step": 520 }, { "epoch": 0.023428609679362743, "grad_norm": 0.6024802923202515, "learning_rate": 1.4056224899598394e-06, "loss": 3.8334468841552733, "step": 525 }, { "epoch": 0.02365173929535667, "grad_norm": 0.6448367238044739, "learning_rate": 1.4190093708165999e-06, "loss": 3.709330749511719, "step": 530 }, { "epoch": 0.023874868911350603, "grad_norm": 0.5610556602478027, "learning_rate": 1.4323962516733602e-06, "loss": 3.608018493652344, "step": 535 }, { "epoch": 0.024097998527344535, "grad_norm": 0.5798413157463074, "learning_rate": 1.4457831325301204e-06, "loss": 3.513960266113281, "step": 540 }, { "epoch": 0.024321128143338467, "grad_norm": 0.5836862921714783, "learning_rate": 1.459170013386881e-06, "loss": 3.5928466796875, "step": 545 }, { "epoch": 0.024544257759332395, "grad_norm": 0.5705676078796387, "learning_rate": 1.4725568942436412e-06, "loss": 3.5890869140625, "step": 550 }, { "epoch": 0.024767387375326327, "grad_norm": 0.6377549767494202, "learning_rate": 1.4859437751004017e-06, "loss": 3.6042083740234374, "step": 555 }, { "epoch": 0.02499051699132026, "grad_norm": 0.5612762570381165, "learning_rate": 1.499330655957162e-06, "loss": 3.511037826538086, "step": 560 }, { "epoch": 0.025213646607314188, "grad_norm": 0.5943495035171509, "learning_rate": 1.5127175368139222e-06, "loss": 3.5007259368896486, "step": 565 }, { "epoch": 0.02543677622330812, "grad_norm": 0.5872607827186584, "learning_rate": 1.5261044176706827e-06, "loss": 3.3742324829101564, "step": 570 }, { "epoch": 0.02565990583930205, "grad_norm": 0.5494210124015808, "learning_rate": 1.5394912985274432e-06, "loss": 3.3373321533203124, "step": 575 }, { "epoch": 0.02588303545529598, "grad_norm": 0.5138423442840576, "learning_rate": 1.5528781793842037e-06, "loss": 3.346783447265625, "step": 580 }, { "epoch": 0.026106165071289912, "grad_norm": 0.5539407730102539, "learning_rate": 1.566265060240964e-06, "loss": 3.183103561401367, "step": 585 }, { "epoch": 0.026329294687283844, "grad_norm": 0.5888996720314026, "learning_rate": 1.5796519410977242e-06, "loss": 3.357350540161133, "step": 590 }, { "epoch": 0.026552424303277773, "grad_norm": 0.5445035696029663, "learning_rate": 1.5930388219544845e-06, "loss": 3.374500274658203, "step": 595 }, { "epoch": 0.026775553919271704, "grad_norm": 0.5013962388038635, "learning_rate": 1.606425702811245e-06, "loss": 3.3687782287597656, "step": 600 }, { "epoch": 0.026998683535265636, "grad_norm": 0.5150508284568787, "learning_rate": 1.6198125836680055e-06, "loss": 3.2590232849121095, "step": 605 }, { "epoch": 0.02722181315125957, "grad_norm": 0.603961169719696, "learning_rate": 1.633199464524766e-06, "loss": 3.1635900497436524, "step": 610 }, { "epoch": 0.027444942767253497, "grad_norm": 0.5365360975265503, "learning_rate": 1.6465863453815263e-06, "loss": 3.3600418090820314, "step": 615 }, { "epoch": 0.02766807238324743, "grad_norm": 0.5401721000671387, "learning_rate": 1.6599732262382863e-06, "loss": 3.183880424499512, "step": 620 }, { "epoch": 0.02789120199924136, "grad_norm": 0.5595034956932068, "learning_rate": 1.6733601070950468e-06, "loss": 3.0762868881225587, "step": 625 }, { "epoch": 0.02811433161523529, "grad_norm": 0.4972860813140869, "learning_rate": 1.6867469879518073e-06, "loss": 3.0724456787109373, "step": 630 }, { "epoch": 0.02833746123122922, "grad_norm": 0.44132477045059204, "learning_rate": 1.7001338688085678e-06, "loss": 3.1175127029418945, "step": 635 }, { "epoch": 0.028560590847223153, "grad_norm": 0.5078781843185425, "learning_rate": 1.713520749665328e-06, "loss": 3.0849445343017576, "step": 640 }, { "epoch": 0.02878372046321708, "grad_norm": 0.4757557809352875, "learning_rate": 1.7269076305220885e-06, "loss": 3.2029571533203125, "step": 645 }, { "epoch": 0.029006850079211013, "grad_norm": 0.4932602345943451, "learning_rate": 1.7402945113788486e-06, "loss": 3.0668067932128906, "step": 650 }, { "epoch": 0.029229979695204945, "grad_norm": 0.5687287449836731, "learning_rate": 1.753681392235609e-06, "loss": 3.113470268249512, "step": 655 }, { "epoch": 0.029453109311198874, "grad_norm": 0.4471919536590576, "learning_rate": 1.7670682730923696e-06, "loss": 2.9105451583862303, "step": 660 }, { "epoch": 0.029676238927192806, "grad_norm": 0.5096343159675598, "learning_rate": 1.7804551539491298e-06, "loss": 2.94341926574707, "step": 665 }, { "epoch": 0.029899368543186738, "grad_norm": 0.5599822402000427, "learning_rate": 1.7938420348058903e-06, "loss": 2.892704391479492, "step": 670 }, { "epoch": 0.03012249815918067, "grad_norm": 0.40055856108665466, "learning_rate": 1.8072289156626508e-06, "loss": 2.896807861328125, "step": 675 }, { "epoch": 0.030345627775174598, "grad_norm": 0.47082021832466125, "learning_rate": 1.8206157965194109e-06, "loss": 2.7966148376464846, "step": 680 }, { "epoch": 0.03056875739116853, "grad_norm": 0.42042815685272217, "learning_rate": 1.8340026773761714e-06, "loss": 2.839730644226074, "step": 685 }, { "epoch": 0.030791887007162462, "grad_norm": 0.49198001623153687, "learning_rate": 1.8473895582329318e-06, "loss": 2.9913705825805663, "step": 690 }, { "epoch": 0.03101501662315639, "grad_norm": 0.4656424820423126, "learning_rate": 1.8607764390896921e-06, "loss": 3.0195621490478515, "step": 695 }, { "epoch": 0.031238146239150322, "grad_norm": 0.46259376406669617, "learning_rate": 1.8741633199464526e-06, "loss": 2.948585319519043, "step": 700 }, { "epoch": 0.03146127585514425, "grad_norm": 0.4671732783317566, "learning_rate": 1.887550200803213e-06, "loss": 2.7285741806030273, "step": 705 }, { "epoch": 0.03168440547113818, "grad_norm": 0.6493679285049438, "learning_rate": 1.9009370816599732e-06, "loss": 2.9263893127441407, "step": 710 }, { "epoch": 0.031907535087132115, "grad_norm": 0.46389710903167725, "learning_rate": 1.9143239625167336e-06, "loss": 2.634868049621582, "step": 715 }, { "epoch": 0.03213066470312605, "grad_norm": 0.44173645973205566, "learning_rate": 1.927710843373494e-06, "loss": 2.7959117889404297, "step": 720 }, { "epoch": 0.03235379431911998, "grad_norm": 0.45440003275871277, "learning_rate": 1.9410977242302546e-06, "loss": 2.9056529998779297, "step": 725 }, { "epoch": 0.03257692393511391, "grad_norm": 0.44541046023368835, "learning_rate": 1.954484605087015e-06, "loss": 2.7214975357055664, "step": 730 }, { "epoch": 0.032800053551107836, "grad_norm": 0.4217804968357086, "learning_rate": 1.967871485943775e-06, "loss": 2.7198978424072267, "step": 735 }, { "epoch": 0.03302318316710177, "grad_norm": 0.46086645126342773, "learning_rate": 1.9812583668005354e-06, "loss": 2.7366127014160155, "step": 740 }, { "epoch": 0.0332463127830957, "grad_norm": 0.41817471385002136, "learning_rate": 1.9946452476572957e-06, "loss": 2.857570457458496, "step": 745 }, { "epoch": 0.03346944239908963, "grad_norm": 0.44163694977760315, "learning_rate": 2.0080321285140564e-06, "loss": 2.8251710891723634, "step": 750 }, { "epoch": 0.03369257201508356, "grad_norm": 0.46645739674568176, "learning_rate": 2.0214190093708167e-06, "loss": 2.548258399963379, "step": 755 }, { "epoch": 0.033915701631077495, "grad_norm": 0.3971084654331207, "learning_rate": 2.034805890227577e-06, "loss": 2.682722473144531, "step": 760 }, { "epoch": 0.03413883124707143, "grad_norm": 0.5373950004577637, "learning_rate": 2.0481927710843377e-06, "loss": 2.8539506912231447, "step": 765 }, { "epoch": 0.03436196086306535, "grad_norm": 0.6133913397789001, "learning_rate": 2.0615796519410975e-06, "loss": 2.7016387939453126, "step": 770 }, { "epoch": 0.034585090479059284, "grad_norm": 0.5247603058815002, "learning_rate": 2.074966532797858e-06, "loss": 2.645602226257324, "step": 775 }, { "epoch": 0.034808220095053216, "grad_norm": 0.4487764537334442, "learning_rate": 2.0883534136546185e-06, "loss": 2.699404716491699, "step": 780 }, { "epoch": 0.03503134971104715, "grad_norm": 0.5962596535682678, "learning_rate": 2.1017402945113788e-06, "loss": 2.7891347885131834, "step": 785 }, { "epoch": 0.03525447932704108, "grad_norm": 0.5410242080688477, "learning_rate": 2.1151271753681395e-06, "loss": 2.675504684448242, "step": 790 }, { "epoch": 0.03547760894303501, "grad_norm": 0.4576985239982605, "learning_rate": 2.1285140562248997e-06, "loss": 2.642281341552734, "step": 795 }, { "epoch": 0.03570073855902894, "grad_norm": 0.45449239015579224, "learning_rate": 2.14190093708166e-06, "loss": 2.7088804244995117, "step": 800 }, { "epoch": 0.03592386817502287, "grad_norm": 0.4782140552997589, "learning_rate": 2.1552878179384203e-06, "loss": 2.698718452453613, "step": 805 }, { "epoch": 0.0361469977910168, "grad_norm": 0.43745940923690796, "learning_rate": 2.168674698795181e-06, "loss": 2.7042917251586913, "step": 810 }, { "epoch": 0.03637012740701073, "grad_norm": 0.6170194149017334, "learning_rate": 2.1820615796519413e-06, "loss": 2.6548912048339846, "step": 815 }, { "epoch": 0.036593257023004665, "grad_norm": 0.5126772522926331, "learning_rate": 2.1954484605087015e-06, "loss": 2.663498306274414, "step": 820 }, { "epoch": 0.0368163866389986, "grad_norm": 0.42503371834754944, "learning_rate": 2.2088353413654622e-06, "loss": 2.766114616394043, "step": 825 }, { "epoch": 0.03703951625499253, "grad_norm": 0.5707629323005676, "learning_rate": 2.222222222222222e-06, "loss": 2.6172439575195314, "step": 830 }, { "epoch": 0.037262645870986454, "grad_norm": 0.45169416069984436, "learning_rate": 2.2356091030789828e-06, "loss": 2.5655603408813477, "step": 835 }, { "epoch": 0.037485775486980386, "grad_norm": 0.44726496934890747, "learning_rate": 2.248995983935743e-06, "loss": 2.5273290634155274, "step": 840 }, { "epoch": 0.03770890510297432, "grad_norm": 0.5287439823150635, "learning_rate": 2.2623828647925033e-06, "loss": 2.6541069030761717, "step": 845 }, { "epoch": 0.03793203471896825, "grad_norm": 0.46616867184638977, "learning_rate": 2.275769745649264e-06, "loss": 2.56517333984375, "step": 850 }, { "epoch": 0.03815516433496218, "grad_norm": 0.4369581937789917, "learning_rate": 2.2891566265060243e-06, "loss": 2.4222312927246095, "step": 855 }, { "epoch": 0.03837829395095611, "grad_norm": 0.4172717332839966, "learning_rate": 2.3025435073627846e-06, "loss": 2.549334716796875, "step": 860 }, { "epoch": 0.03860142356695004, "grad_norm": 0.46124762296676636, "learning_rate": 2.315930388219545e-06, "loss": 2.5675996780395507, "step": 865 }, { "epoch": 0.03882455318294397, "grad_norm": 0.4226182699203491, "learning_rate": 2.329317269076305e-06, "loss": 2.4144569396972657, "step": 870 }, { "epoch": 0.0390476827989379, "grad_norm": 0.4673754870891571, "learning_rate": 2.342704149933066e-06, "loss": 2.3829999923706056, "step": 875 }, { "epoch": 0.039270812414931834, "grad_norm": 0.5152049660682678, "learning_rate": 2.356091030789826e-06, "loss": 2.5030202865600586, "step": 880 }, { "epoch": 0.039493942030925766, "grad_norm": 0.5377767086029053, "learning_rate": 2.3694779116465864e-06, "loss": 2.7714466094970702, "step": 885 }, { "epoch": 0.0397170716469197, "grad_norm": 0.45861950516700745, "learning_rate": 2.3828647925033466e-06, "loss": 2.461492729187012, "step": 890 }, { "epoch": 0.03994020126291363, "grad_norm": 0.4814665615558624, "learning_rate": 2.396251673360107e-06, "loss": 2.4756641387939453, "step": 895 }, { "epoch": 0.040163330878907555, "grad_norm": 0.5024599432945251, "learning_rate": 2.4096385542168676e-06, "loss": 2.5811389923095702, "step": 900 }, { "epoch": 0.04038646049490149, "grad_norm": 0.45141535997390747, "learning_rate": 2.423025435073628e-06, "loss": 2.686309814453125, "step": 905 }, { "epoch": 0.04060959011089542, "grad_norm": 0.4806381165981293, "learning_rate": 2.4364123159303886e-06, "loss": 2.530971717834473, "step": 910 }, { "epoch": 0.04083271972688935, "grad_norm": 0.7574505805969238, "learning_rate": 2.449799196787149e-06, "loss": 2.5287572860717775, "step": 915 }, { "epoch": 0.04105584934288328, "grad_norm": 0.39056211709976196, "learning_rate": 2.4631860776439087e-06, "loss": 2.4373926162719726, "step": 920 }, { "epoch": 0.041278978958877215, "grad_norm": 0.4549945890903473, "learning_rate": 2.4765729585006694e-06, "loss": 2.508747100830078, "step": 925 }, { "epoch": 0.04150210857487114, "grad_norm": 0.49253398180007935, "learning_rate": 2.4899598393574297e-06, "loss": 2.5109813690185545, "step": 930 }, { "epoch": 0.04172523819086507, "grad_norm": 0.4185622036457062, "learning_rate": 2.5033467202141904e-06, "loss": 2.5602550506591797, "step": 935 }, { "epoch": 0.041948367806859004, "grad_norm": 0.5255184769630432, "learning_rate": 2.5167336010709507e-06, "loss": 2.5018032073974608, "step": 940 }, { "epoch": 0.042171497422852935, "grad_norm": 0.40105360746383667, "learning_rate": 2.530120481927711e-06, "loss": 2.3703737258911133, "step": 945 }, { "epoch": 0.04239462703884687, "grad_norm": 0.5396221876144409, "learning_rate": 2.543507362784471e-06, "loss": 2.46767520904541, "step": 950 }, { "epoch": 0.0426177566548408, "grad_norm": 0.4686439633369446, "learning_rate": 2.5568942436412315e-06, "loss": 2.535739517211914, "step": 955 }, { "epoch": 0.04284088627083473, "grad_norm": 0.5110803246498108, "learning_rate": 2.570281124497992e-06, "loss": 2.5248756408691406, "step": 960 }, { "epoch": 0.043064015886828656, "grad_norm": 0.5164591670036316, "learning_rate": 2.5836680053547524e-06, "loss": 2.4275962829589846, "step": 965 }, { "epoch": 0.04328714550282259, "grad_norm": 0.42096105217933655, "learning_rate": 2.5970548862115127e-06, "loss": 2.4086238861083986, "step": 970 }, { "epoch": 0.04351027511881652, "grad_norm": 0.5388748645782471, "learning_rate": 2.6104417670682734e-06, "loss": 2.340771293640137, "step": 975 }, { "epoch": 0.04373340473481045, "grad_norm": 0.4667651653289795, "learning_rate": 2.6238286479250333e-06, "loss": 2.506967544555664, "step": 980 }, { "epoch": 0.043956534350804384, "grad_norm": 0.5090363621711731, "learning_rate": 2.637215528781794e-06, "loss": 2.251785469055176, "step": 985 }, { "epoch": 0.044179663966798316, "grad_norm": 0.4600473940372467, "learning_rate": 2.6506024096385542e-06, "loss": 2.3269046783447265, "step": 990 }, { "epoch": 0.04440279358279224, "grad_norm": 0.4453408718109131, "learning_rate": 2.6639892904953145e-06, "loss": 2.4940771102905273, "step": 995 }, { "epoch": 0.04462592319878617, "grad_norm": 0.4637933671474457, "learning_rate": 2.6773761713520752e-06, "loss": 2.578083801269531, "step": 1000 }, { "epoch": 0.044849052814780105, "grad_norm": 0.48008954524993896, "learning_rate": 2.6907630522088355e-06, "loss": 2.509678077697754, "step": 1005 }, { "epoch": 0.04507218243077404, "grad_norm": 0.5186890363693237, "learning_rate": 2.7041499330655958e-06, "loss": 2.420572280883789, "step": 1010 }, { "epoch": 0.04529531204676797, "grad_norm": 0.5090295076370239, "learning_rate": 2.717536813922356e-06, "loss": 2.42071533203125, "step": 1015 }, { "epoch": 0.0455184416627619, "grad_norm": 0.5519959926605225, "learning_rate": 2.7309236947791163e-06, "loss": 2.458280563354492, "step": 1020 }, { "epoch": 0.04574157127875583, "grad_norm": 0.5525882840156555, "learning_rate": 2.744310575635877e-06, "loss": 2.3400331497192384, "step": 1025 }, { "epoch": 0.04596470089474976, "grad_norm": 0.48508089780807495, "learning_rate": 2.7576974564926373e-06, "loss": 2.4403892517089845, "step": 1030 }, { "epoch": 0.04618783051074369, "grad_norm": 0.5230780839920044, "learning_rate": 2.771084337349398e-06, "loss": 2.3652227401733397, "step": 1035 }, { "epoch": 0.04641096012673762, "grad_norm": 0.5132386088371277, "learning_rate": 2.784471218206158e-06, "loss": 2.313191604614258, "step": 1040 }, { "epoch": 0.04663408974273155, "grad_norm": 0.490017294883728, "learning_rate": 2.7978580990629185e-06, "loss": 2.24790096282959, "step": 1045 }, { "epoch": 0.046857219358725485, "grad_norm": 0.5250972509384155, "learning_rate": 2.811244979919679e-06, "loss": 2.1867441177368163, "step": 1050 }, { "epoch": 0.04708034897471942, "grad_norm": 0.45831558108329773, "learning_rate": 2.824631860776439e-06, "loss": 2.379189300537109, "step": 1055 }, { "epoch": 0.04730347859071334, "grad_norm": 0.5892948508262634, "learning_rate": 2.8380187416331998e-06, "loss": 2.2735191345214845, "step": 1060 }, { "epoch": 0.047526608206707274, "grad_norm": 0.5409689545631409, "learning_rate": 2.85140562248996e-06, "loss": 2.3979333877563476, "step": 1065 }, { "epoch": 0.047749737822701206, "grad_norm": 0.5346503257751465, "learning_rate": 2.8647925033467203e-06, "loss": 2.4819768905639648, "step": 1070 }, { "epoch": 0.04797286743869514, "grad_norm": 0.6150287985801697, "learning_rate": 2.8781793842034806e-06, "loss": 2.5218339920043946, "step": 1075 }, { "epoch": 0.04819599705468907, "grad_norm": 0.4347957968711853, "learning_rate": 2.891566265060241e-06, "loss": 2.1929386138916014, "step": 1080 }, { "epoch": 0.048419126670683, "grad_norm": 0.5238802433013916, "learning_rate": 2.9049531459170016e-06, "loss": 2.215384292602539, "step": 1085 }, { "epoch": 0.048642256286676934, "grad_norm": 0.5371809601783752, "learning_rate": 2.918340026773762e-06, "loss": 2.4019641876220703, "step": 1090 }, { "epoch": 0.04886538590267086, "grad_norm": 0.5125853419303894, "learning_rate": 2.931726907630522e-06, "loss": 2.3410377502441406, "step": 1095 }, { "epoch": 0.04908851551866479, "grad_norm": 0.5087509751319885, "learning_rate": 2.9451137884872824e-06, "loss": 2.3634616851806642, "step": 1100 }, { "epoch": 0.04931164513465872, "grad_norm": 0.5605607628822327, "learning_rate": 2.9585006693440427e-06, "loss": 2.239984130859375, "step": 1105 }, { "epoch": 0.049534774750652655, "grad_norm": 0.6371492743492126, "learning_rate": 2.9718875502008034e-06, "loss": 2.499210739135742, "step": 1110 }, { "epoch": 0.04975790436664659, "grad_norm": 0.5526962280273438, "learning_rate": 2.9852744310575636e-06, "loss": 2.433857536315918, "step": 1115 }, { "epoch": 0.04998103398264052, "grad_norm": 0.5811582803726196, "learning_rate": 2.998661311914324e-06, "loss": 2.372517967224121, "step": 1120 }, { "epoch": 0.050204163598634444, "grad_norm": 0.47710540890693665, "learning_rate": 3.0120481927710846e-06, "loss": 2.441071891784668, "step": 1125 }, { "epoch": 0.050427293214628376, "grad_norm": 0.49463921785354614, "learning_rate": 3.0254350736278445e-06, "loss": 2.218737030029297, "step": 1130 }, { "epoch": 0.05065042283062231, "grad_norm": 0.7574100494384766, "learning_rate": 3.0388219544846056e-06, "loss": 2.1490055084228517, "step": 1135 }, { "epoch": 0.05087355244661624, "grad_norm": 0.5574637651443481, "learning_rate": 3.0522088353413654e-06, "loss": 2.3484785079956056, "step": 1140 }, { "epoch": 0.05109668206261017, "grad_norm": 0.5128051042556763, "learning_rate": 3.0655957161981257e-06, "loss": 2.2358184814453126, "step": 1145 }, { "epoch": 0.0513198116786041, "grad_norm": 0.4747551679611206, "learning_rate": 3.0789825970548864e-06, "loss": 2.2787382125854494, "step": 1150 }, { "epoch": 0.051542941294598035, "grad_norm": 0.4365915358066559, "learning_rate": 3.0923694779116467e-06, "loss": 2.327534484863281, "step": 1155 }, { "epoch": 0.05176607091059196, "grad_norm": 0.6433975696563721, "learning_rate": 3.1057563587684074e-06, "loss": 2.288431167602539, "step": 1160 }, { "epoch": 0.05198920052658589, "grad_norm": 0.6102743744850159, "learning_rate": 3.1191432396251672e-06, "loss": 2.24114933013916, "step": 1165 }, { "epoch": 0.052212330142579824, "grad_norm": 0.5493948459625244, "learning_rate": 3.132530120481928e-06, "loss": 2.317499542236328, "step": 1170 }, { "epoch": 0.052435459758573756, "grad_norm": 0.5411773324012756, "learning_rate": 3.145917001338688e-06, "loss": 2.1778676986694334, "step": 1175 }, { "epoch": 0.05265858937456769, "grad_norm": 0.5491811633110046, "learning_rate": 3.1593038821954485e-06, "loss": 2.260644721984863, "step": 1180 }, { "epoch": 0.05288171899056162, "grad_norm": 0.5138102769851685, "learning_rate": 3.172690763052209e-06, "loss": 2.2743204116821287, "step": 1185 }, { "epoch": 0.053104848606555545, "grad_norm": 0.5310954451560974, "learning_rate": 3.186077643908969e-06, "loss": 2.257237434387207, "step": 1190 }, { "epoch": 0.05332797822254948, "grad_norm": 0.5704927444458008, "learning_rate": 3.1994645247657297e-06, "loss": 2.0418323516845702, "step": 1195 }, { "epoch": 0.05355110783854341, "grad_norm": 0.5295515656471252, "learning_rate": 3.21285140562249e-06, "loss": 2.282415771484375, "step": 1200 }, { "epoch": 0.05377423745453734, "grad_norm": 0.6476492881774902, "learning_rate": 3.2262382864792503e-06, "loss": 2.306402015686035, "step": 1205 }, { "epoch": 0.05399736707053127, "grad_norm": 0.6115656495094299, "learning_rate": 3.239625167336011e-06, "loss": 2.1535377502441406, "step": 1210 }, { "epoch": 0.054220496686525205, "grad_norm": 0.5802296996116638, "learning_rate": 3.253012048192771e-06, "loss": 2.2679920196533203, "step": 1215 }, { "epoch": 0.05444362630251914, "grad_norm": 0.5182514190673828, "learning_rate": 3.266398929049532e-06, "loss": 2.3105010986328125, "step": 1220 }, { "epoch": 0.05466675591851306, "grad_norm": 0.6288827657699585, "learning_rate": 3.279785809906292e-06, "loss": 2.0794536590576174, "step": 1225 }, { "epoch": 0.054889885534506994, "grad_norm": 0.7542652487754822, "learning_rate": 3.2931726907630525e-06, "loss": 2.3161176681518554, "step": 1230 }, { "epoch": 0.055113015150500926, "grad_norm": 0.7416955232620239, "learning_rate": 3.3065595716198128e-06, "loss": 2.279362106323242, "step": 1235 }, { "epoch": 0.05533614476649486, "grad_norm": 0.5682216882705688, "learning_rate": 3.3199464524765726e-06, "loss": 2.2426963806152345, "step": 1240 }, { "epoch": 0.05555927438248879, "grad_norm": 0.6715278625488281, "learning_rate": 3.3333333333333337e-06, "loss": 2.1546775817871096, "step": 1245 }, { "epoch": 0.05578240399848272, "grad_norm": 0.5482577085494995, "learning_rate": 3.3467202141900936e-06, "loss": 2.1484096527099608, "step": 1250 }, { "epoch": 0.056005533614476646, "grad_norm": 0.6009266972541809, "learning_rate": 3.3601070950468543e-06, "loss": 2.221162796020508, "step": 1255 }, { "epoch": 0.05622866323047058, "grad_norm": 0.6640043258666992, "learning_rate": 3.3734939759036146e-06, "loss": 2.338100242614746, "step": 1260 }, { "epoch": 0.05645179284646451, "grad_norm": 0.6514598727226257, "learning_rate": 3.3868808567603744e-06, "loss": 2.2825199127197267, "step": 1265 }, { "epoch": 0.05667492246245844, "grad_norm": 0.6053763031959534, "learning_rate": 3.4002677376171355e-06, "loss": 2.280612754821777, "step": 1270 }, { "epoch": 0.056898052078452374, "grad_norm": 0.8854132294654846, "learning_rate": 3.4136546184738954e-06, "loss": 2.3481237411499025, "step": 1275 }, { "epoch": 0.057121181694446306, "grad_norm": 0.6433872580528259, "learning_rate": 3.427041499330656e-06, "loss": 2.2561552047729494, "step": 1280 }, { "epoch": 0.05734431131044024, "grad_norm": 0.709402859210968, "learning_rate": 3.4404283801874164e-06, "loss": 2.264519691467285, "step": 1285 }, { "epoch": 0.05756744092643416, "grad_norm": 0.5431029796600342, "learning_rate": 3.453815261044177e-06, "loss": 2.163587760925293, "step": 1290 }, { "epoch": 0.057790570542428095, "grad_norm": 0.7146894931793213, "learning_rate": 3.4672021419009373e-06, "loss": 2.2624624252319334, "step": 1295 }, { "epoch": 0.05801370015842203, "grad_norm": 0.6029372215270996, "learning_rate": 3.480589022757697e-06, "loss": 2.0209115982055663, "step": 1300 }, { "epoch": 0.05823682977441596, "grad_norm": 0.6099830865859985, "learning_rate": 3.493975903614458e-06, "loss": 2.177568054199219, "step": 1305 }, { "epoch": 0.05845995939040989, "grad_norm": 0.8421804308891296, "learning_rate": 3.507362784471218e-06, "loss": 2.1980712890625, "step": 1310 }, { "epoch": 0.05868308900640382, "grad_norm": 0.7849622368812561, "learning_rate": 3.520749665327979e-06, "loss": 2.2047344207763673, "step": 1315 }, { "epoch": 0.05890621862239775, "grad_norm": 0.5906243324279785, "learning_rate": 3.534136546184739e-06, "loss": 2.3746465682983398, "step": 1320 }, { "epoch": 0.05912934823839168, "grad_norm": 0.9020318388938904, "learning_rate": 3.547523427041499e-06, "loss": 2.2612667083740234, "step": 1325 }, { "epoch": 0.05935247785438561, "grad_norm": 0.6542187929153442, "learning_rate": 3.5609103078982597e-06, "loss": 2.2060800552368165, "step": 1330 }, { "epoch": 0.059575607470379544, "grad_norm": 0.7428557872772217, "learning_rate": 3.57429718875502e-06, "loss": 2.0877670288085937, "step": 1335 }, { "epoch": 0.059798737086373475, "grad_norm": 0.7468170523643494, "learning_rate": 3.5876840696117807e-06, "loss": 2.3160980224609373, "step": 1340 }, { "epoch": 0.06002186670236741, "grad_norm": 0.8021371960639954, "learning_rate": 3.601070950468541e-06, "loss": 2.1923311233520506, "step": 1345 }, { "epoch": 0.06024499631836134, "grad_norm": 0.6904175281524658, "learning_rate": 3.6144578313253016e-06, "loss": 2.1911346435546877, "step": 1350 }, { "epoch": 0.060468125934355264, "grad_norm": 0.7336163520812988, "learning_rate": 3.627844712182062e-06, "loss": 2.2488683700561523, "step": 1355 }, { "epoch": 0.060691255550349196, "grad_norm": 0.7925560474395752, "learning_rate": 3.6412315930388218e-06, "loss": 2.174961280822754, "step": 1360 }, { "epoch": 0.06091438516634313, "grad_norm": 0.8825748562812805, "learning_rate": 3.6546184738955825e-06, "loss": 2.012922668457031, "step": 1365 }, { "epoch": 0.06113751478233706, "grad_norm": 0.6754929423332214, "learning_rate": 3.6680053547523427e-06, "loss": 2.216470146179199, "step": 1370 }, { "epoch": 0.06136064439833099, "grad_norm": 0.6616571545600891, "learning_rate": 3.6813922356091034e-06, "loss": 2.067160415649414, "step": 1375 }, { "epoch": 0.061583774014324924, "grad_norm": 0.7033371925354004, "learning_rate": 3.6947791164658637e-06, "loss": 2.2197628021240234, "step": 1380 }, { "epoch": 0.06180690363031885, "grad_norm": 0.7256376147270203, "learning_rate": 3.7081659973226235e-06, "loss": 2.2435529708862303, "step": 1385 }, { "epoch": 0.06203003324631278, "grad_norm": 0.6585844159126282, "learning_rate": 3.7215528781793842e-06, "loss": 2.113719177246094, "step": 1390 }, { "epoch": 0.06225316286230671, "grad_norm": 0.7192341685295105, "learning_rate": 3.7349397590361445e-06, "loss": 2.2271860122680662, "step": 1395 }, { "epoch": 0.062476292478300645, "grad_norm": 0.7313777804374695, "learning_rate": 3.7483266398929052e-06, "loss": 2.312770462036133, "step": 1400 }, { "epoch": 0.06269942209429458, "grad_norm": 0.7106947302818298, "learning_rate": 3.7617135207496655e-06, "loss": 2.041255760192871, "step": 1405 }, { "epoch": 0.0629225517102885, "grad_norm": 0.7265296578407288, "learning_rate": 3.775100401606426e-06, "loss": 2.2530691146850588, "step": 1410 }, { "epoch": 0.06314568132628244, "grad_norm": 0.7085316777229309, "learning_rate": 3.788487282463186e-06, "loss": 2.03436222076416, "step": 1415 }, { "epoch": 0.06336881094227637, "grad_norm": 0.7872986793518066, "learning_rate": 3.8018741633199463e-06, "loss": 2.0966476440429687, "step": 1420 }, { "epoch": 0.0635919405582703, "grad_norm": 0.8961313366889954, "learning_rate": 3.8152610441767074e-06, "loss": 2.1432809829711914, "step": 1425 }, { "epoch": 0.06381507017426423, "grad_norm": 0.7127321362495422, "learning_rate": 3.828647925033467e-06, "loss": 2.2022947311401366, "step": 1430 }, { "epoch": 0.06403819979025815, "grad_norm": 0.8228124976158142, "learning_rate": 3.842034805890228e-06, "loss": 2.39965705871582, "step": 1435 }, { "epoch": 0.0642613294062521, "grad_norm": 0.775909960269928, "learning_rate": 3.855421686746988e-06, "loss": 2.1313138961791993, "step": 1440 }, { "epoch": 0.06448445902224602, "grad_norm": 0.7945120334625244, "learning_rate": 3.8688085676037485e-06, "loss": 2.074580955505371, "step": 1445 }, { "epoch": 0.06470758863823996, "grad_norm": 1.014404535293579, "learning_rate": 3.882195448460509e-06, "loss": 2.26824893951416, "step": 1450 }, { "epoch": 0.06493071825423388, "grad_norm": 0.986544132232666, "learning_rate": 3.895582329317269e-06, "loss": 2.207651138305664, "step": 1455 }, { "epoch": 0.06515384787022782, "grad_norm": 0.7337960004806519, "learning_rate": 3.90896921017403e-06, "loss": 2.2294179916381838, "step": 1460 }, { "epoch": 0.06537697748622175, "grad_norm": 1.2697299718856812, "learning_rate": 3.92235609103079e-06, "loss": 2.114596366882324, "step": 1465 }, { "epoch": 0.06560010710221567, "grad_norm": 0.8556989431381226, "learning_rate": 3.93574297188755e-06, "loss": 2.21254940032959, "step": 1470 }, { "epoch": 0.06582323671820961, "grad_norm": 0.9477710723876953, "learning_rate": 3.949129852744311e-06, "loss": 1.977022361755371, "step": 1475 }, { "epoch": 0.06604636633420354, "grad_norm": 1.7189115285873413, "learning_rate": 3.962516733601071e-06, "loss": 2.012314224243164, "step": 1480 }, { "epoch": 0.06626949595019747, "grad_norm": 0.826131284236908, "learning_rate": 3.975903614457832e-06, "loss": 2.193516731262207, "step": 1485 }, { "epoch": 0.0664926255661914, "grad_norm": 1.0023162364959717, "learning_rate": 3.9892904953145914e-06, "loss": 2.2499979019165037, "step": 1490 }, { "epoch": 0.06671575518218534, "grad_norm": 0.7376646399497986, "learning_rate": 4.002677376171352e-06, "loss": 2.1386489868164062, "step": 1495 }, { "epoch": 0.06693888479817926, "grad_norm": 0.7226489782333374, "learning_rate": 4.016064257028113e-06, "loss": 2.11199836730957, "step": 1500 }, { "epoch": 0.06716201441417319, "grad_norm": 0.8427999019622803, "learning_rate": 4.029451137884873e-06, "loss": 2.034942626953125, "step": 1505 }, { "epoch": 0.06738514403016713, "grad_norm": 0.79557204246521, "learning_rate": 4.042838018741633e-06, "loss": 2.145208549499512, "step": 1510 }, { "epoch": 0.06760827364616105, "grad_norm": 0.7423805594444275, "learning_rate": 4.056224899598393e-06, "loss": 2.2282032012939452, "step": 1515 }, { "epoch": 0.06783140326215499, "grad_norm": 0.8701015114784241, "learning_rate": 4.069611780455154e-06, "loss": 2.0600002288818358, "step": 1520 }, { "epoch": 0.06805453287814892, "grad_norm": 0.9967679977416992, "learning_rate": 4.082998661311915e-06, "loss": 2.20194034576416, "step": 1525 }, { "epoch": 0.06827766249414285, "grad_norm": 1.0514639616012573, "learning_rate": 4.096385542168675e-06, "loss": 2.250165557861328, "step": 1530 }, { "epoch": 0.06850079211013678, "grad_norm": 0.7402302026748657, "learning_rate": 4.109772423025435e-06, "loss": 2.0741525650024415, "step": 1535 }, { "epoch": 0.0687239217261307, "grad_norm": 1.0764678716659546, "learning_rate": 4.123159303882195e-06, "loss": 2.068874549865723, "step": 1540 }, { "epoch": 0.06894705134212464, "grad_norm": 0.9718510508537292, "learning_rate": 4.136546184738956e-06, "loss": 1.9750284194946288, "step": 1545 }, { "epoch": 0.06917018095811857, "grad_norm": 0.7534152269363403, "learning_rate": 4.149933065595716e-06, "loss": 2.1882530212402345, "step": 1550 }, { "epoch": 0.06939331057411251, "grad_norm": 0.8103408813476562, "learning_rate": 4.163319946452477e-06, "loss": 2.076101875305176, "step": 1555 }, { "epoch": 0.06961644019010643, "grad_norm": 0.7058753967285156, "learning_rate": 4.176706827309237e-06, "loss": 2.1349681854248046, "step": 1560 }, { "epoch": 0.06983956980610036, "grad_norm": 0.664568305015564, "learning_rate": 4.190093708165997e-06, "loss": 2.1840320587158204, "step": 1565 }, { "epoch": 0.0700626994220943, "grad_norm": 0.9274365901947021, "learning_rate": 4.2034805890227575e-06, "loss": 2.0162349700927735, "step": 1570 }, { "epoch": 0.07028582903808822, "grad_norm": 0.806480348110199, "learning_rate": 4.216867469879518e-06, "loss": 2.1628509521484376, "step": 1575 }, { "epoch": 0.07050895865408216, "grad_norm": 0.9540587067604065, "learning_rate": 4.230254350736279e-06, "loss": 2.239594841003418, "step": 1580 }, { "epoch": 0.07073208827007609, "grad_norm": 0.8032299876213074, "learning_rate": 4.243641231593039e-06, "loss": 2.170275115966797, "step": 1585 }, { "epoch": 0.07095521788607002, "grad_norm": 0.9516310095787048, "learning_rate": 4.2570281124497995e-06, "loss": 2.035562515258789, "step": 1590 }, { "epoch": 0.07117834750206395, "grad_norm": 0.8803707361221313, "learning_rate": 4.270414993306559e-06, "loss": 2.1310283660888674, "step": 1595 }, { "epoch": 0.07140147711805787, "grad_norm": 1.022939682006836, "learning_rate": 4.28380187416332e-06, "loss": 2.1093021392822267, "step": 1600 }, { "epoch": 0.07162460673405181, "grad_norm": 0.8254362344741821, "learning_rate": 4.297188755020081e-06, "loss": 1.8178255081176757, "step": 1605 }, { "epoch": 0.07184773635004574, "grad_norm": 0.779062807559967, "learning_rate": 4.3105756358768406e-06, "loss": 2.261147880554199, "step": 1610 }, { "epoch": 0.07207086596603968, "grad_norm": 0.8894750475883484, "learning_rate": 4.323962516733601e-06, "loss": 2.0283786773681642, "step": 1615 }, { "epoch": 0.0722939955820336, "grad_norm": 0.9776813387870789, "learning_rate": 4.337349397590362e-06, "loss": 2.177045440673828, "step": 1620 }, { "epoch": 0.07251712519802754, "grad_norm": 1.2759140729904175, "learning_rate": 4.350736278447122e-06, "loss": 2.0527717590332033, "step": 1625 }, { "epoch": 0.07274025481402147, "grad_norm": 0.9953216314315796, "learning_rate": 4.3641231593038825e-06, "loss": 2.05969352722168, "step": 1630 }, { "epoch": 0.07296338443001539, "grad_norm": 0.8319527506828308, "learning_rate": 4.377510040160642e-06, "loss": 2.003727340698242, "step": 1635 }, { "epoch": 0.07318651404600933, "grad_norm": 0.8335089087486267, "learning_rate": 4.390896921017403e-06, "loss": 2.0173160552978517, "step": 1640 }, { "epoch": 0.07340964366200325, "grad_norm": 1.1071332693099976, "learning_rate": 4.404283801874164e-06, "loss": 2.1184307098388673, "step": 1645 }, { "epoch": 0.0736327732779972, "grad_norm": 1.2880672216415405, "learning_rate": 4.4176706827309244e-06, "loss": 2.0427175521850587, "step": 1650 }, { "epoch": 0.07385590289399112, "grad_norm": 1.1434367895126343, "learning_rate": 4.431057563587684e-06, "loss": 2.0279747009277345, "step": 1655 }, { "epoch": 0.07407903250998506, "grad_norm": 0.7727090716362, "learning_rate": 4.444444444444444e-06, "loss": 2.1546749114990233, "step": 1660 }, { "epoch": 0.07430216212597898, "grad_norm": 1.0890729427337646, "learning_rate": 4.457831325301205e-06, "loss": 2.1247718811035154, "step": 1665 }, { "epoch": 0.07452529174197291, "grad_norm": 1.0179014205932617, "learning_rate": 4.4712182061579655e-06, "loss": 2.146767234802246, "step": 1670 }, { "epoch": 0.07474842135796685, "grad_norm": 0.745704174041748, "learning_rate": 4.484605087014726e-06, "loss": 2.224173736572266, "step": 1675 }, { "epoch": 0.07497155097396077, "grad_norm": 0.771865963935852, "learning_rate": 4.497991967871486e-06, "loss": 2.2265625, "step": 1680 }, { "epoch": 0.07519468058995471, "grad_norm": 1.352388620376587, "learning_rate": 4.511378848728246e-06, "loss": 2.261865425109863, "step": 1685 }, { "epoch": 0.07541781020594863, "grad_norm": 0.8781198859214783, "learning_rate": 4.524765729585007e-06, "loss": 2.0096199035644533, "step": 1690 }, { "epoch": 0.07564093982194256, "grad_norm": 1.0159834623336792, "learning_rate": 4.538152610441767e-06, "loss": 2.0243366241455076, "step": 1695 }, { "epoch": 0.0758640694379365, "grad_norm": 0.8101127743721008, "learning_rate": 4.551539491298528e-06, "loss": 2.119925117492676, "step": 1700 }, { "epoch": 0.07608719905393042, "grad_norm": 1.0311274528503418, "learning_rate": 4.564926372155288e-06, "loss": 1.9274934768676757, "step": 1705 }, { "epoch": 0.07631032866992436, "grad_norm": 1.315838098526001, "learning_rate": 4.578313253012049e-06, "loss": 2.058757209777832, "step": 1710 }, { "epoch": 0.07653345828591829, "grad_norm": 1.016562819480896, "learning_rate": 4.5917001338688084e-06, "loss": 2.0910503387451174, "step": 1715 }, { "epoch": 0.07675658790191223, "grad_norm": 1.1127480268478394, "learning_rate": 4.605087014725569e-06, "loss": 2.0274993896484377, "step": 1720 }, { "epoch": 0.07697971751790615, "grad_norm": 0.7386570572853088, "learning_rate": 4.61847389558233e-06, "loss": 2.03863468170166, "step": 1725 }, { "epoch": 0.07720284713390008, "grad_norm": 0.9372514486312866, "learning_rate": 4.63186077643909e-06, "loss": 1.9647506713867187, "step": 1730 }, { "epoch": 0.07742597674989402, "grad_norm": 0.866057276725769, "learning_rate": 4.64524765729585e-06, "loss": 2.0213361740112306, "step": 1735 }, { "epoch": 0.07764910636588794, "grad_norm": 0.8000425100326538, "learning_rate": 4.65863453815261e-06, "loss": 2.009931755065918, "step": 1740 }, { "epoch": 0.07787223598188188, "grad_norm": 1.4270923137664795, "learning_rate": 4.672021419009371e-06, "loss": 2.0473087310791014, "step": 1745 }, { "epoch": 0.0780953655978758, "grad_norm": 0.7464487552642822, "learning_rate": 4.685408299866132e-06, "loss": 2.087724304199219, "step": 1750 }, { "epoch": 0.07831849521386974, "grad_norm": 1.0895863771438599, "learning_rate": 4.6987951807228915e-06, "loss": 2.0931013107299803, "step": 1755 }, { "epoch": 0.07854162482986367, "grad_norm": 1.2510039806365967, "learning_rate": 4.712182061579652e-06, "loss": 2.0668258666992188, "step": 1760 }, { "epoch": 0.0787647544458576, "grad_norm": 0.9064348340034485, "learning_rate": 4.725568942436412e-06, "loss": 1.9942918777465821, "step": 1765 }, { "epoch": 0.07898788406185153, "grad_norm": 1.0130219459533691, "learning_rate": 4.738955823293173e-06, "loss": 2.0953414916992186, "step": 1770 }, { "epoch": 0.07921101367784546, "grad_norm": 1.0138216018676758, "learning_rate": 4.7523427041499334e-06, "loss": 2.097544479370117, "step": 1775 }, { "epoch": 0.0794341432938394, "grad_norm": 0.9159922003746033, "learning_rate": 4.765729585006693e-06, "loss": 2.024347686767578, "step": 1780 }, { "epoch": 0.07965727290983332, "grad_norm": 0.6978763937950134, "learning_rate": 4.779116465863454e-06, "loss": 2.032564163208008, "step": 1785 }, { "epoch": 0.07988040252582726, "grad_norm": 0.9015496969223022, "learning_rate": 4.792503346720214e-06, "loss": 2.091436767578125, "step": 1790 }, { "epoch": 0.08010353214182118, "grad_norm": 0.8910471796989441, "learning_rate": 4.8058902275769745e-06, "loss": 2.095396041870117, "step": 1795 }, { "epoch": 0.08032666175781511, "grad_norm": 0.8550083637237549, "learning_rate": 4.819277108433735e-06, "loss": 2.061022186279297, "step": 1800 }, { "epoch": 0.08054979137380905, "grad_norm": 0.888156533241272, "learning_rate": 4.832663989290495e-06, "loss": 2.0070331573486326, "step": 1805 }, { "epoch": 0.08077292098980297, "grad_norm": 1.0453161001205444, "learning_rate": 4.846050870147256e-06, "loss": 2.0485980987548826, "step": 1810 }, { "epoch": 0.08099605060579691, "grad_norm": 0.835533857345581, "learning_rate": 4.859437751004016e-06, "loss": 2.0737443923950196, "step": 1815 }, { "epoch": 0.08121918022179084, "grad_norm": 1.3074698448181152, "learning_rate": 4.872824631860777e-06, "loss": 2.204619216918945, "step": 1820 }, { "epoch": 0.08144230983778476, "grad_norm": 0.952622652053833, "learning_rate": 4.886211512717537e-06, "loss": 1.8864990234375, "step": 1825 }, { "epoch": 0.0816654394537787, "grad_norm": 1.069469690322876, "learning_rate": 4.899598393574298e-06, "loss": 2.0791854858398438, "step": 1830 }, { "epoch": 0.08188856906977263, "grad_norm": 1.1299378871917725, "learning_rate": 4.9129852744310576e-06, "loss": 2.0667266845703125, "step": 1835 }, { "epoch": 0.08211169868576657, "grad_norm": 0.7093334794044495, "learning_rate": 4.926372155287817e-06, "loss": 1.9636056900024415, "step": 1840 }, { "epoch": 0.08233482830176049, "grad_norm": 1.1765010356903076, "learning_rate": 4.939759036144579e-06, "loss": 2.0537750244140627, "step": 1845 }, { "epoch": 0.08255795791775443, "grad_norm": 0.8516542315483093, "learning_rate": 4.953145917001339e-06, "loss": 2.1657791137695312, "step": 1850 }, { "epoch": 0.08278108753374835, "grad_norm": 1.0495437383651733, "learning_rate": 4.9665327978580995e-06, "loss": 2.1656772613525392, "step": 1855 }, { "epoch": 0.08300421714974228, "grad_norm": 0.8000126481056213, "learning_rate": 4.979919678714859e-06, "loss": 2.020354461669922, "step": 1860 }, { "epoch": 0.08322734676573622, "grad_norm": 1.2887593507766724, "learning_rate": 4.993306559571619e-06, "loss": 1.9938453674316405, "step": 1865 }, { "epoch": 0.08345047638173014, "grad_norm": 0.8419036865234375, "learning_rate": 5.006693440428381e-06, "loss": 2.0833532333374025, "step": 1870 }, { "epoch": 0.08367360599772408, "grad_norm": 0.9748121500015259, "learning_rate": 5.020080321285141e-06, "loss": 2.0065855026245116, "step": 1875 }, { "epoch": 0.08389673561371801, "grad_norm": 0.8467640280723572, "learning_rate": 5.033467202141901e-06, "loss": 2.2140008926391603, "step": 1880 }, { "epoch": 0.08411986522971195, "grad_norm": 0.8260283470153809, "learning_rate": 5.046854082998661e-06, "loss": 1.973113441467285, "step": 1885 }, { "epoch": 0.08434299484570587, "grad_norm": 1.2561461925506592, "learning_rate": 5.060240963855422e-06, "loss": 1.9699825286865233, "step": 1890 }, { "epoch": 0.0845661244616998, "grad_norm": 0.7021331191062927, "learning_rate": 5.0736278447121826e-06, "loss": 1.9100543975830078, "step": 1895 }, { "epoch": 0.08478925407769373, "grad_norm": 0.9285181760787964, "learning_rate": 5.087014725568942e-06, "loss": 2.100686454772949, "step": 1900 }, { "epoch": 0.08501238369368766, "grad_norm": 1.0512628555297852, "learning_rate": 5.100401606425703e-06, "loss": 2.092105484008789, "step": 1905 }, { "epoch": 0.0852355133096816, "grad_norm": 0.9993311166763306, "learning_rate": 5.113788487282463e-06, "loss": 2.178700256347656, "step": 1910 }, { "epoch": 0.08545864292567552, "grad_norm": 0.8007418513298035, "learning_rate": 5.127175368139224e-06, "loss": 2.061174774169922, "step": 1915 }, { "epoch": 0.08568177254166946, "grad_norm": 0.9044376611709595, "learning_rate": 5.140562248995984e-06, "loss": 1.905235481262207, "step": 1920 }, { "epoch": 0.08590490215766339, "grad_norm": 1.1532313823699951, "learning_rate": 5.153949129852744e-06, "loss": 2.0893394470214846, "step": 1925 }, { "epoch": 0.08612803177365731, "grad_norm": 1.0485821962356567, "learning_rate": 5.167336010709505e-06, "loss": 1.9315324783325196, "step": 1930 }, { "epoch": 0.08635116138965125, "grad_norm": 1.1166177988052368, "learning_rate": 5.180722891566265e-06, "loss": 1.9648147583007813, "step": 1935 }, { "epoch": 0.08657429100564518, "grad_norm": 1.4419307708740234, "learning_rate": 5.1941097724230254e-06, "loss": 2.0460626602172853, "step": 1940 }, { "epoch": 0.08679742062163912, "grad_norm": 1.0400481224060059, "learning_rate": 5.207496653279786e-06, "loss": 2.034720230102539, "step": 1945 }, { "epoch": 0.08702055023763304, "grad_norm": 1.041756510734558, "learning_rate": 5.220883534136547e-06, "loss": 1.9372877120971679, "step": 1950 }, { "epoch": 0.08724367985362697, "grad_norm": 0.9398750066757202, "learning_rate": 5.234270414993307e-06, "loss": 1.9075811386108399, "step": 1955 }, { "epoch": 0.0874668094696209, "grad_norm": 0.9996817708015442, "learning_rate": 5.2476572958500665e-06, "loss": 2.04040412902832, "step": 1960 }, { "epoch": 0.08768993908561483, "grad_norm": 1.1434073448181152, "learning_rate": 5.261044176706827e-06, "loss": 2.05865535736084, "step": 1965 }, { "epoch": 0.08791306870160877, "grad_norm": 1.1526947021484375, "learning_rate": 5.274431057563588e-06, "loss": 2.0584930419921874, "step": 1970 }, { "epoch": 0.0881361983176027, "grad_norm": 1.1225361824035645, "learning_rate": 5.287817938420349e-06, "loss": 1.869145965576172, "step": 1975 }, { "epoch": 0.08835932793359663, "grad_norm": 1.0828170776367188, "learning_rate": 5.3012048192771085e-06, "loss": 2.0549264907836915, "step": 1980 }, { "epoch": 0.08858245754959056, "grad_norm": 0.9480776786804199, "learning_rate": 5.314591700133868e-06, "loss": 1.987673568725586, "step": 1985 }, { "epoch": 0.08880558716558448, "grad_norm": 0.9799356460571289, "learning_rate": 5.327978580990629e-06, "loss": 1.9657482147216796, "step": 1990 }, { "epoch": 0.08902871678157842, "grad_norm": 1.1955480575561523, "learning_rate": 5.34136546184739e-06, "loss": 1.9616188049316405, "step": 1995 }, { "epoch": 0.08925184639757235, "grad_norm": 0.9528502225875854, "learning_rate": 5.3547523427041504e-06, "loss": 2.1342418670654295, "step": 2000 }, { "epoch": 0.08947497601356628, "grad_norm": 1.1190009117126465, "learning_rate": 5.36813922356091e-06, "loss": 2.1064796447753906, "step": 2005 }, { "epoch": 0.08969810562956021, "grad_norm": 1.0113600492477417, "learning_rate": 5.381526104417671e-06, "loss": 1.9123802185058594, "step": 2010 }, { "epoch": 0.08992123524555415, "grad_norm": 0.9729359745979309, "learning_rate": 5.394912985274431e-06, "loss": 1.9950223922729493, "step": 2015 }, { "epoch": 0.09014436486154807, "grad_norm": 1.035287618637085, "learning_rate": 5.4082998661311915e-06, "loss": 1.968986701965332, "step": 2020 }, { "epoch": 0.090367494477542, "grad_norm": 0.7954384684562683, "learning_rate": 5.421686746987952e-06, "loss": 2.075417327880859, "step": 2025 }, { "epoch": 0.09059062409353594, "grad_norm": 1.2406619787216187, "learning_rate": 5.435073627844712e-06, "loss": 2.097009468078613, "step": 2030 }, { "epoch": 0.09081375370952986, "grad_norm": 0.8746753931045532, "learning_rate": 5.448460508701473e-06, "loss": 2.04738826751709, "step": 2035 }, { "epoch": 0.0910368833255238, "grad_norm": 1.137431025505066, "learning_rate": 5.461847389558233e-06, "loss": 1.9636581420898438, "step": 2040 }, { "epoch": 0.09126001294151773, "grad_norm": 1.4577034711837769, "learning_rate": 5.475234270414993e-06, "loss": 1.8939685821533203, "step": 2045 }, { "epoch": 0.09148314255751167, "grad_norm": 1.3480346202850342, "learning_rate": 5.488621151271754e-06, "loss": 1.9858078002929687, "step": 2050 }, { "epoch": 0.09170627217350559, "grad_norm": 0.8342795372009277, "learning_rate": 5.502008032128514e-06, "loss": 1.7992578506469727, "step": 2055 }, { "epoch": 0.09192940178949952, "grad_norm": 1.1130149364471436, "learning_rate": 5.5153949129852746e-06, "loss": 2.0617441177368163, "step": 2060 }, { "epoch": 0.09215253140549345, "grad_norm": 1.0582903623580933, "learning_rate": 5.528781793842035e-06, "loss": 2.081201934814453, "step": 2065 }, { "epoch": 0.09237566102148738, "grad_norm": 0.8622370958328247, "learning_rate": 5.542168674698796e-06, "loss": 2.184007263183594, "step": 2070 }, { "epoch": 0.09259879063748132, "grad_norm": 1.247605800628662, "learning_rate": 5.555555555555556e-06, "loss": 2.2668291091918946, "step": 2075 }, { "epoch": 0.09282192025347524, "grad_norm": 0.9048439860343933, "learning_rate": 5.568942436412316e-06, "loss": 2.0036815643310546, "step": 2080 }, { "epoch": 0.09304504986946917, "grad_norm": 1.5477705001831055, "learning_rate": 5.582329317269076e-06, "loss": 2.1254167556762695, "step": 2085 }, { "epoch": 0.0932681794854631, "grad_norm": 0.898073673248291, "learning_rate": 5.595716198125837e-06, "loss": 2.0748970031738283, "step": 2090 }, { "epoch": 0.09349130910145703, "grad_norm": 0.976369321346283, "learning_rate": 5.609103078982598e-06, "loss": 1.9387639999389648, "step": 2095 }, { "epoch": 0.09371443871745097, "grad_norm": 1.229954719543457, "learning_rate": 5.622489959839358e-06, "loss": 1.9117536544799805, "step": 2100 }, { "epoch": 0.0939375683334449, "grad_norm": 1.0723400115966797, "learning_rate": 5.6358768406961175e-06, "loss": 1.935175323486328, "step": 2105 }, { "epoch": 0.09416069794943883, "grad_norm": 1.1824326515197754, "learning_rate": 5.649263721552878e-06, "loss": 2.126773452758789, "step": 2110 }, { "epoch": 0.09438382756543276, "grad_norm": 0.8257484436035156, "learning_rate": 5.662650602409639e-06, "loss": 2.01248836517334, "step": 2115 }, { "epoch": 0.09460695718142668, "grad_norm": 1.11427903175354, "learning_rate": 5.6760374832663996e-06, "loss": 2.1134754180908204, "step": 2120 }, { "epoch": 0.09483008679742062, "grad_norm": 1.5741039514541626, "learning_rate": 5.689424364123159e-06, "loss": 2.0824302673339843, "step": 2125 }, { "epoch": 0.09505321641341455, "grad_norm": 0.9258787631988525, "learning_rate": 5.70281124497992e-06, "loss": 1.9319812774658203, "step": 2130 }, { "epoch": 0.09527634602940849, "grad_norm": 1.09774911403656, "learning_rate": 5.71619812583668e-06, "loss": 1.9680803298950196, "step": 2135 }, { "epoch": 0.09549947564540241, "grad_norm": 0.9626898765563965, "learning_rate": 5.729585006693441e-06, "loss": 1.925799560546875, "step": 2140 }, { "epoch": 0.09572260526139635, "grad_norm": 0.9046334624290466, "learning_rate": 5.742971887550201e-06, "loss": 1.9070989608764648, "step": 2145 }, { "epoch": 0.09594573487739028, "grad_norm": 1.0075236558914185, "learning_rate": 5.756358768406961e-06, "loss": 1.9298274993896485, "step": 2150 }, { "epoch": 0.0961688644933842, "grad_norm": 0.8876293301582336, "learning_rate": 5.769745649263722e-06, "loss": 1.995884323120117, "step": 2155 }, { "epoch": 0.09639199410937814, "grad_norm": 0.9397478699684143, "learning_rate": 5.783132530120482e-06, "loss": 2.1033966064453127, "step": 2160 }, { "epoch": 0.09661512372537207, "grad_norm": 1.1887969970703125, "learning_rate": 5.7965194109772425e-06, "loss": 2.2321205139160156, "step": 2165 }, { "epoch": 0.096838253341366, "grad_norm": 0.9165719151496887, "learning_rate": 5.809906291834003e-06, "loss": 1.9332220077514648, "step": 2170 }, { "epoch": 0.09706138295735993, "grad_norm": 1.0620633363723755, "learning_rate": 5.823293172690763e-06, "loss": 1.942098617553711, "step": 2175 }, { "epoch": 0.09728451257335387, "grad_norm": 1.1249985694885254, "learning_rate": 5.836680053547524e-06, "loss": 2.004232406616211, "step": 2180 }, { "epoch": 0.09750764218934779, "grad_norm": 1.009824514389038, "learning_rate": 5.8500669344042836e-06, "loss": 2.060811424255371, "step": 2185 }, { "epoch": 0.09773077180534172, "grad_norm": 1.1177529096603394, "learning_rate": 5.863453815261044e-06, "loss": 1.9775056838989258, "step": 2190 }, { "epoch": 0.09795390142133566, "grad_norm": 0.8883247375488281, "learning_rate": 5.876840696117805e-06, "loss": 2.0491350173950194, "step": 2195 }, { "epoch": 0.09817703103732958, "grad_norm": 1.0781581401824951, "learning_rate": 5.890227576974565e-06, "loss": 2.130450439453125, "step": 2200 }, { "epoch": 0.09840016065332352, "grad_norm": 1.022125244140625, "learning_rate": 5.9036144578313255e-06, "loss": 1.9054424285888671, "step": 2205 }, { "epoch": 0.09862329026931745, "grad_norm": 1.5722389221191406, "learning_rate": 5.917001338688085e-06, "loss": 1.9462669372558594, "step": 2210 }, { "epoch": 0.09884641988531137, "grad_norm": 0.8767709732055664, "learning_rate": 5.930388219544846e-06, "loss": 1.9716548919677734, "step": 2215 }, { "epoch": 0.09906954950130531, "grad_norm": 0.8978337049484253, "learning_rate": 5.943775100401607e-06, "loss": 1.932182502746582, "step": 2220 }, { "epoch": 0.09929267911729923, "grad_norm": 1.1649725437164307, "learning_rate": 5.957161981258367e-06, "loss": 1.949960708618164, "step": 2225 }, { "epoch": 0.09951580873329317, "grad_norm": 0.9707964062690735, "learning_rate": 5.970548862115127e-06, "loss": 1.9415960311889648, "step": 2230 }, { "epoch": 0.0997389383492871, "grad_norm": 0.8075286149978638, "learning_rate": 5.983935742971887e-06, "loss": 1.9884584426879883, "step": 2235 }, { "epoch": 0.09996206796528104, "grad_norm": 0.919001579284668, "learning_rate": 5.997322623828648e-06, "loss": 1.9047340393066405, "step": 2240 }, { "epoch": 0.10018519758127496, "grad_norm": 1.1452890634536743, "learning_rate": 5.999999417648353e-06, "loss": 1.8911170959472656, "step": 2245 }, { "epoch": 0.10040832719726889, "grad_norm": 0.9075614213943481, "learning_rate": 5.999997051845177e-06, "loss": 1.9537294387817383, "step": 2250 }, { "epoch": 0.10063145681326283, "grad_norm": 0.9411978721618652, "learning_rate": 5.999992866194927e-06, "loss": 2.0128080368041994, "step": 2255 }, { "epoch": 0.10085458642925675, "grad_norm": 1.7404388189315796, "learning_rate": 5.999986860700142e-06, "loss": 1.9837495803833007, "step": 2260 }, { "epoch": 0.10107771604525069, "grad_norm": 0.990151047706604, "learning_rate": 5.999979035364464e-06, "loss": 1.9975740432739257, "step": 2265 }, { "epoch": 0.10130084566124462, "grad_norm": 1.042856216430664, "learning_rate": 5.9999693901926435e-06, "loss": 2.016056251525879, "step": 2270 }, { "epoch": 0.10152397527723855, "grad_norm": 0.8757387399673462, "learning_rate": 5.999957925190527e-06, "loss": 1.8990396499633788, "step": 2275 }, { "epoch": 0.10174710489323248, "grad_norm": 1.4858282804489136, "learning_rate": 5.999944640365074e-06, "loss": 2.0470705032348633, "step": 2280 }, { "epoch": 0.1019702345092264, "grad_norm": 1.0206114053726196, "learning_rate": 5.999929535724339e-06, "loss": 2.063922882080078, "step": 2285 }, { "epoch": 0.10219336412522034, "grad_norm": 1.2412887811660767, "learning_rate": 5.999912611277487e-06, "loss": 1.9137586593627929, "step": 2290 }, { "epoch": 0.10241649374121427, "grad_norm": 1.0442379713058472, "learning_rate": 5.999893867034784e-06, "loss": 1.9131647109985352, "step": 2295 }, { "epoch": 0.1026396233572082, "grad_norm": 1.0354846715927124, "learning_rate": 5.9998733030076015e-06, "loss": 1.943178367614746, "step": 2300 }, { "epoch": 0.10286275297320213, "grad_norm": 0.9171436429023743, "learning_rate": 5.999850919208413e-06, "loss": 1.9243003845214843, "step": 2305 }, { "epoch": 0.10308588258919607, "grad_norm": 1.3587474822998047, "learning_rate": 5.999826715650797e-06, "loss": 1.7801437377929688, "step": 2310 }, { "epoch": 0.10330901220519, "grad_norm": 1.6726102828979492, "learning_rate": 5.999800692349435e-06, "loss": 2.0044769287109374, "step": 2315 }, { "epoch": 0.10353214182118392, "grad_norm": 0.909557580947876, "learning_rate": 5.999772849320115e-06, "loss": 1.8790843963623047, "step": 2320 }, { "epoch": 0.10375527143717786, "grad_norm": 1.1021777391433716, "learning_rate": 5.9997431865797255e-06, "loss": 1.9645397186279296, "step": 2325 }, { "epoch": 0.10397840105317178, "grad_norm": 1.0416226387023926, "learning_rate": 5.999711704146261e-06, "loss": 1.9316511154174805, "step": 2330 }, { "epoch": 0.10420153066916572, "grad_norm": 1.014394998550415, "learning_rate": 5.99967840203882e-06, "loss": 1.9602357864379882, "step": 2335 }, { "epoch": 0.10442466028515965, "grad_norm": 1.0431203842163086, "learning_rate": 5.999643280277603e-06, "loss": 1.9662353515625, "step": 2340 }, { "epoch": 0.10464778990115357, "grad_norm": 0.9976010322570801, "learning_rate": 5.999606338883915e-06, "loss": 1.8156768798828125, "step": 2345 }, { "epoch": 0.10487091951714751, "grad_norm": 1.3542364835739136, "learning_rate": 5.999567577880167e-06, "loss": 1.9991931915283203, "step": 2350 }, { "epoch": 0.10509404913314144, "grad_norm": 0.9118264317512512, "learning_rate": 5.99952699728987e-06, "loss": 1.6772186279296875, "step": 2355 }, { "epoch": 0.10531717874913538, "grad_norm": 1.431563138961792, "learning_rate": 5.9994845971376435e-06, "loss": 2.0468074798583986, "step": 2360 }, { "epoch": 0.1055403083651293, "grad_norm": 0.8818404078483582, "learning_rate": 5.999440377449205e-06, "loss": 1.9281387329101562, "step": 2365 }, { "epoch": 0.10576343798112324, "grad_norm": 0.905672013759613, "learning_rate": 5.999394338251381e-06, "loss": 2.0013042449951173, "step": 2370 }, { "epoch": 0.10598656759711717, "grad_norm": 1.0039812326431274, "learning_rate": 5.999346479572099e-06, "loss": 2.0309635162353517, "step": 2375 }, { "epoch": 0.10620969721311109, "grad_norm": 1.1300805807113647, "learning_rate": 5.999296801440391e-06, "loss": 2.013144111633301, "step": 2380 }, { "epoch": 0.10643282682910503, "grad_norm": 1.052232027053833, "learning_rate": 5.999245303886392e-06, "loss": 2.089672660827637, "step": 2385 }, { "epoch": 0.10665595644509895, "grad_norm": 0.8320952653884888, "learning_rate": 5.999191986941342e-06, "loss": 2.0204065322875975, "step": 2390 }, { "epoch": 0.10687908606109289, "grad_norm": 0.8853346705436707, "learning_rate": 5.999136850637584e-06, "loss": 1.9383895874023438, "step": 2395 }, { "epoch": 0.10710221567708682, "grad_norm": 1.3795983791351318, "learning_rate": 5.999079895008564e-06, "loss": 1.808138084411621, "step": 2400 } ], "logging_steps": 5, "max_steps": 22409, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.2148094686056153e+19, "train_batch_size": 4, "trial_name": null, "trial_params": null }