{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 1465, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0034129692832764505, "grad_norm": 2.50749013823737, "learning_rate": 2.702702702702703e-06, "loss": 0.87, "num_tokens": 1933925.0, "step": 5 }, { "epoch": 0.006825938566552901, "grad_norm": 1.6188623255616226, "learning_rate": 6.081081081081082e-06, "loss": 0.8109, "num_tokens": 3759146.0, "step": 10 }, { "epoch": 0.010238907849829351, "grad_norm": 1.0123465050636375, "learning_rate": 9.45945945945946e-06, "loss": 0.7772, "num_tokens": 5644524.0, "step": 15 }, { "epoch": 0.013651877133105802, "grad_norm": 0.7086005802414309, "learning_rate": 1.2837837837837838e-05, "loss": 0.7221, "num_tokens": 7668808.0, "step": 20 }, { "epoch": 0.017064846416382253, "grad_norm": 0.5775312610735582, "learning_rate": 1.6216216216216218e-05, "loss": 0.6862, "num_tokens": 9513010.0, "step": 25 }, { "epoch": 0.020477815699658702, "grad_norm": 0.5015160747913217, "learning_rate": 1.9594594594594595e-05, "loss": 0.648, "num_tokens": 11368873.0, "step": 30 }, { "epoch": 0.023890784982935155, "grad_norm": 0.4755691629562603, "learning_rate": 2.2972972972972976e-05, "loss": 0.6475, "num_tokens": 13219424.0, "step": 35 }, { "epoch": 0.027303754266211604, "grad_norm": 0.45995404550990876, "learning_rate": 2.635135135135135e-05, "loss": 0.6258, "num_tokens": 15118063.0, "step": 40 }, { "epoch": 0.030716723549488054, "grad_norm": 0.4312148613152685, "learning_rate": 2.9729729729729733e-05, "loss": 0.6282, "num_tokens": 17042190.0, "step": 45 }, { "epoch": 0.034129692832764506, "grad_norm": 0.451819013423788, "learning_rate": 3.310810810810811e-05, "loss": 0.6123, "num_tokens": 18906839.0, "step": 50 }, { "epoch": 0.03754266211604096, "grad_norm": 0.5006689200573762, "learning_rate": 3.648648648648649e-05, "loss": 0.5998, "num_tokens": 20791105.0, "step": 55 }, { "epoch": 0.040955631399317405, "grad_norm": 0.46031777713747535, "learning_rate": 3.986486486486487e-05, "loss": 0.6081, "num_tokens": 22641755.0, "step": 60 }, { "epoch": 0.04436860068259386, "grad_norm": 0.46899875801322866, "learning_rate": 4.324324324324325e-05, "loss": 0.6093, "num_tokens": 24604914.0, "step": 65 }, { "epoch": 0.04778156996587031, "grad_norm": 0.6110895849073993, "learning_rate": 4.662162162162162e-05, "loss": 0.6024, "num_tokens": 26636629.0, "step": 70 }, { "epoch": 0.051194539249146756, "grad_norm": 0.6405781895008703, "learning_rate": 5e-05, "loss": 0.5701, "num_tokens": 28441260.0, "step": 75 }, { "epoch": 0.05460750853242321, "grad_norm": 0.581519178723128, "learning_rate": 4.999856539149453e-05, "loss": 0.5927, "num_tokens": 30417967.0, "step": 80 }, { "epoch": 0.05802047781569966, "grad_norm": 0.44619446982283184, "learning_rate": 4.99942617489205e-05, "loss": 0.5881, "num_tokens": 32351323.0, "step": 85 }, { "epoch": 0.06143344709897611, "grad_norm": 0.5141883933664836, "learning_rate": 4.998708962108165e-05, "loss": 0.5876, "num_tokens": 34231333.0, "step": 90 }, { "epoch": 0.06484641638225255, "grad_norm": 0.5192244585208947, "learning_rate": 4.9977049922573155e-05, "loss": 0.584, "num_tokens": 36068371.0, "step": 95 }, { "epoch": 0.06825938566552901, "grad_norm": 0.46267202773061866, "learning_rate": 4.9964143933664945e-05, "loss": 0.5962, "num_tokens": 37961424.0, "step": 100 }, { "epoch": 0.07167235494880546, "grad_norm": 0.551678204091152, "learning_rate": 4.99483733001385e-05, "loss": 0.5953, "num_tokens": 39846738.0, "step": 105 }, { "epoch": 0.07508532423208192, "grad_norm": 0.5571076806342042, "learning_rate": 4.9929740033076915e-05, "loss": 0.5899, "num_tokens": 41826860.0, "step": 110 }, { "epoch": 0.07849829351535836, "grad_norm": 0.5411739076594996, "learning_rate": 4.9908246508608506e-05, "loss": 0.5848, "num_tokens": 43675049.0, "step": 115 }, { "epoch": 0.08191126279863481, "grad_norm": 0.4485161126231456, "learning_rate": 4.9883895467603764e-05, "loss": 0.5773, "num_tokens": 45543403.0, "step": 120 }, { "epoch": 0.08532423208191127, "grad_norm": 0.4648271403988684, "learning_rate": 4.985669001532583e-05, "loss": 0.5904, "num_tokens": 47466657.0, "step": 125 }, { "epoch": 0.08873720136518772, "grad_norm": 0.4638975404274955, "learning_rate": 4.982663362103456e-05, "loss": 0.5818, "num_tokens": 49369486.0, "step": 130 }, { "epoch": 0.09215017064846416, "grad_norm": 0.533501406227831, "learning_rate": 4.979373011754409e-05, "loss": 0.5773, "num_tokens": 51193029.0, "step": 135 }, { "epoch": 0.09556313993174062, "grad_norm": 0.5456590197317953, "learning_rate": 4.975798370073404e-05, "loss": 0.5767, "num_tokens": 53010874.0, "step": 140 }, { "epoch": 0.09897610921501707, "grad_norm": 0.47396184664647756, "learning_rate": 4.971939892901451e-05, "loss": 0.5696, "num_tokens": 54919625.0, "step": 145 }, { "epoch": 0.10238907849829351, "grad_norm": 0.4121898187313886, "learning_rate": 4.9677980722744774e-05, "loss": 0.5881, "num_tokens": 56909889.0, "step": 150 }, { "epoch": 0.10580204778156997, "grad_norm": 0.4504445666388642, "learning_rate": 4.9633734363605805e-05, "loss": 0.5585, "num_tokens": 58759721.0, "step": 155 }, { "epoch": 0.10921501706484642, "grad_norm": 0.43511695023636615, "learning_rate": 4.9586665493926773e-05, "loss": 0.5655, "num_tokens": 60650570.0, "step": 160 }, { "epoch": 0.11262798634812286, "grad_norm": 0.4529892002607441, "learning_rate": 4.9536780115965514e-05, "loss": 0.5499, "num_tokens": 62662896.0, "step": 165 }, { "epoch": 0.11604095563139932, "grad_norm": 0.4983273663357267, "learning_rate": 4.9484084591143154e-05, "loss": 0.562, "num_tokens": 64564660.0, "step": 170 }, { "epoch": 0.11945392491467577, "grad_norm": 0.4483226301683336, "learning_rate": 4.942858563923285e-05, "loss": 0.5669, "num_tokens": 66485190.0, "step": 175 }, { "epoch": 0.12286689419795221, "grad_norm": 0.49399353340120483, "learning_rate": 4.9370290337502924e-05, "loss": 0.5805, "num_tokens": 68426882.0, "step": 180 }, { "epoch": 0.12627986348122866, "grad_norm": 0.40977009590857616, "learning_rate": 4.930920611981431e-05, "loss": 0.5699, "num_tokens": 70332714.0, "step": 185 }, { "epoch": 0.1296928327645051, "grad_norm": 0.542445950494901, "learning_rate": 4.9245340775672634e-05, "loss": 0.5691, "num_tokens": 72252819.0, "step": 190 }, { "epoch": 0.13310580204778158, "grad_norm": 0.4622793694221624, "learning_rate": 4.917870244923486e-05, "loss": 0.5439, "num_tokens": 74221358.0, "step": 195 }, { "epoch": 0.13651877133105803, "grad_norm": 0.5759125508857202, "learning_rate": 4.910929963827078e-05, "loss": 0.5683, "num_tokens": 76160914.0, "step": 200 }, { "epoch": 0.13993174061433447, "grad_norm": 0.39779641276955163, "learning_rate": 4.903714119307929e-05, "loss": 0.5564, "num_tokens": 78170639.0, "step": 205 }, { "epoch": 0.14334470989761092, "grad_norm": 0.4401620614285602, "learning_rate": 4.8962236315359896e-05, "loss": 0.5603, "num_tokens": 80152955.0, "step": 210 }, { "epoch": 0.14675767918088736, "grad_norm": 0.4752332395853894, "learning_rate": 4.8884594557039224e-05, "loss": 0.5587, "num_tokens": 81973764.0, "step": 215 }, { "epoch": 0.15017064846416384, "grad_norm": 0.4004170259025375, "learning_rate": 4.8804225819052994e-05, "loss": 0.5664, "num_tokens": 83901702.0, "step": 220 }, { "epoch": 0.15358361774744028, "grad_norm": 0.5178389649751411, "learning_rate": 4.872114035008345e-05, "loss": 0.5416, "num_tokens": 85675897.0, "step": 225 }, { "epoch": 0.15699658703071673, "grad_norm": 0.5112947588191477, "learning_rate": 4.863534874525241e-05, "loss": 0.582, "num_tokens": 87691018.0, "step": 230 }, { "epoch": 0.16040955631399317, "grad_norm": 0.5199172012585387, "learning_rate": 4.854686194477017e-05, "loss": 0.5514, "num_tokens": 89579835.0, "step": 235 }, { "epoch": 0.16382252559726962, "grad_norm": 0.5079655450026499, "learning_rate": 4.845569123254044e-05, "loss": 0.5503, "num_tokens": 91542428.0, "step": 240 }, { "epoch": 0.16723549488054607, "grad_norm": 0.4384736028856376, "learning_rate": 4.836184823472136e-05, "loss": 0.5511, "num_tokens": 93492833.0, "step": 245 }, { "epoch": 0.17064846416382254, "grad_norm": 0.4196228017966155, "learning_rate": 4.8265344918242974e-05, "loss": 0.565, "num_tokens": 95438170.0, "step": 250 }, { "epoch": 0.17406143344709898, "grad_norm": 0.5223124856503392, "learning_rate": 4.8166193589281154e-05, "loss": 0.5565, "num_tokens": 97303144.0, "step": 255 }, { "epoch": 0.17747440273037543, "grad_norm": 0.5446858466087424, "learning_rate": 4.806440689168833e-05, "loss": 0.5504, "num_tokens": 99383692.0, "step": 260 }, { "epoch": 0.18088737201365188, "grad_norm": 0.5859886624013145, "learning_rate": 4.795999780538113e-05, "loss": 0.5555, "num_tokens": 101357212.0, "step": 265 }, { "epoch": 0.18430034129692832, "grad_norm": 0.5022040294550851, "learning_rate": 4.785297964468519e-05, "loss": 0.5494, "num_tokens": 103223537.0, "step": 270 }, { "epoch": 0.18771331058020477, "grad_norm": 0.43644085987482656, "learning_rate": 4.7743366056637266e-05, "loss": 0.5607, "num_tokens": 105044617.0, "step": 275 }, { "epoch": 0.19112627986348124, "grad_norm": 0.4171727799014969, "learning_rate": 4.7631171019244974e-05, "loss": 0.5691, "num_tokens": 106901687.0, "step": 280 }, { "epoch": 0.1945392491467577, "grad_norm": 0.41261747021762124, "learning_rate": 4.7516408839704316e-05, "loss": 0.5251, "num_tokens": 108801293.0, "step": 285 }, { "epoch": 0.19795221843003413, "grad_norm": 0.40680116615519046, "learning_rate": 4.739909415257518e-05, "loss": 0.5489, "num_tokens": 110840758.0, "step": 290 }, { "epoch": 0.20136518771331058, "grad_norm": 0.3913282639543854, "learning_rate": 4.727924191791518e-05, "loss": 0.552, "num_tokens": 112713541.0, "step": 295 }, { "epoch": 0.20477815699658702, "grad_norm": 0.4331430104517169, "learning_rate": 4.7156867419371866e-05, "loss": 0.5525, "num_tokens": 114521504.0, "step": 300 }, { "epoch": 0.20819112627986347, "grad_norm": 0.4119990823055153, "learning_rate": 4.703198626223383e-05, "loss": 0.5378, "num_tokens": 116445027.0, "step": 305 }, { "epoch": 0.21160409556313994, "grad_norm": 0.3887498249238803, "learning_rate": 4.6904614371440654e-05, "loss": 0.5374, "num_tokens": 118445759.0, "step": 310 }, { "epoch": 0.2150170648464164, "grad_norm": 0.4309808604161364, "learning_rate": 4.677476798955213e-05, "loss": 0.5441, "num_tokens": 120303683.0, "step": 315 }, { "epoch": 0.21843003412969283, "grad_norm": 0.42995753849368423, "learning_rate": 4.664246367467707e-05, "loss": 0.5485, "num_tokens": 122311693.0, "step": 320 }, { "epoch": 0.22184300341296928, "grad_norm": 0.3731200866529834, "learning_rate": 4.6507718298361716e-05, "loss": 0.5508, "num_tokens": 124240526.0, "step": 325 }, { "epoch": 0.22525597269624573, "grad_norm": 0.39306487790311306, "learning_rate": 4.637054904343833e-05, "loss": 0.5465, "num_tokens": 126094524.0, "step": 330 }, { "epoch": 0.22866894197952217, "grad_norm": 0.42607532993380426, "learning_rate": 4.623097340183401e-05, "loss": 0.5296, "num_tokens": 128075174.0, "step": 335 }, { "epoch": 0.23208191126279865, "grad_norm": 0.42217543753738046, "learning_rate": 4.60890091723401e-05, "loss": 0.5547, "num_tokens": 129971056.0, "step": 340 }, { "epoch": 0.2354948805460751, "grad_norm": 0.4373865514759956, "learning_rate": 4.5944674458342473e-05, "loss": 0.5535, "num_tokens": 131847305.0, "step": 345 }, { "epoch": 0.23890784982935154, "grad_norm": 0.4780257445539072, "learning_rate": 4.579798766551298e-05, "loss": 0.5603, "num_tokens": 133765982.0, "step": 350 }, { "epoch": 0.24232081911262798, "grad_norm": 0.39373360953599906, "learning_rate": 4.564896749946234e-05, "loss": 0.5332, "num_tokens": 135541389.0, "step": 355 }, { "epoch": 0.24573378839590443, "grad_norm": 0.48408274088956066, "learning_rate": 4.549763296335481e-05, "loss": 0.5417, "num_tokens": 137522281.0, "step": 360 }, { "epoch": 0.24914675767918087, "grad_norm": 0.38885053329494734, "learning_rate": 4.534400335548484e-05, "loss": 0.5302, "num_tokens": 139412841.0, "step": 365 }, { "epoch": 0.2525597269624573, "grad_norm": 0.4284079243560024, "learning_rate": 4.518809826681616e-05, "loss": 0.5338, "num_tokens": 141196084.0, "step": 370 }, { "epoch": 0.25597269624573377, "grad_norm": 0.40280789037829534, "learning_rate": 4.5029937578483566e-05, "loss": 0.5545, "num_tokens": 143277757.0, "step": 375 }, { "epoch": 0.2593856655290102, "grad_norm": 0.4139479242813224, "learning_rate": 4.486954145925761e-05, "loss": 0.5327, "num_tokens": 145136704.0, "step": 380 }, { "epoch": 0.2627986348122867, "grad_norm": 0.39471160328650434, "learning_rate": 4.470693036297268e-05, "loss": 0.5229, "num_tokens": 147007396.0, "step": 385 }, { "epoch": 0.26621160409556316, "grad_norm": 0.4071301905270605, "learning_rate": 4.454212502591871e-05, "loss": 0.5396, "num_tokens": 148940122.0, "step": 390 }, { "epoch": 0.2696245733788396, "grad_norm": 0.4385891426510694, "learning_rate": 4.437514646419687e-05, "loss": 0.5473, "num_tokens": 150912454.0, "step": 395 }, { "epoch": 0.27303754266211605, "grad_norm": 0.4067463607989996, "learning_rate": 4.4206015971039625e-05, "loss": 0.5346, "num_tokens": 152809678.0, "step": 400 }, { "epoch": 0.2764505119453925, "grad_norm": 0.4680396603333508, "learning_rate": 4.403475511409531e-05, "loss": 0.5192, "num_tokens": 154713852.0, "step": 405 }, { "epoch": 0.27986348122866894, "grad_norm": 0.39429793945119174, "learning_rate": 4.386138573267788e-05, "loss": 0.5495, "num_tokens": 156683573.0, "step": 410 }, { "epoch": 0.2832764505119454, "grad_norm": 0.4016748463060468, "learning_rate": 4.368592993498194e-05, "loss": 0.525, "num_tokens": 158515202.0, "step": 415 }, { "epoch": 0.28668941979522183, "grad_norm": 0.48270745046669, "learning_rate": 4.350841009526346e-05, "loss": 0.5464, "num_tokens": 160433326.0, "step": 420 }, { "epoch": 0.2901023890784983, "grad_norm": 0.43091858228129815, "learning_rate": 4.33288488509866e-05, "loss": 0.5305, "num_tokens": 162331530.0, "step": 425 }, { "epoch": 0.2935153583617747, "grad_norm": 0.45890742725656564, "learning_rate": 4.3147269099937014e-05, "loss": 0.5472, "num_tokens": 164374316.0, "step": 430 }, { "epoch": 0.29692832764505117, "grad_norm": 0.4169343021344466, "learning_rate": 4.2963693997301814e-05, "loss": 0.5139, "num_tokens": 166249544.0, "step": 435 }, { "epoch": 0.3003412969283277, "grad_norm": 0.38041559154226434, "learning_rate": 4.2778146952716884e-05, "loss": 0.5236, "num_tokens": 168223299.0, "step": 440 }, { "epoch": 0.3037542662116041, "grad_norm": 0.376843828158071, "learning_rate": 4.259065162728163e-05, "loss": 0.5424, "num_tokens": 170189422.0, "step": 445 }, { "epoch": 0.30716723549488056, "grad_norm": 0.39092535624386915, "learning_rate": 4.240123193054172e-05, "loss": 0.5204, "num_tokens": 172132000.0, "step": 450 }, { "epoch": 0.310580204778157, "grad_norm": 0.438997522031723, "learning_rate": 4.2209912017440066e-05, "loss": 0.5322, "num_tokens": 174083393.0, "step": 455 }, { "epoch": 0.31399317406143346, "grad_norm": 0.3770965790416271, "learning_rate": 4.2016716285236655e-05, "loss": 0.5213, "num_tokens": 176086331.0, "step": 460 }, { "epoch": 0.3174061433447099, "grad_norm": 0.44000928731950867, "learning_rate": 4.182166937039731e-05, "loss": 0.5061, "num_tokens": 178019887.0, "step": 465 }, { "epoch": 0.32081911262798635, "grad_norm": 0.4170515747782146, "learning_rate": 4.1624796145452075e-05, "loss": 0.5327, "num_tokens": 179917640.0, "step": 470 }, { "epoch": 0.3242320819112628, "grad_norm": 0.3715898097749977, "learning_rate": 4.1426121715823455e-05, "loss": 0.5366, "num_tokens": 181861425.0, "step": 475 }, { "epoch": 0.32764505119453924, "grad_norm": 0.3953828116703113, "learning_rate": 4.1225671416624886e-05, "loss": 0.5265, "num_tokens": 183746129.0, "step": 480 }, { "epoch": 0.3310580204778157, "grad_norm": 0.43082731831273297, "learning_rate": 4.1023470809430115e-05, "loss": 0.543, "num_tokens": 185690559.0, "step": 485 }, { "epoch": 0.33447098976109213, "grad_norm": 0.3905669096501847, "learning_rate": 4.081954567901341e-05, "loss": 0.5364, "num_tokens": 187699370.0, "step": 490 }, { "epoch": 0.3378839590443686, "grad_norm": 0.38820279646990474, "learning_rate": 4.0613922030061593e-05, "loss": 0.5214, "num_tokens": 189580008.0, "step": 495 }, { "epoch": 0.3412969283276451, "grad_norm": 0.42021271993551723, "learning_rate": 4.040662608385783e-05, "loss": 0.5236, "num_tokens": 191512142.0, "step": 500 }, { "epoch": 0.3447098976109215, "grad_norm": 0.3994553277975334, "learning_rate": 4.019768427493792e-05, "loss": 0.5196, "num_tokens": 193419599.0, "step": 505 }, { "epoch": 0.34812286689419797, "grad_norm": 0.3666548627034848, "learning_rate": 3.99871232477193e-05, "loss": 0.5278, "num_tokens": 195367749.0, "step": 510 }, { "epoch": 0.3515358361774744, "grad_norm": 0.3822859965000655, "learning_rate": 3.9774969853103334e-05, "loss": 0.5079, "num_tokens": 197256984.0, "step": 515 }, { "epoch": 0.35494880546075086, "grad_norm": 0.3731101615447913, "learning_rate": 3.9561251145051325e-05, "loss": 0.5172, "num_tokens": 199215371.0, "step": 520 }, { "epoch": 0.3583617747440273, "grad_norm": 0.4024382355228473, "learning_rate": 3.934599437713449e-05, "loss": 0.5435, "num_tokens": 201102127.0, "step": 525 }, { "epoch": 0.36177474402730375, "grad_norm": 0.42416341608468616, "learning_rate": 3.912922699905861e-05, "loss": 0.5203, "num_tokens": 202903048.0, "step": 530 }, { "epoch": 0.3651877133105802, "grad_norm": 0.4507790300235238, "learning_rate": 3.8910976653163655e-05, "loss": 0.5212, "num_tokens": 204769389.0, "step": 535 }, { "epoch": 0.36860068259385664, "grad_norm": 0.40038853454858814, "learning_rate": 3.869127117089871e-05, "loss": 0.5207, "num_tokens": 206761213.0, "step": 540 }, { "epoch": 0.3720136518771331, "grad_norm": 0.38456031528474893, "learning_rate": 3.847013856927298e-05, "loss": 0.5171, "num_tokens": 208772401.0, "step": 545 }, { "epoch": 0.37542662116040953, "grad_norm": 0.4050638677352149, "learning_rate": 3.824760704728303e-05, "loss": 0.5167, "num_tokens": 210585632.0, "step": 550 }, { "epoch": 0.378839590443686, "grad_norm": 0.41821430327774767, "learning_rate": 3.802370498231677e-05, "loss": 0.5265, "num_tokens": 212446145.0, "step": 555 }, { "epoch": 0.3822525597269625, "grad_norm": 0.3819310209614732, "learning_rate": 3.7798460926534814e-05, "loss": 0.5186, "num_tokens": 214261738.0, "step": 560 }, { "epoch": 0.3856655290102389, "grad_norm": 0.3755106192945339, "learning_rate": 3.757190360322943e-05, "loss": 0.5329, "num_tokens": 216231144.0, "step": 565 }, { "epoch": 0.3890784982935154, "grad_norm": 0.38208455166534944, "learning_rate": 3.7344061903161796e-05, "loss": 0.5223, "num_tokens": 218144024.0, "step": 570 }, { "epoch": 0.3924914675767918, "grad_norm": 0.3539646707114066, "learning_rate": 3.711496488087775e-05, "loss": 0.5212, "num_tokens": 220108684.0, "step": 575 }, { "epoch": 0.39590443686006827, "grad_norm": 0.3589926928321897, "learning_rate": 3.6884641751002784e-05, "loss": 0.5236, "num_tokens": 222057295.0, "step": 580 }, { "epoch": 0.3993174061433447, "grad_norm": 0.3562958553389008, "learning_rate": 3.6653121884516555e-05, "loss": 0.5092, "num_tokens": 223892351.0, "step": 585 }, { "epoch": 0.40273037542662116, "grad_norm": 0.35407235016262395, "learning_rate": 3.6420434805007476e-05, "loss": 0.5179, "num_tokens": 225828573.0, "step": 590 }, { "epoch": 0.4061433447098976, "grad_norm": 0.34046241079182105, "learning_rate": 3.618661018490784e-05, "loss": 0.5134, "num_tokens": 227748575.0, "step": 595 }, { "epoch": 0.40955631399317405, "grad_norm": 0.3329107263813499, "learning_rate": 3.595167784170998e-05, "loss": 0.512, "num_tokens": 229710487.0, "step": 600 }, { "epoch": 0.4129692832764505, "grad_norm": 0.4437872930329917, "learning_rate": 3.571566773416394e-05, "loss": 0.5192, "num_tokens": 231678126.0, "step": 605 }, { "epoch": 0.41638225255972694, "grad_norm": 0.37300167458989936, "learning_rate": 3.5478609958457056e-05, "loss": 0.5368, "num_tokens": 233617525.0, "step": 610 }, { "epoch": 0.4197952218430034, "grad_norm": 0.35893741375216137, "learning_rate": 3.524053474437611e-05, "loss": 0.5231, "num_tokens": 235464027.0, "step": 615 }, { "epoch": 0.4232081911262799, "grad_norm": 0.35587624591010675, "learning_rate": 3.500147245145238e-05, "loss": 0.528, "num_tokens": 237394471.0, "step": 620 }, { "epoch": 0.42662116040955633, "grad_norm": 0.3708746214599896, "learning_rate": 3.476145356509023e-05, "loss": 0.5181, "num_tokens": 239271145.0, "step": 625 }, { "epoch": 0.4300341296928328, "grad_norm": 0.35856346963397057, "learning_rate": 3.452050869267951e-05, "loss": 0.5225, "num_tokens": 241114261.0, "step": 630 }, { "epoch": 0.4334470989761092, "grad_norm": 0.3136928173032959, "learning_rate": 3.427866855969249e-05, "loss": 0.5327, "num_tokens": 242992742.0, "step": 635 }, { "epoch": 0.43686006825938567, "grad_norm": 0.37325632854349405, "learning_rate": 3.4035964005765786e-05, "loss": 0.5081, "num_tokens": 244967987.0, "step": 640 }, { "epoch": 0.4402730375426621, "grad_norm": 0.4618580338006964, "learning_rate": 3.379242598076761e-05, "loss": 0.5427, "num_tokens": 246976387.0, "step": 645 }, { "epoch": 0.44368600682593856, "grad_norm": 0.3586335818960914, "learning_rate": 3.354808554085103e-05, "loss": 0.5107, "num_tokens": 248751095.0, "step": 650 }, { "epoch": 0.447098976109215, "grad_norm": 0.40711927274059334, "learning_rate": 3.330297384449369e-05, "loss": 0.5198, "num_tokens": 250671856.0, "step": 655 }, { "epoch": 0.45051194539249145, "grad_norm": 0.3839556864854868, "learning_rate": 3.3057122148524394e-05, "loss": 0.5082, "num_tokens": 252607265.0, "step": 660 }, { "epoch": 0.4539249146757679, "grad_norm": 0.3965949942432131, "learning_rate": 3.2810561804137316e-05, "loss": 0.5213, "num_tokens": 254497259.0, "step": 665 }, { "epoch": 0.45733788395904434, "grad_norm": 0.3790046349369836, "learning_rate": 3.2563324252893964e-05, "loss": 0.5202, "num_tokens": 256417909.0, "step": 670 }, { "epoch": 0.46075085324232085, "grad_norm": 0.35112804548112325, "learning_rate": 3.2315441022713796e-05, "loss": 0.5024, "num_tokens": 258396985.0, "step": 675 }, { "epoch": 0.4641638225255973, "grad_norm": 0.3498528091945533, "learning_rate": 3.206694372385375e-05, "loss": 0.524, "num_tokens": 260456429.0, "step": 680 }, { "epoch": 0.46757679180887374, "grad_norm": 0.33304357758211456, "learning_rate": 3.1817864044877236e-05, "loss": 0.5041, "num_tokens": 262384995.0, "step": 685 }, { "epoch": 0.4709897610921502, "grad_norm": 0.33178642606056385, "learning_rate": 3.156823374861324e-05, "loss": 0.5078, "num_tokens": 264287905.0, "step": 690 }, { "epoch": 0.47440273037542663, "grad_norm": 0.3667901940014392, "learning_rate": 3.131808466810586e-05, "loss": 0.5146, "num_tokens": 266198448.0, "step": 695 }, { "epoch": 0.4778156996587031, "grad_norm": 0.37162692372701855, "learning_rate": 3.1067448702554976e-05, "loss": 0.5143, "num_tokens": 268098790.0, "step": 700 }, { "epoch": 0.4812286689419795, "grad_norm": 0.3307745019851869, "learning_rate": 3.08163578132484e-05, "loss": 0.5257, "num_tokens": 270033989.0, "step": 705 }, { "epoch": 0.48464163822525597, "grad_norm": 0.34029736690615986, "learning_rate": 3.056484401948618e-05, "loss": 0.522, "num_tokens": 271974065.0, "step": 710 }, { "epoch": 0.4880546075085324, "grad_norm": 0.39322097601658623, "learning_rate": 3.0312939394497442e-05, "loss": 0.5317, "num_tokens": 273982161.0, "step": 715 }, { "epoch": 0.49146757679180886, "grad_norm": 0.41826615673656964, "learning_rate": 3.0060676061350423e-05, "loss": 0.5239, "num_tokens": 275852045.0, "step": 720 }, { "epoch": 0.4948805460750853, "grad_norm": 0.3374695252050787, "learning_rate": 2.9808086188856084e-05, "loss": 0.4941, "num_tokens": 277687831.0, "step": 725 }, { "epoch": 0.49829351535836175, "grad_norm": 0.3304039048198173, "learning_rate": 2.9555201987465932e-05, "loss": 0.5171, "num_tokens": 279504484.0, "step": 730 }, { "epoch": 0.5017064846416383, "grad_norm": 0.34845109052739065, "learning_rate": 2.9302055705164516e-05, "loss": 0.5019, "num_tokens": 281456293.0, "step": 735 }, { "epoch": 0.5051194539249146, "grad_norm": 0.33157077662528966, "learning_rate": 2.9048679623357155e-05, "loss": 0.5132, "num_tokens": 283461546.0, "step": 740 }, { "epoch": 0.5085324232081911, "grad_norm": 0.38033162519773295, "learning_rate": 2.8795106052753356e-05, "loss": 0.5087, "num_tokens": 285433304.0, "step": 745 }, { "epoch": 0.5119453924914675, "grad_norm": 0.34335625928994096, "learning_rate": 2.854136732924656e-05, "loss": 0.5219, "num_tokens": 287371918.0, "step": 750 }, { "epoch": 0.515358361774744, "grad_norm": 0.40565497458200667, "learning_rate": 2.828749580979061e-05, "loss": 0.4941, "num_tokens": 289259239.0, "step": 755 }, { "epoch": 0.5187713310580204, "grad_norm": 0.3720298836144228, "learning_rate": 2.8033523868273625e-05, "loss": 0.5106, "num_tokens": 291057738.0, "step": 760 }, { "epoch": 0.5221843003412969, "grad_norm": 0.31625729147636455, "learning_rate": 2.777948389138959e-05, "loss": 0.5157, "num_tokens": 293034663.0, "step": 765 }, { "epoch": 0.5255972696245734, "grad_norm": 0.3457494085179469, "learning_rate": 2.7525408274508424e-05, "loss": 0.5062, "num_tokens": 294881963.0, "step": 770 }, { "epoch": 0.5290102389078498, "grad_norm": 0.3216688679579312, "learning_rate": 2.727132941754488e-05, "loss": 0.5144, "num_tokens": 296794020.0, "step": 775 }, { "epoch": 0.5324232081911263, "grad_norm": 0.3909025772373753, "learning_rate": 2.70172797208269e-05, "loss": 0.5125, "num_tokens": 298677895.0, "step": 780 }, { "epoch": 0.5358361774744027, "grad_norm": 0.3250766580673669, "learning_rate": 2.676329158096388e-05, "loss": 0.5031, "num_tokens": 300610709.0, "step": 785 }, { "epoch": 0.5392491467576792, "grad_norm": 0.338726222176612, "learning_rate": 2.650939738671543e-05, "loss": 0.4924, "num_tokens": 302387985.0, "step": 790 }, { "epoch": 0.5426621160409556, "grad_norm": 0.3383596322502192, "learning_rate": 2.6255629514861156e-05, "loss": 0.523, "num_tokens": 304293467.0, "step": 795 }, { "epoch": 0.5460750853242321, "grad_norm": 0.39543644326281013, "learning_rate": 2.6002020326071958e-05, "loss": 0.5172, "num_tokens": 306130593.0, "step": 800 }, { "epoch": 0.5494880546075085, "grad_norm": 0.32006279775540386, "learning_rate": 2.5748602160783308e-05, "loss": 0.5222, "num_tokens": 308025370.0, "step": 805 }, { "epoch": 0.552901023890785, "grad_norm": 0.4085796376264559, "learning_rate": 2.5495407335071247e-05, "loss": 0.512, "num_tokens": 309952008.0, "step": 810 }, { "epoch": 0.5563139931740614, "grad_norm": 0.3661536040322409, "learning_rate": 2.5242468136531354e-05, "loss": 0.4993, "num_tokens": 311856787.0, "step": 815 }, { "epoch": 0.5597269624573379, "grad_norm": 0.3752261876585624, "learning_rate": 2.4989816820161497e-05, "loss": 0.5054, "num_tokens": 313725489.0, "step": 820 }, { "epoch": 0.5631399317406144, "grad_norm": 0.36687049267487, "learning_rate": 2.4737485604248534e-05, "loss": 0.5039, "num_tokens": 315690477.0, "step": 825 }, { "epoch": 0.5665529010238908, "grad_norm": 0.3524144778865341, "learning_rate": 2.4485506666259895e-05, "loss": 0.5111, "num_tokens": 317525148.0, "step": 830 }, { "epoch": 0.5699658703071673, "grad_norm": 0.33122901402654253, "learning_rate": 2.4233912138740244e-05, "loss": 0.5053, "num_tokens": 319505322.0, "step": 835 }, { "epoch": 0.5733788395904437, "grad_norm": 0.3303079569668191, "learning_rate": 2.3982734105213967e-05, "loss": 0.5087, "num_tokens": 321410101.0, "step": 840 }, { "epoch": 0.5767918088737202, "grad_norm": 0.33915287853036713, "learning_rate": 2.373200459609379e-05, "loss": 0.5008, "num_tokens": 323302156.0, "step": 845 }, { "epoch": 0.5802047781569966, "grad_norm": 0.3831127722312039, "learning_rate": 2.348175558459628e-05, "loss": 0.5038, "num_tokens": 325102278.0, "step": 850 }, { "epoch": 0.5836177474402731, "grad_norm": 0.37741976583347536, "learning_rate": 2.323201898266458e-05, "loss": 0.5082, "num_tokens": 326903105.0, "step": 855 }, { "epoch": 0.5870307167235495, "grad_norm": 0.4285613277811094, "learning_rate": 2.2982826636898975e-05, "loss": 0.5057, "num_tokens": 328831071.0, "step": 860 }, { "epoch": 0.590443686006826, "grad_norm": 0.36448317241953704, "learning_rate": 2.2734210324495812e-05, "loss": 0.5043, "num_tokens": 330823718.0, "step": 865 }, { "epoch": 0.5938566552901023, "grad_norm": 0.3798280882427834, "learning_rate": 2.24862017491952e-05, "loss": 0.4885, "num_tokens": 332767044.0, "step": 870 }, { "epoch": 0.5972696245733788, "grad_norm": 0.34626645123019356, "learning_rate": 2.2238832537238174e-05, "loss": 0.5074, "num_tokens": 334743310.0, "step": 875 }, { "epoch": 0.6006825938566553, "grad_norm": 0.3163069903882375, "learning_rate": 2.199213423333365e-05, "loss": 0.4798, "num_tokens": 336595638.0, "step": 880 }, { "epoch": 0.6040955631399317, "grad_norm": 0.3377440234044467, "learning_rate": 2.174613829663587e-05, "loss": 0.502, "num_tokens": 338382273.0, "step": 885 }, { "epoch": 0.6075085324232082, "grad_norm": 0.31815725634886977, "learning_rate": 2.1500876096732664e-05, "loss": 0.5112, "num_tokens": 340358925.0, "step": 890 }, { "epoch": 0.6109215017064846, "grad_norm": 0.367940975007064, "learning_rate": 2.1256378909645187e-05, "loss": 0.5098, "num_tokens": 342188299.0, "step": 895 }, { "epoch": 0.6143344709897611, "grad_norm": 0.34518431957216267, "learning_rate": 2.101267791383959e-05, "loss": 0.5011, "num_tokens": 344239058.0, "step": 900 }, { "epoch": 0.6177474402730375, "grad_norm": 0.35244931572919835, "learning_rate": 2.0769804186251096e-05, "loss": 0.5115, "num_tokens": 346206452.0, "step": 905 }, { "epoch": 0.621160409556314, "grad_norm": 0.34420042600455253, "learning_rate": 2.052778869832108e-05, "loss": 0.5066, "num_tokens": 348080585.0, "step": 910 }, { "epoch": 0.6245733788395904, "grad_norm": 0.3764360844364085, "learning_rate": 2.0286662312047537e-05, "loss": 0.5138, "num_tokens": 349930176.0, "step": 915 }, { "epoch": 0.6279863481228669, "grad_norm": 0.30919462082768023, "learning_rate": 2.0046455776049545e-05, "loss": 0.4909, "num_tokens": 351817695.0, "step": 920 }, { "epoch": 0.6313993174061433, "grad_norm": 0.35475689674485694, "learning_rate": 1.980719972164617e-05, "loss": 0.4965, "num_tokens": 353737820.0, "step": 925 }, { "epoch": 0.6348122866894198, "grad_norm": 0.33400288717046317, "learning_rate": 1.9568924658950376e-05, "loss": 0.497, "num_tokens": 355639183.0, "step": 930 }, { "epoch": 0.6382252559726962, "grad_norm": 0.3328103815591411, "learning_rate": 1.9331660972978294e-05, "loss": 0.5171, "num_tokens": 357548627.0, "step": 935 }, { "epoch": 0.6416382252559727, "grad_norm": 0.32058485303964424, "learning_rate": 1.909543891977454e-05, "loss": 0.5134, "num_tokens": 359437581.0, "step": 940 }, { "epoch": 0.6450511945392492, "grad_norm": 0.41282587646488744, "learning_rate": 1.8860288622553922e-05, "loss": 0.4891, "num_tokens": 361329513.0, "step": 945 }, { "epoch": 0.6484641638225256, "grad_norm": 0.3316340082900624, "learning_rate": 1.862624006786014e-05, "loss": 0.4935, "num_tokens": 363189983.0, "step": 950 }, { "epoch": 0.6518771331058021, "grad_norm": 0.36678552775077355, "learning_rate": 1.839332310174184e-05, "loss": 0.4985, "num_tokens": 365024705.0, "step": 955 }, { "epoch": 0.6552901023890785, "grad_norm": 0.3179539334053606, "learning_rate": 1.8161567425946623e-05, "loss": 0.4889, "num_tokens": 366863209.0, "step": 960 }, { "epoch": 0.658703071672355, "grad_norm": 0.33328111865289556, "learning_rate": 1.7931002594133494e-05, "loss": 0.5158, "num_tokens": 368837047.0, "step": 965 }, { "epoch": 0.6621160409556314, "grad_norm": 0.30756457223233913, "learning_rate": 1.7701658008104122e-05, "loss": 0.4994, "num_tokens": 370725860.0, "step": 970 }, { "epoch": 0.6655290102389079, "grad_norm": 0.3031309186317247, "learning_rate": 1.7473562914053505e-05, "loss": 0.4817, "num_tokens": 372569301.0, "step": 975 }, { "epoch": 0.6689419795221843, "grad_norm": 0.32568354226631835, "learning_rate": 1.7246746398840484e-05, "loss": 0.5045, "num_tokens": 374566515.0, "step": 980 }, { "epoch": 0.6723549488054608, "grad_norm": 0.3241787945238758, "learning_rate": 1.702123738627857e-05, "loss": 0.4854, "num_tokens": 376431969.0, "step": 985 }, { "epoch": 0.6757679180887372, "grad_norm": 0.34099547260942087, "learning_rate": 1.679706463344754e-05, "loss": 0.4861, "num_tokens": 378330489.0, "step": 990 }, { "epoch": 0.6791808873720137, "grad_norm": 0.3105997343666536, "learning_rate": 1.657425672702631e-05, "loss": 0.4786, "num_tokens": 380245892.0, "step": 995 }, { "epoch": 0.6825938566552902, "grad_norm": 0.31926125903551006, "learning_rate": 1.6352842079647567e-05, "loss": 0.5024, "num_tokens": 382103468.0, "step": 1000 }, { "epoch": 0.6860068259385665, "grad_norm": 0.35427861114618725, "learning_rate": 1.6132848926274537e-05, "loss": 0.5109, "num_tokens": 384011945.0, "step": 1005 }, { "epoch": 0.689419795221843, "grad_norm": 0.3318228585555292, "learning_rate": 1.5914305320600474e-05, "loss": 0.4874, "num_tokens": 385837297.0, "step": 1010 }, { "epoch": 0.6928327645051194, "grad_norm": 0.4616084679923675, "learning_rate": 1.5697239131471175e-05, "loss": 0.501, "num_tokens": 387718762.0, "step": 1015 }, { "epoch": 0.6962457337883959, "grad_norm": 0.33671130456556214, "learning_rate": 1.5481678039331217e-05, "loss": 0.4907, "num_tokens": 389586680.0, "step": 1020 }, { "epoch": 0.6996587030716723, "grad_norm": 0.3306227060452227, "learning_rate": 1.526764953269403e-05, "loss": 0.508, "num_tokens": 391567792.0, "step": 1025 }, { "epoch": 0.7030716723549488, "grad_norm": 0.31813025756926777, "learning_rate": 1.5055180904636602e-05, "loss": 0.4924, "num_tokens": 393428760.0, "step": 1030 }, { "epoch": 0.7064846416382252, "grad_norm": 0.3302134363815304, "learning_rate": 1.4844299249319e-05, "loss": 0.5088, "num_tokens": 395264025.0, "step": 1035 }, { "epoch": 0.7098976109215017, "grad_norm": 0.32408702255799543, "learning_rate": 1.4635031458529366e-05, "loss": 0.5156, "num_tokens": 397158700.0, "step": 1040 }, { "epoch": 0.7133105802047781, "grad_norm": 0.3146304024133278, "learning_rate": 1.4427404218254598e-05, "loss": 0.4876, "num_tokens": 399092348.0, "step": 1045 }, { "epoch": 0.7167235494880546, "grad_norm": 0.33224101912248266, "learning_rate": 1.422144400527733e-05, "loss": 0.5054, "num_tokens": 400923938.0, "step": 1050 }, { "epoch": 0.7201365187713311, "grad_norm": 0.33570961644767916, "learning_rate": 1.401717708379964e-05, "loss": 0.485, "num_tokens": 402794748.0, "step": 1055 }, { "epoch": 0.7235494880546075, "grad_norm": 0.33759516313845783, "learning_rate": 1.3814629502093803e-05, "loss": 0.4961, "num_tokens": 404685655.0, "step": 1060 }, { "epoch": 0.726962457337884, "grad_norm": 0.2849134557184847, "learning_rate": 1.3613827089180576e-05, "loss": 0.4966, "num_tokens": 406613744.0, "step": 1065 }, { "epoch": 0.7303754266211604, "grad_norm": 0.33308470866419615, "learning_rate": 1.3414795451535448e-05, "loss": 0.4955, "num_tokens": 408582799.0, "step": 1070 }, { "epoch": 0.7337883959044369, "grad_norm": 0.3218010396658918, "learning_rate": 1.3217559969823334e-05, "loss": 0.5041, "num_tokens": 410442926.0, "step": 1075 }, { "epoch": 0.7372013651877133, "grad_norm": 0.2948387124715625, "learning_rate": 1.3022145795661988e-05, "loss": 0.5002, "num_tokens": 412386009.0, "step": 1080 }, { "epoch": 0.7406143344709898, "grad_norm": 0.31655676114802184, "learning_rate": 1.2828577848414636e-05, "loss": 0.5102, "num_tokens": 414357432.0, "step": 1085 }, { "epoch": 0.7440273037542662, "grad_norm": 0.2869995128416729, "learning_rate": 1.2636880812012253e-05, "loss": 0.499, "num_tokens": 416372039.0, "step": 1090 }, { "epoch": 0.7474402730375427, "grad_norm": 0.32259657238383393, "learning_rate": 1.2447079131805857e-05, "loss": 0.4956, "num_tokens": 418265973.0, "step": 1095 }, { "epoch": 0.7508532423208191, "grad_norm": 0.30899676272076165, "learning_rate": 1.2259197011449203e-05, "loss": 0.4829, "num_tokens": 420051975.0, "step": 1100 }, { "epoch": 0.7542662116040956, "grad_norm": 0.29770792665255996, "learning_rate": 1.2073258409812328e-05, "loss": 0.4844, "num_tokens": 421905923.0, "step": 1105 }, { "epoch": 0.757679180887372, "grad_norm": 0.3251270162529146, "learning_rate": 1.1889287037926289e-05, "loss": 0.4948, "num_tokens": 423685709.0, "step": 1110 }, { "epoch": 0.7610921501706485, "grad_norm": 0.33226182684232963, "learning_rate": 1.1707306355959516e-05, "loss": 0.4859, "num_tokens": 425515443.0, "step": 1115 }, { "epoch": 0.764505119453925, "grad_norm": 0.2820363582513593, "learning_rate": 1.152733957022617e-05, "loss": 0.4855, "num_tokens": 427405904.0, "step": 1120 }, { "epoch": 0.7679180887372014, "grad_norm": 0.29505926000276406, "learning_rate": 1.1349409630226804e-05, "loss": 0.4983, "num_tokens": 429329969.0, "step": 1125 }, { "epoch": 0.7713310580204779, "grad_norm": 0.29155650643706393, "learning_rate": 1.1173539225721916e-05, "loss": 0.4842, "num_tokens": 431288378.0, "step": 1130 }, { "epoch": 0.7747440273037542, "grad_norm": 0.30656473872840284, "learning_rate": 1.0999750783838442e-05, "loss": 0.4931, "num_tokens": 433213317.0, "step": 1135 }, { "epoch": 0.7781569965870307, "grad_norm": 0.3103072512057092, "learning_rate": 1.0828066466209891e-05, "loss": 0.4853, "num_tokens": 435077995.0, "step": 1140 }, { "epoch": 0.7815699658703071, "grad_norm": 0.3119583099379152, "learning_rate": 1.0658508166150224e-05, "loss": 0.4902, "num_tokens": 437063353.0, "step": 1145 }, { "epoch": 0.7849829351535836, "grad_norm": 0.2899025649371279, "learning_rate": 1.0491097505862085e-05, "loss": 0.4876, "num_tokens": 439006864.0, "step": 1150 }, { "epoch": 0.78839590443686, "grad_norm": 0.29349884858317443, "learning_rate": 1.0325855833679438e-05, "loss": 0.5039, "num_tokens": 441123044.0, "step": 1155 }, { "epoch": 0.7918088737201365, "grad_norm": 0.31417942708495006, "learning_rate": 1.0162804221345235e-05, "loss": 0.5038, "num_tokens": 443045688.0, "step": 1160 }, { "epoch": 0.7952218430034129, "grad_norm": 0.31804865960369644, "learning_rate": 1.0001963461324357e-05, "loss": 0.4925, "num_tokens": 444874851.0, "step": 1165 }, { "epoch": 0.7986348122866894, "grad_norm": 0.34762843018816847, "learning_rate": 9.843354064152135e-06, "loss": 0.49, "num_tokens": 446852018.0, "step": 1170 }, { "epoch": 0.8020477815699659, "grad_norm": 0.29778977328595035, "learning_rate": 9.686996255818815e-06, "loss": 0.4966, "num_tokens": 448665204.0, "step": 1175 }, { "epoch": 0.8054607508532423, "grad_norm": 0.28927996228687575, "learning_rate": 9.532909975190317e-06, "loss": 0.5137, "num_tokens": 450687287.0, "step": 1180 }, { "epoch": 0.8088737201365188, "grad_norm": 0.3094536714350483, "learning_rate": 9.381114871465644e-06, "loss": 0.4907, "num_tokens": 452550699.0, "step": 1185 }, { "epoch": 0.8122866894197952, "grad_norm": 0.31642622912231355, "learning_rate": 9.231630301671188e-06, "loss": 0.4907, "num_tokens": 454483896.0, "step": 1190 }, { "epoch": 0.8156996587030717, "grad_norm": 0.3081353624354231, "learning_rate": 9.084475328192308e-06, "loss": 0.4759, "num_tokens": 456418788.0, "step": 1195 }, { "epoch": 0.8191126279863481, "grad_norm": 0.365599455782294, "learning_rate": 8.939668716342473e-06, "loss": 0.5015, "num_tokens": 458325861.0, "step": 1200 }, { "epoch": 0.8225255972696246, "grad_norm": 0.3194677820320597, "learning_rate": 8.797228931970305e-06, "loss": 0.5069, "num_tokens": 460227692.0, "step": 1205 }, { "epoch": 0.825938566552901, "grad_norm": 0.2940713904786176, "learning_rate": 8.657174139104807e-06, "loss": 0.4807, "num_tokens": 462055868.0, "step": 1210 }, { "epoch": 0.8293515358361775, "grad_norm": 0.2859069818865439, "learning_rate": 8.519522197639043e-06, "loss": 0.4809, "num_tokens": 463988579.0, "step": 1215 }, { "epoch": 0.8327645051194539, "grad_norm": 0.2817816657796972, "learning_rate": 8.384290661052662e-06, "loss": 0.4955, "num_tokens": 465973876.0, "step": 1220 }, { "epoch": 0.8361774744027304, "grad_norm": 0.2925807997562712, "learning_rate": 8.251496774173451e-06, "loss": 0.4887, "num_tokens": 467962214.0, "step": 1225 }, { "epoch": 0.8395904436860068, "grad_norm": 0.306769088717932, "learning_rate": 8.121157470978268e-06, "loss": 0.4826, "num_tokens": 469791498.0, "step": 1230 }, { "epoch": 0.8430034129692833, "grad_norm": 0.3026860853513822, "learning_rate": 7.993289372433582e-06, "loss": 0.4844, "num_tokens": 471616902.0, "step": 1235 }, { "epoch": 0.8464163822525598, "grad_norm": 0.3002876341550466, "learning_rate": 7.867908784376006e-06, "loss": 0.482, "num_tokens": 473570581.0, "step": 1240 }, { "epoch": 0.8498293515358362, "grad_norm": 0.30847727739844566, "learning_rate": 7.745031695432923e-06, "loss": 0.4959, "num_tokens": 475518680.0, "step": 1245 }, { "epoch": 0.8532423208191127, "grad_norm": 0.3159715882865048, "learning_rate": 7.624673774983614e-06, "loss": 0.5024, "num_tokens": 477401353.0, "step": 1250 }, { "epoch": 0.856655290102389, "grad_norm": 0.28952971357131746, "learning_rate": 7.506850371161078e-06, "loss": 0.4969, "num_tokens": 479371988.0, "step": 1255 }, { "epoch": 0.8600682593856656, "grad_norm": 0.2845959933463602, "learning_rate": 7.391576508894865e-06, "loss": 0.4995, "num_tokens": 481348892.0, "step": 1260 }, { "epoch": 0.863481228668942, "grad_norm": 0.3011207792019779, "learning_rate": 7.27886688799504e-06, "loss": 0.4859, "num_tokens": 483220037.0, "step": 1265 }, { "epoch": 0.8668941979522184, "grad_norm": 0.293919145898071, "learning_rate": 7.1687358812776715e-06, "loss": 0.507, "num_tokens": 485171910.0, "step": 1270 }, { "epoch": 0.8703071672354948, "grad_norm": 0.301426754517889, "learning_rate": 7.061197532731992e-06, "loss": 0.4834, "num_tokens": 487118187.0, "step": 1275 }, { "epoch": 0.8737201365187713, "grad_norm": 0.30097229400923386, "learning_rate": 6.956265555729524e-06, "loss": 0.4801, "num_tokens": 489059548.0, "step": 1280 }, { "epoch": 0.8771331058020477, "grad_norm": 0.2685038385521516, "learning_rate": 6.853953331275304e-06, "loss": 0.4834, "num_tokens": 490973976.0, "step": 1285 }, { "epoch": 0.8805460750853242, "grad_norm": 0.28920434789008614, "learning_rate": 6.754273906301544e-06, "loss": 0.4869, "num_tokens": 492893029.0, "step": 1290 }, { "epoch": 0.8839590443686007, "grad_norm": 0.26489876231378706, "learning_rate": 6.6572399920038775e-06, "loss": 0.506, "num_tokens": 494980675.0, "step": 1295 }, { "epoch": 0.8873720136518771, "grad_norm": 0.2788460104910853, "learning_rate": 6.562863962220414e-06, "loss": 0.4783, "num_tokens": 496905674.0, "step": 1300 }, { "epoch": 0.8907849829351536, "grad_norm": 0.2804986007877692, "learning_rate": 6.47115785185384e-06, "loss": 0.4922, "num_tokens": 498889766.0, "step": 1305 }, { "epoch": 0.89419795221843, "grad_norm": 0.3194121397231278, "learning_rate": 6.382133355336667e-06, "loss": 0.4831, "num_tokens": 500864542.0, "step": 1310 }, { "epoch": 0.8976109215017065, "grad_norm": 0.2826959872970924, "learning_rate": 6.29580182514e-06, "loss": 0.5, "num_tokens": 502876203.0, "step": 1315 }, { "epoch": 0.9010238907849829, "grad_norm": 0.2686904123571036, "learning_rate": 6.212174270325845e-06, "loss": 0.4828, "num_tokens": 504810366.0, "step": 1320 }, { "epoch": 0.9044368600682594, "grad_norm": 0.2663595760500798, "learning_rate": 6.1312613551432135e-06, "loss": 0.4865, "num_tokens": 506777983.0, "step": 1325 }, { "epoch": 0.9078498293515358, "grad_norm": 0.2907833181872873, "learning_rate": 6.053073397668226e-06, "loss": 0.4887, "num_tokens": 508607232.0, "step": 1330 }, { "epoch": 0.9112627986348123, "grad_norm": 0.29397858235192875, "learning_rate": 5.977620368488328e-06, "loss": 0.4882, "num_tokens": 510543906.0, "step": 1335 }, { "epoch": 0.9146757679180887, "grad_norm": 0.31013908532336765, "learning_rate": 5.904911889430853e-06, "loss": 0.501, "num_tokens": 512474084.0, "step": 1340 }, { "epoch": 0.9180887372013652, "grad_norm": 0.283505071193333, "learning_rate": 5.834957232336018e-06, "loss": 0.4763, "num_tokens": 514322576.0, "step": 1345 }, { "epoch": 0.9215017064846417, "grad_norm": 0.28906365848402543, "learning_rate": 5.7677653178745805e-06, "loss": 0.4939, "num_tokens": 516216104.0, "step": 1350 }, { "epoch": 0.9249146757679181, "grad_norm": 0.2727767003049317, "learning_rate": 5.7033447144102785e-06, "loss": 0.4993, "num_tokens": 518137999.0, "step": 1355 }, { "epoch": 0.9283276450511946, "grad_norm": 0.3011258273126128, "learning_rate": 5.641703636907171e-06, "loss": 0.4895, "num_tokens": 519966345.0, "step": 1360 }, { "epoch": 0.931740614334471, "grad_norm": 0.28411785811363793, "learning_rate": 5.582849945882055e-06, "loss": 0.4755, "num_tokens": 521824255.0, "step": 1365 }, { "epoch": 0.9351535836177475, "grad_norm": 0.3050666821162905, "learning_rate": 5.526791146402112e-06, "loss": 0.485, "num_tokens": 523793837.0, "step": 1370 }, { "epoch": 0.9385665529010239, "grad_norm": 0.2677439239480107, "learning_rate": 5.473534387127838e-06, "loss": 0.492, "num_tokens": 525746360.0, "step": 1375 }, { "epoch": 0.9419795221843004, "grad_norm": 0.28833438367340264, "learning_rate": 5.423086459401437e-06, "loss": 0.496, "num_tokens": 527666864.0, "step": 1380 }, { "epoch": 0.9453924914675768, "grad_norm": 0.27631357977330695, "learning_rate": 5.37545379638079e-06, "loss": 0.4946, "num_tokens": 529624010.0, "step": 1385 }, { "epoch": 0.9488054607508533, "grad_norm": 0.29599901993590805, "learning_rate": 5.330642472219102e-06, "loss": 0.4883, "num_tokens": 531466359.0, "step": 1390 }, { "epoch": 0.9522184300341296, "grad_norm": 0.29245998431871206, "learning_rate": 5.28865820129031e-06, "loss": 0.503, "num_tokens": 533360398.0, "step": 1395 }, { "epoch": 0.9556313993174061, "grad_norm": 0.29809224367055454, "learning_rate": 5.24950633746038e-06, "loss": 0.4848, "num_tokens": 535253217.0, "step": 1400 }, { "epoch": 0.9590443686006825, "grad_norm": 0.2681357355557361, "learning_rate": 5.213191873404592e-06, "loss": 0.4753, "num_tokens": 537247787.0, "step": 1405 }, { "epoch": 0.962457337883959, "grad_norm": 0.26573908919210415, "learning_rate": 5.179719439970854e-06, "loss": 0.4811, "num_tokens": 539137436.0, "step": 1410 }, { "epoch": 0.9658703071672355, "grad_norm": 0.3012993878942698, "learning_rate": 5.14909330558919e-06, "loss": 0.4856, "num_tokens": 540954808.0, "step": 1415 }, { "epoch": 0.9692832764505119, "grad_norm": 0.2734104791087793, "learning_rate": 5.121317375727405e-06, "loss": 0.492, "num_tokens": 542824098.0, "step": 1420 }, { "epoch": 0.9726962457337884, "grad_norm": 0.28445668344186853, "learning_rate": 5.0963951923930745e-06, "loss": 0.5046, "num_tokens": 544761712.0, "step": 1425 }, { "epoch": 0.9761092150170648, "grad_norm": 0.3111290218032905, "learning_rate": 5.074329933681866e-06, "loss": 0.4941, "num_tokens": 546578047.0, "step": 1430 }, { "epoch": 0.9795221843003413, "grad_norm": 0.2841539130555513, "learning_rate": 5.055124413372245e-06, "loss": 0.4855, "num_tokens": 548490659.0, "step": 1435 }, { "epoch": 0.9829351535836177, "grad_norm": 0.28667073975075474, "learning_rate": 5.038781080566688e-06, "loss": 0.4854, "num_tokens": 550377811.0, "step": 1440 }, { "epoch": 0.9863481228668942, "grad_norm": 0.30158413834264025, "learning_rate": 5.025302019379348e-06, "loss": 0.4891, "num_tokens": 552306122.0, "step": 1445 }, { "epoch": 0.9897610921501706, "grad_norm": 0.2699230549724662, "learning_rate": 5.014688948670296e-06, "loss": 0.474, "num_tokens": 554224024.0, "step": 1450 }, { "epoch": 0.9931740614334471, "grad_norm": 0.33061635195442485, "learning_rate": 5.006943221826337e-06, "loss": 0.4955, "num_tokens": 556055771.0, "step": 1455 }, { "epoch": 0.9965870307167235, "grad_norm": 0.25966538380747367, "learning_rate": 5.002065826588409e-06, "loss": 0.4832, "num_tokens": 558001366.0, "step": 1460 }, { "epoch": 1.0, "grad_norm": 0.26645961168127524, "learning_rate": 5.000057384925645e-06, "loss": 0.4887, "num_tokens": 559917574.0, "step": 1465 }, { "epoch": 1.0, "step": 1465, "total_flos": 1146656934330368.0, "train_loss": 0.5250163195483106, "train_runtime": 26005.4598, "train_samples_per_second": 3.604, "train_steps_per_second": 0.056 } ], "logging_steps": 5, "max_steps": 1465, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 100, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1146656934330368.0, "train_batch_size": 16, "trial_name": null, "trial_params": null }