{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.18321729571271528, "eval_steps": 500, "global_step": 1000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00018321729571271528, "grad_norm": 7.481241625973833, "learning_rate": 0.0, "loss": 0.764, "step": 1 }, { "epoch": 0.00036643459142543056, "grad_norm": 7.936839352385908, "learning_rate": 1.8315018315018317e-08, "loss": 0.7505, "step": 2 }, { "epoch": 0.0005496518871381459, "grad_norm": 8.381229660678587, "learning_rate": 3.6630036630036635e-08, "loss": 0.7767, "step": 3 }, { "epoch": 0.0007328691828508611, "grad_norm": 7.533434041856819, "learning_rate": 5.494505494505495e-08, "loss": 0.7545, "step": 4 }, { "epoch": 0.0009160864785635764, "grad_norm": 6.4876435204860545, "learning_rate": 7.326007326007327e-08, "loss": 0.7254, "step": 5 }, { "epoch": 0.0010993037742762918, "grad_norm": 6.984439562746247, "learning_rate": 9.157509157509159e-08, "loss": 0.7238, "step": 6 }, { "epoch": 0.001282521069989007, "grad_norm": 8.432795226889532, "learning_rate": 1.098901098901099e-07, "loss": 0.7667, "step": 7 }, { "epoch": 0.0014657383657017222, "grad_norm": 6.615389400475257, "learning_rate": 1.282051282051282e-07, "loss": 0.739, "step": 8 }, { "epoch": 0.0016489556614144375, "grad_norm": 7.043438048491804, "learning_rate": 1.4652014652014654e-07, "loss": 0.7412, "step": 9 }, { "epoch": 0.001832172957127153, "grad_norm": 7.763888342467084, "learning_rate": 1.6483516483516484e-07, "loss": 0.739, "step": 10 }, { "epoch": 0.002015390252839868, "grad_norm": 7.115680896600135, "learning_rate": 1.8315018315018317e-07, "loss": 0.6814, "step": 11 }, { "epoch": 0.0021986075485525836, "grad_norm": 9.472751553447468, "learning_rate": 2.014652014652015e-07, "loss": 0.7694, "step": 12 }, { "epoch": 0.0023818248442652986, "grad_norm": 9.164365397122241, "learning_rate": 2.197802197802198e-07, "loss": 0.8101, "step": 13 }, { "epoch": 0.002565042139978014, "grad_norm": 7.831934204465113, "learning_rate": 2.3809523809523811e-07, "loss": 0.7504, "step": 14 }, { "epoch": 0.002748259435690729, "grad_norm": 6.7423320112288865, "learning_rate": 2.564102564102564e-07, "loss": 0.7442, "step": 15 }, { "epoch": 0.0029314767314034445, "grad_norm": 7.8103949840893145, "learning_rate": 2.7472527472527475e-07, "loss": 0.7782, "step": 16 }, { "epoch": 0.00311469402711616, "grad_norm": 6.1519268993583465, "learning_rate": 2.930402930402931e-07, "loss": 0.663, "step": 17 }, { "epoch": 0.003297911322828875, "grad_norm": 8.049047835270914, "learning_rate": 3.113553113553114e-07, "loss": 0.7934, "step": 18 }, { "epoch": 0.0034811286185415903, "grad_norm": 6.388709624292669, "learning_rate": 3.296703296703297e-07, "loss": 0.7241, "step": 19 }, { "epoch": 0.003664345914254306, "grad_norm": 7.039170133500172, "learning_rate": 3.47985347985348e-07, "loss": 0.7133, "step": 20 }, { "epoch": 0.003847563209967021, "grad_norm": 8.416142106876258, "learning_rate": 3.6630036630036635e-07, "loss": 0.7787, "step": 21 }, { "epoch": 0.004030780505679736, "grad_norm": 5.942542945633484, "learning_rate": 3.846153846153847e-07, "loss": 0.692, "step": 22 }, { "epoch": 0.004213997801392451, "grad_norm": 6.397620608740201, "learning_rate": 4.02930402930403e-07, "loss": 0.6885, "step": 23 }, { "epoch": 0.004397215097105167, "grad_norm": 4.789506080162222, "learning_rate": 4.212454212454213e-07, "loss": 0.6818, "step": 24 }, { "epoch": 0.004580432392817882, "grad_norm": 5.4323837663846275, "learning_rate": 4.395604395604396e-07, "loss": 0.7202, "step": 25 }, { "epoch": 0.004763649688530597, "grad_norm": 6.001418955741802, "learning_rate": 4.578754578754579e-07, "loss": 0.6915, "step": 26 }, { "epoch": 0.004946866984243312, "grad_norm": 6.931290264891086, "learning_rate": 4.7619047619047623e-07, "loss": 0.7214, "step": 27 }, { "epoch": 0.005130084279956028, "grad_norm": 5.423177642103934, "learning_rate": 4.945054945054946e-07, "loss": 0.6627, "step": 28 }, { "epoch": 0.005313301575668743, "grad_norm": 5.824082478920574, "learning_rate": 5.128205128205128e-07, "loss": 0.7128, "step": 29 }, { "epoch": 0.005496518871381458, "grad_norm": 4.837515053542817, "learning_rate": 5.311355311355311e-07, "loss": 0.7063, "step": 30 }, { "epoch": 0.005679736167094174, "grad_norm": 4.2178492878732, "learning_rate": 5.494505494505495e-07, "loss": 0.6667, "step": 31 }, { "epoch": 0.005862953462806889, "grad_norm": 3.7310273500960283, "learning_rate": 5.677655677655678e-07, "loss": 0.6724, "step": 32 }, { "epoch": 0.006046170758519604, "grad_norm": 3.4436166557785013, "learning_rate": 5.860805860805862e-07, "loss": 0.6158, "step": 33 }, { "epoch": 0.00622938805423232, "grad_norm": 2.6914219359650264, "learning_rate": 6.043956043956044e-07, "loss": 0.6384, "step": 34 }, { "epoch": 0.006412605349945035, "grad_norm": 2.8611154351289803, "learning_rate": 6.227106227106228e-07, "loss": 0.6564, "step": 35 }, { "epoch": 0.00659582264565775, "grad_norm": 2.915608387835296, "learning_rate": 6.41025641025641e-07, "loss": 0.6556, "step": 36 }, { "epoch": 0.006779039941370466, "grad_norm": 3.104210670394775, "learning_rate": 6.593406593406594e-07, "loss": 0.6571, "step": 37 }, { "epoch": 0.006962257237083181, "grad_norm": 2.416024185643152, "learning_rate": 6.776556776556777e-07, "loss": 0.6118, "step": 38 }, { "epoch": 0.007145474532795896, "grad_norm": 2.2840327527366187, "learning_rate": 6.95970695970696e-07, "loss": 0.5773, "step": 39 }, { "epoch": 0.007328691828508612, "grad_norm": 2.994955066139323, "learning_rate": 7.142857142857143e-07, "loss": 0.6621, "step": 40 }, { "epoch": 0.007511909124221327, "grad_norm": 2.4977146662969956, "learning_rate": 7.326007326007327e-07, "loss": 0.5878, "step": 41 }, { "epoch": 0.007695126419934042, "grad_norm": 2.2697538551105443, "learning_rate": 7.50915750915751e-07, "loss": 0.6161, "step": 42 }, { "epoch": 0.007878343715646757, "grad_norm": 2.040084870467292, "learning_rate": 7.692307692307694e-07, "loss": 0.625, "step": 43 }, { "epoch": 0.008061561011359472, "grad_norm": 1.4264410515550054, "learning_rate": 7.875457875457876e-07, "loss": 0.5991, "step": 44 }, { "epoch": 0.008244778307072188, "grad_norm": 1.3307329745484617, "learning_rate": 8.05860805860806e-07, "loss": 0.5802, "step": 45 }, { "epoch": 0.008427995602784902, "grad_norm": 1.3186002922814006, "learning_rate": 8.241758241758242e-07, "loss": 0.5971, "step": 46 }, { "epoch": 0.008611212898497618, "grad_norm": 1.4454419288227918, "learning_rate": 8.424908424908426e-07, "loss": 0.6152, "step": 47 }, { "epoch": 0.008794430194210334, "grad_norm": 1.2728865809977838, "learning_rate": 8.608058608058609e-07, "loss": 0.6176, "step": 48 }, { "epoch": 0.008977647489923048, "grad_norm": 1.2387163683388445, "learning_rate": 8.791208791208792e-07, "loss": 0.6233, "step": 49 }, { "epoch": 0.009160864785635764, "grad_norm": 1.1465107255795282, "learning_rate": 8.974358974358975e-07, "loss": 0.5795, "step": 50 }, { "epoch": 0.00934408208134848, "grad_norm": 1.264265250343769, "learning_rate": 9.157509157509158e-07, "loss": 0.6262, "step": 51 }, { "epoch": 0.009527299377061194, "grad_norm": 0.919312835052228, "learning_rate": 9.340659340659342e-07, "loss": 0.5535, "step": 52 }, { "epoch": 0.00971051667277391, "grad_norm": 0.8672735287012353, "learning_rate": 9.523809523809525e-07, "loss": 0.5327, "step": 53 }, { "epoch": 0.009893733968486624, "grad_norm": 0.8288169490471791, "learning_rate": 9.706959706959708e-07, "loss": 0.5673, "step": 54 }, { "epoch": 0.01007695126419934, "grad_norm": 0.8603030677007734, "learning_rate": 9.890109890109891e-07, "loss": 0.5547, "step": 55 }, { "epoch": 0.010260168559912056, "grad_norm": 0.7482006424960487, "learning_rate": 1.0073260073260074e-06, "loss": 0.6012, "step": 56 }, { "epoch": 0.01044338585562477, "grad_norm": 0.6144742944660406, "learning_rate": 1.0256410256410257e-06, "loss": 0.5231, "step": 57 }, { "epoch": 0.010626603151337486, "grad_norm": 0.634897676727318, "learning_rate": 1.0439560439560442e-06, "loss": 0.5726, "step": 58 }, { "epoch": 0.010809820447050202, "grad_norm": 0.6410545094813134, "learning_rate": 1.0622710622710622e-06, "loss": 0.5695, "step": 59 }, { "epoch": 0.010993037742762916, "grad_norm": 0.6609456014192859, "learning_rate": 1.0805860805860807e-06, "loss": 0.5805, "step": 60 }, { "epoch": 0.011176255038475632, "grad_norm": 0.7416419125176478, "learning_rate": 1.098901098901099e-06, "loss": 0.634, "step": 61 }, { "epoch": 0.011359472334188348, "grad_norm": 0.6851293138177442, "learning_rate": 1.1172161172161173e-06, "loss": 0.5544, "step": 62 }, { "epoch": 0.011542689629901062, "grad_norm": 0.6383351643137376, "learning_rate": 1.1355311355311355e-06, "loss": 0.557, "step": 63 }, { "epoch": 0.011725906925613778, "grad_norm": 0.6348085314283121, "learning_rate": 1.153846153846154e-06, "loss": 0.5632, "step": 64 }, { "epoch": 0.011909124221326494, "grad_norm": 0.7024363928163538, "learning_rate": 1.1721611721611723e-06, "loss": 0.5839, "step": 65 }, { "epoch": 0.012092341517039208, "grad_norm": 0.5900911656210056, "learning_rate": 1.1904761904761906e-06, "loss": 0.5344, "step": 66 }, { "epoch": 0.012275558812751924, "grad_norm": 0.6275963867777459, "learning_rate": 1.2087912087912089e-06, "loss": 0.5584, "step": 67 }, { "epoch": 0.01245877610846464, "grad_norm": 0.5865171487667508, "learning_rate": 1.2271062271062271e-06, "loss": 0.575, "step": 68 }, { "epoch": 0.012641993404177354, "grad_norm": 0.674046870317235, "learning_rate": 1.2454212454212456e-06, "loss": 0.5744, "step": 69 }, { "epoch": 0.01282521069989007, "grad_norm": 0.60918911566989, "learning_rate": 1.2637362637362637e-06, "loss": 0.5423, "step": 70 }, { "epoch": 0.013008427995602785, "grad_norm": 0.6221860528344966, "learning_rate": 1.282051282051282e-06, "loss": 0.5652, "step": 71 }, { "epoch": 0.0131916452913155, "grad_norm": 0.606665778835579, "learning_rate": 1.3003663003663005e-06, "loss": 0.6123, "step": 72 }, { "epoch": 0.013374862587028215, "grad_norm": 0.5884723188789052, "learning_rate": 1.3186813186813187e-06, "loss": 0.5863, "step": 73 }, { "epoch": 0.013558079882740931, "grad_norm": 0.5824062487652404, "learning_rate": 1.336996336996337e-06, "loss": 0.5918, "step": 74 }, { "epoch": 0.013741297178453645, "grad_norm": 0.5460196146561194, "learning_rate": 1.3553113553113553e-06, "loss": 0.5912, "step": 75 }, { "epoch": 0.013924514474166361, "grad_norm": 0.49213553293102813, "learning_rate": 1.3736263736263738e-06, "loss": 0.5421, "step": 76 }, { "epoch": 0.014107731769879077, "grad_norm": 0.5594168254817149, "learning_rate": 1.391941391941392e-06, "loss": 0.5122, "step": 77 }, { "epoch": 0.014290949065591791, "grad_norm": 0.5277088034821339, "learning_rate": 1.4102564102564104e-06, "loss": 0.5678, "step": 78 }, { "epoch": 0.014474166361304507, "grad_norm": 0.49574136870511754, "learning_rate": 1.4285714285714286e-06, "loss": 0.5498, "step": 79 }, { "epoch": 0.014657383657017223, "grad_norm": 0.5944104688680958, "learning_rate": 1.4468864468864471e-06, "loss": 0.5469, "step": 80 }, { "epoch": 0.014840600952729937, "grad_norm": 0.4732970613815555, "learning_rate": 1.4652014652014654e-06, "loss": 0.5233, "step": 81 }, { "epoch": 0.015023818248442653, "grad_norm": 0.49591541086638596, "learning_rate": 1.4835164835164837e-06, "loss": 0.5367, "step": 82 }, { "epoch": 0.015207035544155369, "grad_norm": 0.4883648219378977, "learning_rate": 1.501831501831502e-06, "loss": 0.5782, "step": 83 }, { "epoch": 0.015390252839868083, "grad_norm": 0.48559293967760114, "learning_rate": 1.5201465201465202e-06, "loss": 0.5578, "step": 84 }, { "epoch": 0.015573470135580799, "grad_norm": 0.5802435990379926, "learning_rate": 1.5384615384615387e-06, "loss": 0.5562, "step": 85 }, { "epoch": 0.015756687431293513, "grad_norm": 0.6655689083082568, "learning_rate": 1.556776556776557e-06, "loss": 0.5916, "step": 86 }, { "epoch": 0.01593990472700623, "grad_norm": 0.5272919501805459, "learning_rate": 1.5750915750915753e-06, "loss": 0.5462, "step": 87 }, { "epoch": 0.016123122022718945, "grad_norm": 0.5025768009972991, "learning_rate": 1.5934065934065933e-06, "loss": 0.564, "step": 88 }, { "epoch": 0.01630633931843166, "grad_norm": 0.4868371012830415, "learning_rate": 1.611721611721612e-06, "loss": 0.557, "step": 89 }, { "epoch": 0.016489556614144377, "grad_norm": 0.4724914851279723, "learning_rate": 1.6300366300366301e-06, "loss": 0.534, "step": 90 }, { "epoch": 0.01667277390985709, "grad_norm": 0.4933114328584066, "learning_rate": 1.6483516483516484e-06, "loss": 0.5628, "step": 91 }, { "epoch": 0.016855991205569805, "grad_norm": 0.5429724547645147, "learning_rate": 1.6666666666666667e-06, "loss": 0.5693, "step": 92 }, { "epoch": 0.017039208501282523, "grad_norm": 0.46970029586030615, "learning_rate": 1.6849816849816852e-06, "loss": 0.5761, "step": 93 }, { "epoch": 0.017222425796995237, "grad_norm": 0.5466204682716642, "learning_rate": 1.7032967032967034e-06, "loss": 0.5407, "step": 94 }, { "epoch": 0.01740564309270795, "grad_norm": 0.46448768993410167, "learning_rate": 1.7216117216117217e-06, "loss": 0.5784, "step": 95 }, { "epoch": 0.01758886038842067, "grad_norm": 0.46769210989227256, "learning_rate": 1.73992673992674e-06, "loss": 0.5262, "step": 96 }, { "epoch": 0.017772077684133383, "grad_norm": 0.4412027593954725, "learning_rate": 1.7582417582417585e-06, "loss": 0.5608, "step": 97 }, { "epoch": 0.017955294979846097, "grad_norm": 0.5975733130220022, "learning_rate": 1.7765567765567768e-06, "loss": 0.5633, "step": 98 }, { "epoch": 0.018138512275558814, "grad_norm": 0.46156861116006753, "learning_rate": 1.794871794871795e-06, "loss": 0.5576, "step": 99 }, { "epoch": 0.01832172957127153, "grad_norm": 0.5193644526534718, "learning_rate": 1.8131868131868133e-06, "loss": 0.5533, "step": 100 }, { "epoch": 0.018504946866984243, "grad_norm": 0.479596247036775, "learning_rate": 1.8315018315018316e-06, "loss": 0.5337, "step": 101 }, { "epoch": 0.01868816416269696, "grad_norm": 0.4610500867184236, "learning_rate": 1.84981684981685e-06, "loss": 0.5314, "step": 102 }, { "epoch": 0.018871381458409674, "grad_norm": 0.481950984787821, "learning_rate": 1.8681318681318684e-06, "loss": 0.5865, "step": 103 }, { "epoch": 0.01905459875412239, "grad_norm": 0.5225151739123198, "learning_rate": 1.8864468864468866e-06, "loss": 0.5694, "step": 104 }, { "epoch": 0.019237816049835106, "grad_norm": 0.46381468108353424, "learning_rate": 1.904761904761905e-06, "loss": 0.557, "step": 105 }, { "epoch": 0.01942103334554782, "grad_norm": 0.45411743556679485, "learning_rate": 1.9230769230769234e-06, "loss": 0.5648, "step": 106 }, { "epoch": 0.019604250641260534, "grad_norm": 0.4859210831259208, "learning_rate": 1.9413919413919417e-06, "loss": 0.5685, "step": 107 }, { "epoch": 0.01978746793697325, "grad_norm": 0.4205701080898951, "learning_rate": 1.95970695970696e-06, "loss": 0.5426, "step": 108 }, { "epoch": 0.019970685232685966, "grad_norm": 0.5044757429436201, "learning_rate": 1.9780219780219782e-06, "loss": 0.5509, "step": 109 }, { "epoch": 0.02015390252839868, "grad_norm": 0.47571345015861244, "learning_rate": 1.9963369963369965e-06, "loss": 0.5281, "step": 110 }, { "epoch": 0.020337119824111394, "grad_norm": 0.5681283820477757, "learning_rate": 2.0146520146520148e-06, "loss": 0.5398, "step": 111 }, { "epoch": 0.020520337119824112, "grad_norm": 0.48199138528425167, "learning_rate": 2.032967032967033e-06, "loss": 0.56, "step": 112 }, { "epoch": 0.020703554415536826, "grad_norm": 0.4841461198775233, "learning_rate": 2.0512820512820513e-06, "loss": 0.5454, "step": 113 }, { "epoch": 0.02088677171124954, "grad_norm": 0.4680411378560794, "learning_rate": 2.0695970695970696e-06, "loss": 0.516, "step": 114 }, { "epoch": 0.021069989006962258, "grad_norm": 0.4426302566203345, "learning_rate": 2.0879120879120883e-06, "loss": 0.5493, "step": 115 }, { "epoch": 0.021253206302674972, "grad_norm": 0.5379521772056074, "learning_rate": 2.1062271062271066e-06, "loss": 0.5642, "step": 116 }, { "epoch": 0.021436423598387686, "grad_norm": 0.42644965401865687, "learning_rate": 2.1245421245421245e-06, "loss": 0.564, "step": 117 }, { "epoch": 0.021619640894100404, "grad_norm": 1.3341683924694292, "learning_rate": 2.1428571428571427e-06, "loss": 0.5831, "step": 118 }, { "epoch": 0.021802858189813118, "grad_norm": 0.5613781478585862, "learning_rate": 2.1611721611721614e-06, "loss": 0.5598, "step": 119 }, { "epoch": 0.021986075485525832, "grad_norm": 0.4389899980262906, "learning_rate": 2.1794871794871797e-06, "loss": 0.5498, "step": 120 }, { "epoch": 0.02216929278123855, "grad_norm": 0.5433793107490897, "learning_rate": 2.197802197802198e-06, "loss": 0.5838, "step": 121 }, { "epoch": 0.022352510076951264, "grad_norm": 0.49992126724035435, "learning_rate": 2.2161172161172163e-06, "loss": 0.5581, "step": 122 }, { "epoch": 0.022535727372663978, "grad_norm": 0.5192253683114394, "learning_rate": 2.2344322344322345e-06, "loss": 0.5963, "step": 123 }, { "epoch": 0.022718944668376696, "grad_norm": 0.5218720529273816, "learning_rate": 2.252747252747253e-06, "loss": 0.5568, "step": 124 }, { "epoch": 0.02290216196408941, "grad_norm": 0.41135006043138017, "learning_rate": 2.271062271062271e-06, "loss": 0.5256, "step": 125 }, { "epoch": 0.023085379259802124, "grad_norm": 0.4960132808732071, "learning_rate": 2.2893772893772894e-06, "loss": 0.5636, "step": 126 }, { "epoch": 0.02326859655551484, "grad_norm": 0.40294874528313, "learning_rate": 2.307692307692308e-06, "loss": 0.5373, "step": 127 }, { "epoch": 0.023451813851227556, "grad_norm": 0.49129951716432685, "learning_rate": 2.3260073260073264e-06, "loss": 0.5571, "step": 128 }, { "epoch": 0.02363503114694027, "grad_norm": 0.4541626724950761, "learning_rate": 2.3443223443223446e-06, "loss": 0.5407, "step": 129 }, { "epoch": 0.023818248442652987, "grad_norm": 0.44854257290105054, "learning_rate": 2.362637362637363e-06, "loss": 0.5516, "step": 130 }, { "epoch": 0.0240014657383657, "grad_norm": 0.4439039076319426, "learning_rate": 2.380952380952381e-06, "loss": 0.5192, "step": 131 }, { "epoch": 0.024184683034078416, "grad_norm": 0.4410959524536652, "learning_rate": 2.3992673992673995e-06, "loss": 0.5678, "step": 132 }, { "epoch": 0.024367900329791133, "grad_norm": 0.49361868947998544, "learning_rate": 2.4175824175824177e-06, "loss": 0.5591, "step": 133 }, { "epoch": 0.024551117625503847, "grad_norm": 0.5023022664764156, "learning_rate": 2.435897435897436e-06, "loss": 0.5818, "step": 134 }, { "epoch": 0.02473433492121656, "grad_norm": 0.4287249843742935, "learning_rate": 2.4542124542124543e-06, "loss": 0.5532, "step": 135 }, { "epoch": 0.02491755221692928, "grad_norm": 0.43595844702277614, "learning_rate": 2.472527472527473e-06, "loss": 0.534, "step": 136 }, { "epoch": 0.025100769512641993, "grad_norm": 0.5546618594341888, "learning_rate": 2.4908424908424913e-06, "loss": 0.5058, "step": 137 }, { "epoch": 0.025283986808354707, "grad_norm": 0.5017549598118495, "learning_rate": 2.509157509157509e-06, "loss": 0.5416, "step": 138 }, { "epoch": 0.025467204104067425, "grad_norm": 0.47047331776441903, "learning_rate": 2.5274725274725274e-06, "loss": 0.5557, "step": 139 }, { "epoch": 0.02565042139978014, "grad_norm": 0.4881037742451912, "learning_rate": 2.5457875457875457e-06, "loss": 0.5494, "step": 140 }, { "epoch": 0.025833638695492853, "grad_norm": 0.41933017046045795, "learning_rate": 2.564102564102564e-06, "loss": 0.5729, "step": 141 }, { "epoch": 0.02601685599120557, "grad_norm": 0.39974209370594543, "learning_rate": 2.582417582417583e-06, "loss": 0.5381, "step": 142 }, { "epoch": 0.026200073286918285, "grad_norm": 0.4473481461077892, "learning_rate": 2.600732600732601e-06, "loss": 0.5336, "step": 143 }, { "epoch": 0.026383290582631, "grad_norm": 0.45197895778035446, "learning_rate": 2.6190476190476192e-06, "loss": 0.5271, "step": 144 }, { "epoch": 0.026566507878343717, "grad_norm": 0.49529182062519755, "learning_rate": 2.6373626373626375e-06, "loss": 0.5468, "step": 145 }, { "epoch": 0.02674972517405643, "grad_norm": 0.5413416851523152, "learning_rate": 2.6556776556776558e-06, "loss": 0.5807, "step": 146 }, { "epoch": 0.026932942469769145, "grad_norm": 0.4703582853460863, "learning_rate": 2.673992673992674e-06, "loss": 0.538, "step": 147 }, { "epoch": 0.027116159765481863, "grad_norm": 0.4332295907365602, "learning_rate": 2.6923076923076923e-06, "loss": 0.5167, "step": 148 }, { "epoch": 0.027299377061194577, "grad_norm": 0.458616583095158, "learning_rate": 2.7106227106227106e-06, "loss": 0.5226, "step": 149 }, { "epoch": 0.02748259435690729, "grad_norm": 0.43740744109233864, "learning_rate": 2.728937728937729e-06, "loss": 0.541, "step": 150 }, { "epoch": 0.02766581165262001, "grad_norm": 0.5735574364330706, "learning_rate": 2.7472527472527476e-06, "loss": 0.5835, "step": 151 }, { "epoch": 0.027849028948332723, "grad_norm": 0.5055518766772779, "learning_rate": 2.765567765567766e-06, "loss": 0.5631, "step": 152 }, { "epoch": 0.028032246244045437, "grad_norm": 0.43271616931082146, "learning_rate": 2.783882783882784e-06, "loss": 0.5426, "step": 153 }, { "epoch": 0.028215463539758154, "grad_norm": 0.4971463373765023, "learning_rate": 2.8021978021978024e-06, "loss": 0.5389, "step": 154 }, { "epoch": 0.02839868083547087, "grad_norm": 0.4591734918280654, "learning_rate": 2.8205128205128207e-06, "loss": 0.5309, "step": 155 }, { "epoch": 0.028581898131183583, "grad_norm": 0.45247087926212365, "learning_rate": 2.838827838827839e-06, "loss": 0.5536, "step": 156 }, { "epoch": 0.0287651154268963, "grad_norm": 0.46626015341742694, "learning_rate": 2.8571428571428573e-06, "loss": 0.559, "step": 157 }, { "epoch": 0.028948332722609015, "grad_norm": 0.4834854871636071, "learning_rate": 2.8754578754578755e-06, "loss": 0.5476, "step": 158 }, { "epoch": 0.02913155001832173, "grad_norm": 0.570546858013933, "learning_rate": 2.8937728937728942e-06, "loss": 0.5394, "step": 159 }, { "epoch": 0.029314767314034446, "grad_norm": 0.45441480104130544, "learning_rate": 2.9120879120879125e-06, "loss": 0.5029, "step": 160 }, { "epoch": 0.02949798460974716, "grad_norm": 0.5206871098196969, "learning_rate": 2.930402930402931e-06, "loss": 0.528, "step": 161 }, { "epoch": 0.029681201905459875, "grad_norm": 0.43359542229542136, "learning_rate": 2.948717948717949e-06, "loss": 0.5502, "step": 162 }, { "epoch": 0.029864419201172592, "grad_norm": 0.46302617065984364, "learning_rate": 2.9670329670329673e-06, "loss": 0.5447, "step": 163 }, { "epoch": 0.030047636496885306, "grad_norm": 0.3915642271916536, "learning_rate": 2.9853479853479856e-06, "loss": 0.503, "step": 164 }, { "epoch": 0.03023085379259802, "grad_norm": 0.42991246482419715, "learning_rate": 3.003663003663004e-06, "loss": 0.521, "step": 165 }, { "epoch": 0.030414071088310738, "grad_norm": 0.468175424095518, "learning_rate": 3.021978021978022e-06, "loss": 0.5101, "step": 166 }, { "epoch": 0.030597288384023452, "grad_norm": 0.6468735604295471, "learning_rate": 3.0402930402930405e-06, "loss": 0.5617, "step": 167 }, { "epoch": 0.030780505679736166, "grad_norm": 0.5058923699848836, "learning_rate": 3.058608058608059e-06, "loss": 0.5154, "step": 168 }, { "epoch": 0.030963722975448884, "grad_norm": 0.45437537978064513, "learning_rate": 3.0769230769230774e-06, "loss": 0.4993, "step": 169 }, { "epoch": 0.031146940271161598, "grad_norm": 0.467183819204265, "learning_rate": 3.0952380952380957e-06, "loss": 0.5701, "step": 170 }, { "epoch": 0.031330157566874316, "grad_norm": 0.3835674459614267, "learning_rate": 3.113553113553114e-06, "loss": 0.4902, "step": 171 }, { "epoch": 0.031513374862587026, "grad_norm": 0.4292795112150875, "learning_rate": 3.1318681318681323e-06, "loss": 0.5237, "step": 172 }, { "epoch": 0.031696592158299744, "grad_norm": 0.6593051731116806, "learning_rate": 3.1501831501831505e-06, "loss": 0.5361, "step": 173 }, { "epoch": 0.03187980945401246, "grad_norm": 0.4718436483558741, "learning_rate": 3.1684981684981684e-06, "loss": 0.5324, "step": 174 }, { "epoch": 0.03206302674972517, "grad_norm": 0.48018416222395494, "learning_rate": 3.1868131868131867e-06, "loss": 0.5273, "step": 175 }, { "epoch": 0.03224624404543789, "grad_norm": 0.6316028239985647, "learning_rate": 3.205128205128206e-06, "loss": 0.5346, "step": 176 }, { "epoch": 0.03242946134115061, "grad_norm": 0.4596646339699305, "learning_rate": 3.223443223443224e-06, "loss": 0.5263, "step": 177 }, { "epoch": 0.03261267863686332, "grad_norm": 0.6529252419894329, "learning_rate": 3.2417582417582424e-06, "loss": 0.5442, "step": 178 }, { "epoch": 0.032795895932576036, "grad_norm": 0.5538108341969676, "learning_rate": 3.2600732600732602e-06, "loss": 0.543, "step": 179 }, { "epoch": 0.03297911322828875, "grad_norm": 0.46058359751530825, "learning_rate": 3.2783882783882785e-06, "loss": 0.5101, "step": 180 }, { "epoch": 0.033162330524001464, "grad_norm": 0.42238032667898895, "learning_rate": 3.2967032967032968e-06, "loss": 0.5487, "step": 181 }, { "epoch": 0.03334554781971418, "grad_norm": 0.49973876221605035, "learning_rate": 3.315018315018315e-06, "loss": 0.5333, "step": 182 }, { "epoch": 0.0335287651154269, "grad_norm": 0.49185079032879564, "learning_rate": 3.3333333333333333e-06, "loss": 0.5733, "step": 183 }, { "epoch": 0.03371198241113961, "grad_norm": 0.41357500203470565, "learning_rate": 3.3516483516483516e-06, "loss": 0.5108, "step": 184 }, { "epoch": 0.03389519970685233, "grad_norm": 0.45597659979596383, "learning_rate": 3.3699633699633703e-06, "loss": 0.5316, "step": 185 }, { "epoch": 0.034078417002565045, "grad_norm": 0.523873134306111, "learning_rate": 3.3882783882783886e-06, "loss": 0.5516, "step": 186 }, { "epoch": 0.034261634298277756, "grad_norm": 0.4604705346503309, "learning_rate": 3.406593406593407e-06, "loss": 0.5463, "step": 187 }, { "epoch": 0.03444485159399047, "grad_norm": 0.4494660179181277, "learning_rate": 3.424908424908425e-06, "loss": 0.5195, "step": 188 }, { "epoch": 0.03462806888970319, "grad_norm": 0.47161535336220833, "learning_rate": 3.4432234432234434e-06, "loss": 0.5495, "step": 189 }, { "epoch": 0.0348112861854159, "grad_norm": 0.41422931205807795, "learning_rate": 3.4615384615384617e-06, "loss": 0.5165, "step": 190 }, { "epoch": 0.03499450348112862, "grad_norm": 0.48552414748511474, "learning_rate": 3.47985347985348e-06, "loss": 0.5207, "step": 191 }, { "epoch": 0.03517772077684134, "grad_norm": 0.49918231460984896, "learning_rate": 3.4981684981684982e-06, "loss": 0.5055, "step": 192 }, { "epoch": 0.03536093807255405, "grad_norm": 0.5047784031000427, "learning_rate": 3.516483516483517e-06, "loss": 0.5541, "step": 193 }, { "epoch": 0.035544155368266765, "grad_norm": 0.45515393167769386, "learning_rate": 3.5347985347985352e-06, "loss": 0.5324, "step": 194 }, { "epoch": 0.03572737266397948, "grad_norm": 0.4697755766255993, "learning_rate": 3.5531135531135535e-06, "loss": 0.5317, "step": 195 }, { "epoch": 0.035910589959692193, "grad_norm": 0.47670842684148323, "learning_rate": 3.5714285714285718e-06, "loss": 0.5331, "step": 196 }, { "epoch": 0.03609380725540491, "grad_norm": 0.46160363428859325, "learning_rate": 3.58974358974359e-06, "loss": 0.5645, "step": 197 }, { "epoch": 0.03627702455111763, "grad_norm": 0.5456279444043086, "learning_rate": 3.6080586080586083e-06, "loss": 0.4923, "step": 198 }, { "epoch": 0.03646024184683034, "grad_norm": 0.45679324633890633, "learning_rate": 3.6263736263736266e-06, "loss": 0.5408, "step": 199 }, { "epoch": 0.03664345914254306, "grad_norm": 0.4038620329916062, "learning_rate": 3.644688644688645e-06, "loss": 0.5542, "step": 200 }, { "epoch": 0.036826676438255775, "grad_norm": 0.42338470275172285, "learning_rate": 3.663003663003663e-06, "loss": 0.5509, "step": 201 }, { "epoch": 0.037009893733968485, "grad_norm": 0.48862631901451187, "learning_rate": 3.681318681318682e-06, "loss": 0.533, "step": 202 }, { "epoch": 0.0371931110296812, "grad_norm": 0.4339739485667441, "learning_rate": 3.6996336996337e-06, "loss": 0.5331, "step": 203 }, { "epoch": 0.03737632832539392, "grad_norm": 0.4040593227125272, "learning_rate": 3.7179487179487184e-06, "loss": 0.568, "step": 204 }, { "epoch": 0.03755954562110663, "grad_norm": 0.5025361443024049, "learning_rate": 3.7362637362637367e-06, "loss": 0.5616, "step": 205 }, { "epoch": 0.03774276291681935, "grad_norm": 0.43671528407867277, "learning_rate": 3.754578754578755e-06, "loss": 0.5468, "step": 206 }, { "epoch": 0.037925980212532066, "grad_norm": 0.5426409729355149, "learning_rate": 3.7728937728937733e-06, "loss": 0.5274, "step": 207 }, { "epoch": 0.03810919750824478, "grad_norm": 0.4383066270524436, "learning_rate": 3.7912087912087915e-06, "loss": 0.5491, "step": 208 }, { "epoch": 0.038292414803957495, "grad_norm": 0.4541571165503901, "learning_rate": 3.80952380952381e-06, "loss": 0.511, "step": 209 }, { "epoch": 0.03847563209967021, "grad_norm": 0.42914694472434756, "learning_rate": 3.827838827838828e-06, "loss": 0.4891, "step": 210 }, { "epoch": 0.03865884939538292, "grad_norm": 0.4583586099579229, "learning_rate": 3.846153846153847e-06, "loss": 0.5359, "step": 211 }, { "epoch": 0.03884206669109564, "grad_norm": 0.46332307938408596, "learning_rate": 3.864468864468865e-06, "loss": 0.506, "step": 212 }, { "epoch": 0.03902528398680836, "grad_norm": 0.4979093495563886, "learning_rate": 3.882783882783883e-06, "loss": 0.5523, "step": 213 }, { "epoch": 0.03920850128252107, "grad_norm": 0.524225251704003, "learning_rate": 3.901098901098901e-06, "loss": 0.543, "step": 214 }, { "epoch": 0.039391718578233786, "grad_norm": 0.5147235908704881, "learning_rate": 3.91941391941392e-06, "loss": 0.5465, "step": 215 }, { "epoch": 0.0395749358739465, "grad_norm": 0.4170039710750516, "learning_rate": 3.937728937728938e-06, "loss": 0.557, "step": 216 }, { "epoch": 0.039758153169659215, "grad_norm": 0.4459151957597281, "learning_rate": 3.9560439560439565e-06, "loss": 0.5176, "step": 217 }, { "epoch": 0.03994137046537193, "grad_norm": 0.4387028825678643, "learning_rate": 3.974358974358974e-06, "loss": 0.5433, "step": 218 }, { "epoch": 0.04012458776108464, "grad_norm": 0.4606220492260897, "learning_rate": 3.992673992673993e-06, "loss": 0.5558, "step": 219 }, { "epoch": 0.04030780505679736, "grad_norm": 0.43876788524484817, "learning_rate": 4.010989010989012e-06, "loss": 0.496, "step": 220 }, { "epoch": 0.04049102235251008, "grad_norm": 0.5354442317270937, "learning_rate": 4.0293040293040296e-06, "loss": 0.4883, "step": 221 }, { "epoch": 0.04067423964822279, "grad_norm": 0.4471338769246311, "learning_rate": 4.047619047619048e-06, "loss": 0.5248, "step": 222 }, { "epoch": 0.040857456943935506, "grad_norm": 0.4728450428012797, "learning_rate": 4.065934065934066e-06, "loss": 0.5412, "step": 223 }, { "epoch": 0.041040674239648224, "grad_norm": 0.5021628515290991, "learning_rate": 4.084249084249085e-06, "loss": 0.5398, "step": 224 }, { "epoch": 0.041223891535360935, "grad_norm": 0.379469323319607, "learning_rate": 4.102564102564103e-06, "loss": 0.5484, "step": 225 }, { "epoch": 0.04140710883107365, "grad_norm": 0.4682923741442823, "learning_rate": 4.120879120879121e-06, "loss": 0.5221, "step": 226 }, { "epoch": 0.04159032612678637, "grad_norm": 0.5298230539031403, "learning_rate": 4.139194139194139e-06, "loss": 0.5248, "step": 227 }, { "epoch": 0.04177354342249908, "grad_norm": 0.448919135267925, "learning_rate": 4.157509157509158e-06, "loss": 0.5492, "step": 228 }, { "epoch": 0.0419567607182118, "grad_norm": 0.4651550006247672, "learning_rate": 4.175824175824177e-06, "loss": 0.5244, "step": 229 }, { "epoch": 0.042139978013924516, "grad_norm": 0.46885786974024124, "learning_rate": 4.1941391941391945e-06, "loss": 0.5145, "step": 230 }, { "epoch": 0.04232319530963723, "grad_norm": 0.9939588576814024, "learning_rate": 4.212454212454213e-06, "loss": 0.507, "step": 231 }, { "epoch": 0.042506412605349944, "grad_norm": 0.42445387506164906, "learning_rate": 4.230769230769231e-06, "loss": 0.528, "step": 232 }, { "epoch": 0.04268962990106266, "grad_norm": 0.4386077325175301, "learning_rate": 4.249084249084249e-06, "loss": 0.5235, "step": 233 }, { "epoch": 0.04287284719677537, "grad_norm": 0.4275787850644743, "learning_rate": 4.267399267399268e-06, "loss": 0.5268, "step": 234 }, { "epoch": 0.04305606449248809, "grad_norm": 0.3819534900324145, "learning_rate": 4.2857142857142855e-06, "loss": 0.5108, "step": 235 }, { "epoch": 0.04323928178820081, "grad_norm": 0.5069196499403993, "learning_rate": 4.304029304029305e-06, "loss": 0.4914, "step": 236 }, { "epoch": 0.04342249908391352, "grad_norm": 0.4224671659155117, "learning_rate": 4.322344322344323e-06, "loss": 0.531, "step": 237 }, { "epoch": 0.043605716379626236, "grad_norm": 0.49892435842001814, "learning_rate": 4.340659340659341e-06, "loss": 0.5309, "step": 238 }, { "epoch": 0.043788933675338954, "grad_norm": 0.4435906661481072, "learning_rate": 4.358974358974359e-06, "loss": 0.5386, "step": 239 }, { "epoch": 0.043972150971051664, "grad_norm": 0.4646701721197805, "learning_rate": 4.377289377289377e-06, "loss": 0.5445, "step": 240 }, { "epoch": 0.04415536826676438, "grad_norm": 0.42345412346700445, "learning_rate": 4.395604395604396e-06, "loss": 0.5171, "step": 241 }, { "epoch": 0.0443385855624771, "grad_norm": 0.4664513677280782, "learning_rate": 4.413919413919414e-06, "loss": 0.5471, "step": 242 }, { "epoch": 0.04452180285818981, "grad_norm": 0.5007403600133091, "learning_rate": 4.4322344322344325e-06, "loss": 0.5224, "step": 243 }, { "epoch": 0.04470502015390253, "grad_norm": 0.4202775330369337, "learning_rate": 4.45054945054945e-06, "loss": 0.5283, "step": 244 }, { "epoch": 0.044888237449615245, "grad_norm": 0.485312343325994, "learning_rate": 4.468864468864469e-06, "loss": 0.5365, "step": 245 }, { "epoch": 0.045071454745327956, "grad_norm": 0.4796744192562041, "learning_rate": 4.487179487179488e-06, "loss": 0.5481, "step": 246 }, { "epoch": 0.045254672041040674, "grad_norm": 0.4699525411145978, "learning_rate": 4.505494505494506e-06, "loss": 0.526, "step": 247 }, { "epoch": 0.04543788933675339, "grad_norm": 0.37235134296143163, "learning_rate": 4.523809523809524e-06, "loss": 0.4929, "step": 248 }, { "epoch": 0.0456211066324661, "grad_norm": 0.4594065464998237, "learning_rate": 4.542124542124542e-06, "loss": 0.5435, "step": 249 }, { "epoch": 0.04580432392817882, "grad_norm": 0.4812140084395847, "learning_rate": 4.560439560439561e-06, "loss": 0.4714, "step": 250 }, { "epoch": 0.04598754122389154, "grad_norm": 0.44275937617791644, "learning_rate": 4.578754578754579e-06, "loss": 0.5289, "step": 251 }, { "epoch": 0.04617075851960425, "grad_norm": 0.543477861772032, "learning_rate": 4.5970695970695975e-06, "loss": 0.5176, "step": 252 }, { "epoch": 0.046353975815316965, "grad_norm": 0.9195336625704912, "learning_rate": 4.615384615384616e-06, "loss": 0.5315, "step": 253 }, { "epoch": 0.04653719311102968, "grad_norm": 0.4922315670719196, "learning_rate": 4.633699633699634e-06, "loss": 0.5364, "step": 254 }, { "epoch": 0.046720410406742394, "grad_norm": 0.47361690353516367, "learning_rate": 4.652014652014653e-06, "loss": 0.5591, "step": 255 }, { "epoch": 0.04690362770245511, "grad_norm": 0.5164779414047217, "learning_rate": 4.6703296703296706e-06, "loss": 0.5538, "step": 256 }, { "epoch": 0.04708684499816783, "grad_norm": 0.421397036777767, "learning_rate": 4.688644688644689e-06, "loss": 0.5338, "step": 257 }, { "epoch": 0.04727006229388054, "grad_norm": 0.5053871259325204, "learning_rate": 4.706959706959707e-06, "loss": 0.544, "step": 258 }, { "epoch": 0.04745327958959326, "grad_norm": 0.4607447877406368, "learning_rate": 4.725274725274726e-06, "loss": 0.4812, "step": 259 }, { "epoch": 0.047636496885305975, "grad_norm": 0.4875422302168998, "learning_rate": 4.743589743589744e-06, "loss": 0.5614, "step": 260 }, { "epoch": 0.047819714181018685, "grad_norm": 0.41579068200919733, "learning_rate": 4.761904761904762e-06, "loss": 0.4729, "step": 261 }, { "epoch": 0.0480029314767314, "grad_norm": 0.46282242693556186, "learning_rate": 4.780219780219781e-06, "loss": 0.5224, "step": 262 }, { "epoch": 0.04818614877244412, "grad_norm": 0.4482830307148575, "learning_rate": 4.798534798534799e-06, "loss": 0.5326, "step": 263 }, { "epoch": 0.04836936606815683, "grad_norm": 0.4342637761169385, "learning_rate": 4.816849816849818e-06, "loss": 0.5056, "step": 264 }, { "epoch": 0.04855258336386955, "grad_norm": 0.42342307710917526, "learning_rate": 4.8351648351648355e-06, "loss": 0.5049, "step": 265 }, { "epoch": 0.04873580065958227, "grad_norm": 0.41567729548709964, "learning_rate": 4.853479853479854e-06, "loss": 0.5313, "step": 266 }, { "epoch": 0.04891901795529498, "grad_norm": 0.44877730848158315, "learning_rate": 4.871794871794872e-06, "loss": 0.5027, "step": 267 }, { "epoch": 0.049102235251007695, "grad_norm": 0.46298601715996757, "learning_rate": 4.890109890109891e-06, "loss": 0.5418, "step": 268 }, { "epoch": 0.04928545254672041, "grad_norm": 0.4233152082129357, "learning_rate": 4.908424908424909e-06, "loss": 0.534, "step": 269 }, { "epoch": 0.04946866984243312, "grad_norm": 0.42327618076780654, "learning_rate": 4.926739926739927e-06, "loss": 0.5073, "step": 270 }, { "epoch": 0.04965188713814584, "grad_norm": 0.47132160003804374, "learning_rate": 4.945054945054946e-06, "loss": 0.564, "step": 271 }, { "epoch": 0.04983510443385856, "grad_norm": 0.467395758671848, "learning_rate": 4.963369963369964e-06, "loss": 0.5194, "step": 272 }, { "epoch": 0.05001832172957127, "grad_norm": 0.4377241688268797, "learning_rate": 4.9816849816849826e-06, "loss": 0.5473, "step": 273 }, { "epoch": 0.05020153902528399, "grad_norm": 0.4344310773487788, "learning_rate": 5e-06, "loss": 0.5198, "step": 274 }, { "epoch": 0.050384756320996704, "grad_norm": 0.505691771505538, "learning_rate": 5.018315018315018e-06, "loss": 0.5001, "step": 275 }, { "epoch": 0.050567973616709415, "grad_norm": 0.4741702696479342, "learning_rate": 5.036630036630037e-06, "loss": 0.5269, "step": 276 }, { "epoch": 0.05075119091242213, "grad_norm": 0.4123760853657366, "learning_rate": 5.054945054945055e-06, "loss": 0.5366, "step": 277 }, { "epoch": 0.05093440820813485, "grad_norm": 0.5165952852785715, "learning_rate": 5.0732600732600735e-06, "loss": 0.5629, "step": 278 }, { "epoch": 0.05111762550384756, "grad_norm": 0.6018743338441076, "learning_rate": 5.091575091575091e-06, "loss": 0.5268, "step": 279 }, { "epoch": 0.05130084279956028, "grad_norm": 0.4647905088113548, "learning_rate": 5.10989010989011e-06, "loss": 0.5353, "step": 280 }, { "epoch": 0.051484060095272996, "grad_norm": 0.4666862939383661, "learning_rate": 5.128205128205128e-06, "loss": 0.5582, "step": 281 }, { "epoch": 0.05166727739098571, "grad_norm": 0.42295571485014016, "learning_rate": 5.146520146520147e-06, "loss": 0.5389, "step": 282 }, { "epoch": 0.051850494686698424, "grad_norm": 0.5088982589681916, "learning_rate": 5.164835164835166e-06, "loss": 0.5114, "step": 283 }, { "epoch": 0.05203371198241114, "grad_norm": 0.39252264391052066, "learning_rate": 5.183150183150184e-06, "loss": 0.5078, "step": 284 }, { "epoch": 0.05221692927812385, "grad_norm": 0.47464881831711925, "learning_rate": 5.201465201465202e-06, "loss": 0.5719, "step": 285 }, { "epoch": 0.05240014657383657, "grad_norm": 0.42811181510690394, "learning_rate": 5.219780219780221e-06, "loss": 0.536, "step": 286 }, { "epoch": 0.05258336386954929, "grad_norm": 0.4983761768332983, "learning_rate": 5.2380952380952384e-06, "loss": 0.5454, "step": 287 }, { "epoch": 0.052766581165262, "grad_norm": 0.3871752405078846, "learning_rate": 5.256410256410257e-06, "loss": 0.5172, "step": 288 }, { "epoch": 0.052949798460974716, "grad_norm": 0.44878294094458826, "learning_rate": 5.274725274725275e-06, "loss": 0.4965, "step": 289 }, { "epoch": 0.053133015756687434, "grad_norm": 0.4193937449864018, "learning_rate": 5.293040293040294e-06, "loss": 0.5418, "step": 290 }, { "epoch": 0.053316233052400144, "grad_norm": 0.47326184468203625, "learning_rate": 5.3113553113553116e-06, "loss": 0.5465, "step": 291 }, { "epoch": 0.05349945034811286, "grad_norm": 0.46890596965933473, "learning_rate": 5.32967032967033e-06, "loss": 0.5427, "step": 292 }, { "epoch": 0.05368266764382558, "grad_norm": 0.4718404244115825, "learning_rate": 5.347985347985348e-06, "loss": 0.5315, "step": 293 }, { "epoch": 0.05386588493953829, "grad_norm": 0.5313479853203268, "learning_rate": 5.366300366300367e-06, "loss": 0.5261, "step": 294 }, { "epoch": 0.05404910223525101, "grad_norm": 0.4619862699623299, "learning_rate": 5.384615384615385e-06, "loss": 0.5224, "step": 295 }, { "epoch": 0.054232319530963725, "grad_norm": 0.4235637361399875, "learning_rate": 5.402930402930403e-06, "loss": 0.5484, "step": 296 }, { "epoch": 0.054415536826676436, "grad_norm": 0.4826898937600368, "learning_rate": 5.421245421245421e-06, "loss": 0.524, "step": 297 }, { "epoch": 0.054598754122389154, "grad_norm": 0.43904512964940123, "learning_rate": 5.43956043956044e-06, "loss": 0.5486, "step": 298 }, { "epoch": 0.05478197141810187, "grad_norm": 0.5045757484045217, "learning_rate": 5.457875457875458e-06, "loss": 0.5407, "step": 299 }, { "epoch": 0.05496518871381458, "grad_norm": 0.47829971819207484, "learning_rate": 5.476190476190477e-06, "loss": 0.5344, "step": 300 }, { "epoch": 0.0551484060095273, "grad_norm": 0.416644246441645, "learning_rate": 5.494505494505495e-06, "loss": 0.5111, "step": 301 }, { "epoch": 0.05533162330524002, "grad_norm": 0.488275746902462, "learning_rate": 5.512820512820514e-06, "loss": 0.54, "step": 302 }, { "epoch": 0.05551484060095273, "grad_norm": 0.43082352297647686, "learning_rate": 5.531135531135532e-06, "loss": 0.5219, "step": 303 }, { "epoch": 0.055698057896665445, "grad_norm": 0.41708996725660685, "learning_rate": 5.5494505494505504e-06, "loss": 0.5272, "step": 304 }, { "epoch": 0.05588127519237816, "grad_norm": 0.4748217492221608, "learning_rate": 5.567765567765568e-06, "loss": 0.5439, "step": 305 }, { "epoch": 0.056064492488090874, "grad_norm": 0.5257169187612324, "learning_rate": 5.586080586080587e-06, "loss": 0.5671, "step": 306 }, { "epoch": 0.05624770978380359, "grad_norm": 0.4243472668028098, "learning_rate": 5.604395604395605e-06, "loss": 0.5044, "step": 307 }, { "epoch": 0.05643092707951631, "grad_norm": 0.43876355547814727, "learning_rate": 5.6227106227106235e-06, "loss": 0.4815, "step": 308 }, { "epoch": 0.05661414437522902, "grad_norm": 0.43342398130791976, "learning_rate": 5.641025641025641e-06, "loss": 0.5308, "step": 309 }, { "epoch": 0.05679736167094174, "grad_norm": 0.4660705177035744, "learning_rate": 5.65934065934066e-06, "loss": 0.5686, "step": 310 }, { "epoch": 0.056980578966654455, "grad_norm": 0.4170384834874546, "learning_rate": 5.677655677655678e-06, "loss": 0.4965, "step": 311 }, { "epoch": 0.057163796262367166, "grad_norm": 0.4568771189697619, "learning_rate": 5.695970695970696e-06, "loss": 0.5074, "step": 312 }, { "epoch": 0.05734701355807988, "grad_norm": 0.4666988354433752, "learning_rate": 5.7142857142857145e-06, "loss": 0.5225, "step": 313 }, { "epoch": 0.0575302308537926, "grad_norm": 0.38305953336818827, "learning_rate": 5.732600732600732e-06, "loss": 0.5057, "step": 314 }, { "epoch": 0.05771344814950531, "grad_norm": 0.8971752127635111, "learning_rate": 5.750915750915751e-06, "loss": 0.5556, "step": 315 }, { "epoch": 0.05789666544521803, "grad_norm": 0.4580428245754137, "learning_rate": 5.769230769230769e-06, "loss": 0.5344, "step": 316 }, { "epoch": 0.05807988274093075, "grad_norm": 0.414647309534276, "learning_rate": 5.7875457875457885e-06, "loss": 0.5177, "step": 317 }, { "epoch": 0.05826310003664346, "grad_norm": 0.45495854796733415, "learning_rate": 5.805860805860807e-06, "loss": 0.5214, "step": 318 }, { "epoch": 0.058446317332356175, "grad_norm": 0.47098300998769715, "learning_rate": 5.824175824175825e-06, "loss": 0.5519, "step": 319 }, { "epoch": 0.05862953462806889, "grad_norm": 0.43335718768494447, "learning_rate": 5.842490842490844e-06, "loss": 0.5282, "step": 320 }, { "epoch": 0.0588127519237816, "grad_norm": 0.4649559782288625, "learning_rate": 5.860805860805862e-06, "loss": 0.5163, "step": 321 }, { "epoch": 0.05899596921949432, "grad_norm": 0.44623794630893965, "learning_rate": 5.8791208791208794e-06, "loss": 0.514, "step": 322 }, { "epoch": 0.05917918651520704, "grad_norm": 0.4521538486948526, "learning_rate": 5.897435897435898e-06, "loss": 0.5248, "step": 323 }, { "epoch": 0.05936240381091975, "grad_norm": 0.453853173278213, "learning_rate": 5.915750915750916e-06, "loss": 0.5105, "step": 324 }, { "epoch": 0.05954562110663247, "grad_norm": 0.45615939787874893, "learning_rate": 5.934065934065935e-06, "loss": 0.5321, "step": 325 }, { "epoch": 0.059728838402345184, "grad_norm": 0.5103507598238842, "learning_rate": 5.9523809523809525e-06, "loss": 0.4889, "step": 326 }, { "epoch": 0.059912055698057895, "grad_norm": 0.4755868373014201, "learning_rate": 5.970695970695971e-06, "loss": 0.5233, "step": 327 }, { "epoch": 0.06009527299377061, "grad_norm": 0.48186711326734255, "learning_rate": 5.989010989010989e-06, "loss": 0.5216, "step": 328 }, { "epoch": 0.06027849028948333, "grad_norm": 0.406290487409726, "learning_rate": 6.007326007326008e-06, "loss": 0.5048, "step": 329 }, { "epoch": 0.06046170758519604, "grad_norm": 0.47965200235867606, "learning_rate": 6.025641025641026e-06, "loss": 0.513, "step": 330 }, { "epoch": 0.06064492488090876, "grad_norm": 0.492103246179344, "learning_rate": 6.043956043956044e-06, "loss": 0.5553, "step": 331 }, { "epoch": 0.060828142176621476, "grad_norm": 0.49171785130479284, "learning_rate": 6.062271062271062e-06, "loss": 0.5289, "step": 332 }, { "epoch": 0.06101135947233419, "grad_norm": 0.4659535600554528, "learning_rate": 6.080586080586081e-06, "loss": 0.5143, "step": 333 }, { "epoch": 0.061194576768046904, "grad_norm": 0.43136677550682173, "learning_rate": 6.0989010989011e-06, "loss": 0.469, "step": 334 }, { "epoch": 0.06137779406375962, "grad_norm": 0.6094192173431031, "learning_rate": 6.117216117216118e-06, "loss": 0.5121, "step": 335 }, { "epoch": 0.06156101135947233, "grad_norm": 0.45352827270619606, "learning_rate": 6.135531135531136e-06, "loss": 0.5097, "step": 336 }, { "epoch": 0.06174422865518505, "grad_norm": 0.42626268219917746, "learning_rate": 6.153846153846155e-06, "loss": 0.5234, "step": 337 }, { "epoch": 0.06192744595089777, "grad_norm": 0.4438479757326601, "learning_rate": 6.172161172161173e-06, "loss": 0.4852, "step": 338 }, { "epoch": 0.06211066324661048, "grad_norm": 0.688188614062373, "learning_rate": 6.1904761904761914e-06, "loss": 0.502, "step": 339 }, { "epoch": 0.062293880542323196, "grad_norm": 0.5160627484540854, "learning_rate": 6.208791208791209e-06, "loss": 0.521, "step": 340 }, { "epoch": 0.062477097838035914, "grad_norm": 0.4356067335955304, "learning_rate": 6.227106227106228e-06, "loss": 0.4918, "step": 341 }, { "epoch": 0.06266031513374863, "grad_norm": 0.5001791021027777, "learning_rate": 6.245421245421246e-06, "loss": 0.5173, "step": 342 }, { "epoch": 0.06284353242946134, "grad_norm": 0.45621279623031163, "learning_rate": 6.2637362637362645e-06, "loss": 0.5532, "step": 343 }, { "epoch": 0.06302674972517405, "grad_norm": 0.4686583276600699, "learning_rate": 6.282051282051282e-06, "loss": 0.5544, "step": 344 }, { "epoch": 0.06320996702088677, "grad_norm": 0.4500457038114704, "learning_rate": 6.300366300366301e-06, "loss": 0.5094, "step": 345 }, { "epoch": 0.06339318431659949, "grad_norm": 0.4707435610591325, "learning_rate": 6.318681318681319e-06, "loss": 0.5317, "step": 346 }, { "epoch": 0.0635764016123122, "grad_norm": 0.44910822533973516, "learning_rate": 6.336996336996337e-06, "loss": 0.5063, "step": 347 }, { "epoch": 0.06375961890802492, "grad_norm": 0.4262957774336457, "learning_rate": 6.3553113553113555e-06, "loss": 0.5389, "step": 348 }, { "epoch": 0.06394283620373763, "grad_norm": 0.4958715530578741, "learning_rate": 6.373626373626373e-06, "loss": 0.5253, "step": 349 }, { "epoch": 0.06412605349945034, "grad_norm": 0.4706503863353741, "learning_rate": 6.391941391941392e-06, "loss": 0.5078, "step": 350 }, { "epoch": 0.06430927079516306, "grad_norm": 0.5025375698465898, "learning_rate": 6.410256410256412e-06, "loss": 0.5366, "step": 351 }, { "epoch": 0.06449248809087578, "grad_norm": 0.43307402952194485, "learning_rate": 6.4285714285714295e-06, "loss": 0.5053, "step": 352 }, { "epoch": 0.0646757053865885, "grad_norm": 0.4145957842766705, "learning_rate": 6.446886446886448e-06, "loss": 0.5134, "step": 353 }, { "epoch": 0.06485892268230121, "grad_norm": 0.42558146560441634, "learning_rate": 6.465201465201466e-06, "loss": 0.498, "step": 354 }, { "epoch": 0.06504213997801392, "grad_norm": 0.6808736114735602, "learning_rate": 6.483516483516485e-06, "loss": 0.4865, "step": 355 }, { "epoch": 0.06522535727372664, "grad_norm": 0.4858578720351213, "learning_rate": 6.5018315018315026e-06, "loss": 0.4977, "step": 356 }, { "epoch": 0.06540857456943935, "grad_norm": 0.5523209851617275, "learning_rate": 6.5201465201465204e-06, "loss": 0.5278, "step": 357 }, { "epoch": 0.06559179186515207, "grad_norm": 0.4756243917905379, "learning_rate": 6.538461538461539e-06, "loss": 0.5018, "step": 358 }, { "epoch": 0.06577500916086479, "grad_norm": 0.5270789359516691, "learning_rate": 6.556776556776557e-06, "loss": 0.5127, "step": 359 }, { "epoch": 0.0659582264565775, "grad_norm": 0.5213313487503423, "learning_rate": 6.575091575091576e-06, "loss": 0.504, "step": 360 }, { "epoch": 0.06614144375229021, "grad_norm": 0.47699985237826076, "learning_rate": 6.5934065934065935e-06, "loss": 0.5356, "step": 361 }, { "epoch": 0.06632466104800293, "grad_norm": 0.4398115760336562, "learning_rate": 6.611721611721612e-06, "loss": 0.5038, "step": 362 }, { "epoch": 0.06650787834371565, "grad_norm": 0.5848808647247892, "learning_rate": 6.63003663003663e-06, "loss": 0.5393, "step": 363 }, { "epoch": 0.06669109563942836, "grad_norm": 0.40297449884724584, "learning_rate": 6.648351648351649e-06, "loss": 0.5231, "step": 364 }, { "epoch": 0.06687431293514108, "grad_norm": 0.4615954062532406, "learning_rate": 6.666666666666667e-06, "loss": 0.5017, "step": 365 }, { "epoch": 0.0670575302308538, "grad_norm": 0.4204668889117905, "learning_rate": 6.684981684981685e-06, "loss": 0.5651, "step": 366 }, { "epoch": 0.0672407475265665, "grad_norm": 0.4765525618556606, "learning_rate": 6.703296703296703e-06, "loss": 0.5253, "step": 367 }, { "epoch": 0.06742396482227922, "grad_norm": 0.431663431335195, "learning_rate": 6.721611721611723e-06, "loss": 0.5398, "step": 368 }, { "epoch": 0.06760718211799194, "grad_norm": 0.4403652862248983, "learning_rate": 6.739926739926741e-06, "loss": 0.5252, "step": 369 }, { "epoch": 0.06779039941370466, "grad_norm": 0.483725795531489, "learning_rate": 6.758241758241759e-06, "loss": 0.5349, "step": 370 }, { "epoch": 0.06797361670941737, "grad_norm": 0.8892216780264888, "learning_rate": 6.776556776556777e-06, "loss": 0.5184, "step": 371 }, { "epoch": 0.06815683400513009, "grad_norm": 0.48775432933817636, "learning_rate": 6.794871794871796e-06, "loss": 0.5275, "step": 372 }, { "epoch": 0.0683400513008428, "grad_norm": 0.4772597017011689, "learning_rate": 6.813186813186814e-06, "loss": 0.5517, "step": 373 }, { "epoch": 0.06852326859655551, "grad_norm": 0.46418441159832313, "learning_rate": 6.831501831501832e-06, "loss": 0.5429, "step": 374 }, { "epoch": 0.06870648589226823, "grad_norm": 0.45474945807496187, "learning_rate": 6.84981684981685e-06, "loss": 0.4999, "step": 375 }, { "epoch": 0.06888970318798095, "grad_norm": 0.5576640330118253, "learning_rate": 6.868131868131869e-06, "loss": 0.5198, "step": 376 }, { "epoch": 0.06907292048369366, "grad_norm": 0.4921371781399045, "learning_rate": 6.886446886446887e-06, "loss": 0.5379, "step": 377 }, { "epoch": 0.06925613777940638, "grad_norm": 0.41639097386056484, "learning_rate": 6.9047619047619055e-06, "loss": 0.519, "step": 378 }, { "epoch": 0.06943935507511909, "grad_norm": 0.4778909271586564, "learning_rate": 6.923076923076923e-06, "loss": 0.5336, "step": 379 }, { "epoch": 0.0696225723708318, "grad_norm": 0.4735414706699025, "learning_rate": 6.941391941391942e-06, "loss": 0.5, "step": 380 }, { "epoch": 0.06980578966654452, "grad_norm": 0.5646322357431656, "learning_rate": 6.95970695970696e-06, "loss": 0.5073, "step": 381 }, { "epoch": 0.06998900696225724, "grad_norm": 0.48231469572609614, "learning_rate": 6.978021978021979e-06, "loss": 0.519, "step": 382 }, { "epoch": 0.07017222425796996, "grad_norm": 0.5051213072897166, "learning_rate": 6.9963369963369965e-06, "loss": 0.5315, "step": 383 }, { "epoch": 0.07035544155368267, "grad_norm": 0.5166250853237789, "learning_rate": 7.014652014652014e-06, "loss": 0.4978, "step": 384 }, { "epoch": 0.07053865884939538, "grad_norm": 0.43900982727967125, "learning_rate": 7.032967032967034e-06, "loss": 0.4913, "step": 385 }, { "epoch": 0.0707218761451081, "grad_norm": 0.5262819691743885, "learning_rate": 7.051282051282053e-06, "loss": 0.5189, "step": 386 }, { "epoch": 0.07090509344082081, "grad_norm": 0.4906825009780104, "learning_rate": 7.0695970695970705e-06, "loss": 0.5586, "step": 387 }, { "epoch": 0.07108831073653353, "grad_norm": 0.48825547184252527, "learning_rate": 7.087912087912089e-06, "loss": 0.5218, "step": 388 }, { "epoch": 0.07127152803224625, "grad_norm": 0.5005437189224704, "learning_rate": 7.106227106227107e-06, "loss": 0.5088, "step": 389 }, { "epoch": 0.07145474532795897, "grad_norm": 0.4820090004987099, "learning_rate": 7.124542124542126e-06, "loss": 0.5132, "step": 390 }, { "epoch": 0.07163796262367167, "grad_norm": 0.46585258023856246, "learning_rate": 7.1428571428571436e-06, "loss": 0.5198, "step": 391 }, { "epoch": 0.07182117991938439, "grad_norm": 0.5351346796257885, "learning_rate": 7.161172161172162e-06, "loss": 0.5174, "step": 392 }, { "epoch": 0.0720043972150971, "grad_norm": 0.505980784032641, "learning_rate": 7.17948717948718e-06, "loss": 0.517, "step": 393 }, { "epoch": 0.07218761451080982, "grad_norm": 0.3991493735258601, "learning_rate": 7.197802197802198e-06, "loss": 0.5193, "step": 394 }, { "epoch": 0.07237083180652254, "grad_norm": 0.47050897190858826, "learning_rate": 7.216117216117217e-06, "loss": 0.5278, "step": 395 }, { "epoch": 0.07255404910223526, "grad_norm": 0.42709223237684557, "learning_rate": 7.2344322344322345e-06, "loss": 0.5233, "step": 396 }, { "epoch": 0.07273726639794796, "grad_norm": 0.40701893214503493, "learning_rate": 7.252747252747253e-06, "loss": 0.4594, "step": 397 }, { "epoch": 0.07292048369366068, "grad_norm": 0.4793723507064191, "learning_rate": 7.271062271062271e-06, "loss": 0.5022, "step": 398 }, { "epoch": 0.0731037009893734, "grad_norm": 0.4084855491847111, "learning_rate": 7.28937728937729e-06, "loss": 0.5177, "step": 399 }, { "epoch": 0.07328691828508611, "grad_norm": 0.48791204362548757, "learning_rate": 7.307692307692308e-06, "loss": 0.5336, "step": 400 }, { "epoch": 0.07347013558079883, "grad_norm": 0.4721917787188149, "learning_rate": 7.326007326007326e-06, "loss": 0.5291, "step": 401 }, { "epoch": 0.07365335287651155, "grad_norm": 0.4482564665450733, "learning_rate": 7.344322344322346e-06, "loss": 0.516, "step": 402 }, { "epoch": 0.07383657017222425, "grad_norm": 0.503350703275625, "learning_rate": 7.362637362637364e-06, "loss": 0.5232, "step": 403 }, { "epoch": 0.07401978746793697, "grad_norm": 0.5152368271826678, "learning_rate": 7.380952380952382e-06, "loss": 0.5258, "step": 404 }, { "epoch": 0.07420300476364969, "grad_norm": 0.4523690267979874, "learning_rate": 7.3992673992674e-06, "loss": 0.5007, "step": 405 }, { "epoch": 0.0743862220593624, "grad_norm": 0.44122338596098665, "learning_rate": 7.417582417582418e-06, "loss": 0.5325, "step": 406 }, { "epoch": 0.07456943935507512, "grad_norm": 0.418021168922302, "learning_rate": 7.435897435897437e-06, "loss": 0.5048, "step": 407 }, { "epoch": 0.07475265665078784, "grad_norm": 0.44537098554090276, "learning_rate": 7.454212454212455e-06, "loss": 0.5294, "step": 408 }, { "epoch": 0.07493587394650054, "grad_norm": 0.41384101894590325, "learning_rate": 7.472527472527473e-06, "loss": 0.5119, "step": 409 }, { "epoch": 0.07511909124221326, "grad_norm": 0.4647470180215214, "learning_rate": 7.490842490842491e-06, "loss": 0.5202, "step": 410 }, { "epoch": 0.07530230853792598, "grad_norm": 0.4836409216489063, "learning_rate": 7.50915750915751e-06, "loss": 0.5056, "step": 411 }, { "epoch": 0.0754855258336387, "grad_norm": 0.45195196209865357, "learning_rate": 7.527472527472528e-06, "loss": 0.5213, "step": 412 }, { "epoch": 0.07566874312935142, "grad_norm": 0.5036829282617328, "learning_rate": 7.5457875457875465e-06, "loss": 0.5347, "step": 413 }, { "epoch": 0.07585196042506413, "grad_norm": 0.4666038283310089, "learning_rate": 7.564102564102564e-06, "loss": 0.5151, "step": 414 }, { "epoch": 0.07603517772077684, "grad_norm": 0.5067769344340564, "learning_rate": 7.582417582417583e-06, "loss": 0.5171, "step": 415 }, { "epoch": 0.07621839501648955, "grad_norm": 0.4093230258713508, "learning_rate": 7.600732600732601e-06, "loss": 0.5039, "step": 416 }, { "epoch": 0.07640161231220227, "grad_norm": 0.6498006485071868, "learning_rate": 7.61904761904762e-06, "loss": 0.5232, "step": 417 }, { "epoch": 0.07658482960791499, "grad_norm": 0.4558152414900924, "learning_rate": 7.637362637362638e-06, "loss": 0.5232, "step": 418 }, { "epoch": 0.0767680469036277, "grad_norm": 0.4804930639066785, "learning_rate": 7.655677655677656e-06, "loss": 0.5088, "step": 419 }, { "epoch": 0.07695126419934042, "grad_norm": 0.5081120095319347, "learning_rate": 7.673992673992676e-06, "loss": 0.5308, "step": 420 }, { "epoch": 0.07713448149505313, "grad_norm": 0.42393212923875745, "learning_rate": 7.692307692307694e-06, "loss": 0.5084, "step": 421 }, { "epoch": 0.07731769879076585, "grad_norm": 0.44672075335716155, "learning_rate": 7.710622710622711e-06, "loss": 0.483, "step": 422 }, { "epoch": 0.07750091608647856, "grad_norm": 0.4942211324726324, "learning_rate": 7.72893772893773e-06, "loss": 0.5236, "step": 423 }, { "epoch": 0.07768413338219128, "grad_norm": 0.39817279842841763, "learning_rate": 7.747252747252749e-06, "loss": 0.5332, "step": 424 }, { "epoch": 0.077867350677904, "grad_norm": 0.4608671155196199, "learning_rate": 7.765567765567767e-06, "loss": 0.523, "step": 425 }, { "epoch": 0.07805056797361672, "grad_norm": 0.5225991254379231, "learning_rate": 7.783882783882785e-06, "loss": 0.5171, "step": 426 }, { "epoch": 0.07823378526932942, "grad_norm": 0.5423490196220466, "learning_rate": 7.802197802197802e-06, "loss": 0.4985, "step": 427 }, { "epoch": 0.07841700256504214, "grad_norm": 0.502419506962013, "learning_rate": 7.820512820512822e-06, "loss": 0.5087, "step": 428 }, { "epoch": 0.07860021986075486, "grad_norm": 0.5102984182578104, "learning_rate": 7.83882783882784e-06, "loss": 0.5358, "step": 429 }, { "epoch": 0.07878343715646757, "grad_norm": 0.4742126281062651, "learning_rate": 7.857142857142858e-06, "loss": 0.5585, "step": 430 }, { "epoch": 0.07896665445218029, "grad_norm": 0.508787409495283, "learning_rate": 7.875457875457876e-06, "loss": 0.5411, "step": 431 }, { "epoch": 0.079149871747893, "grad_norm": 0.4559258370924022, "learning_rate": 7.893772893772893e-06, "loss": 0.5229, "step": 432 }, { "epoch": 0.07933308904360571, "grad_norm": 0.6320276655690735, "learning_rate": 7.912087912087913e-06, "loss": 0.4688, "step": 433 }, { "epoch": 0.07951630633931843, "grad_norm": 0.9797165625621964, "learning_rate": 7.93040293040293e-06, "loss": 0.5372, "step": 434 }, { "epoch": 0.07969952363503115, "grad_norm": 0.43648526458810954, "learning_rate": 7.948717948717949e-06, "loss": 0.4528, "step": 435 }, { "epoch": 0.07988274093074386, "grad_norm": 0.46101717931557235, "learning_rate": 7.967032967032966e-06, "loss": 0.4836, "step": 436 }, { "epoch": 0.08006595822645658, "grad_norm": 0.46019219132505085, "learning_rate": 7.985347985347986e-06, "loss": 0.4874, "step": 437 }, { "epoch": 0.08024917552216929, "grad_norm": 0.4228905699110607, "learning_rate": 8.003663003663006e-06, "loss": 0.483, "step": 438 }, { "epoch": 0.080432392817882, "grad_norm": 0.4920935432281776, "learning_rate": 8.021978021978023e-06, "loss": 0.489, "step": 439 }, { "epoch": 0.08061561011359472, "grad_norm": 0.45385267234753407, "learning_rate": 8.040293040293041e-06, "loss": 0.5262, "step": 440 }, { "epoch": 0.08079882740930744, "grad_norm": 0.48390990285633756, "learning_rate": 8.058608058608059e-06, "loss": 0.4969, "step": 441 }, { "epoch": 0.08098204470502016, "grad_norm": 0.4424118373260341, "learning_rate": 8.076923076923077e-06, "loss": 0.5217, "step": 442 }, { "epoch": 0.08116526200073287, "grad_norm": 0.4647327426734162, "learning_rate": 8.095238095238097e-06, "loss": 0.5103, "step": 443 }, { "epoch": 0.08134847929644558, "grad_norm": 0.42019112334691044, "learning_rate": 8.113553113553114e-06, "loss": 0.5378, "step": 444 }, { "epoch": 0.0815316965921583, "grad_norm": 0.40127483204178854, "learning_rate": 8.131868131868132e-06, "loss": 0.4813, "step": 445 }, { "epoch": 0.08171491388787101, "grad_norm": 0.4606579465800615, "learning_rate": 8.15018315018315e-06, "loss": 0.5067, "step": 446 }, { "epoch": 0.08189813118358373, "grad_norm": 0.4715535246345655, "learning_rate": 8.16849816849817e-06, "loss": 0.5181, "step": 447 }, { "epoch": 0.08208134847929645, "grad_norm": 0.5386430386856308, "learning_rate": 8.186813186813188e-06, "loss": 0.5188, "step": 448 }, { "epoch": 0.08226456577500917, "grad_norm": 0.5389030998275688, "learning_rate": 8.205128205128205e-06, "loss": 0.516, "step": 449 }, { "epoch": 0.08244778307072187, "grad_norm": 0.5181954163327931, "learning_rate": 8.223443223443223e-06, "loss": 0.5099, "step": 450 }, { "epoch": 0.08263100036643459, "grad_norm": 0.5425277965290457, "learning_rate": 8.241758241758243e-06, "loss": 0.5508, "step": 451 }, { "epoch": 0.0828142176621473, "grad_norm": 0.4366152129754919, "learning_rate": 8.26007326007326e-06, "loss": 0.4772, "step": 452 }, { "epoch": 0.08299743495786002, "grad_norm": 0.45607475441878936, "learning_rate": 8.278388278388278e-06, "loss": 0.5319, "step": 453 }, { "epoch": 0.08318065225357274, "grad_norm": 0.4712644997156548, "learning_rate": 8.296703296703298e-06, "loss": 0.5477, "step": 454 }, { "epoch": 0.08336386954928546, "grad_norm": 0.6856852588986854, "learning_rate": 8.315018315018316e-06, "loss": 0.5235, "step": 455 }, { "epoch": 0.08354708684499816, "grad_norm": 0.5019345602931549, "learning_rate": 8.333333333333334e-06, "loss": 0.5101, "step": 456 }, { "epoch": 0.08373030414071088, "grad_norm": 0.4155972768165925, "learning_rate": 8.351648351648353e-06, "loss": 0.4986, "step": 457 }, { "epoch": 0.0839135214364236, "grad_norm": 0.45725631484653695, "learning_rate": 8.369963369963371e-06, "loss": 0.5233, "step": 458 }, { "epoch": 0.08409673873213631, "grad_norm": 0.5500464841633228, "learning_rate": 8.388278388278389e-06, "loss": 0.5225, "step": 459 }, { "epoch": 0.08427995602784903, "grad_norm": 0.5210381864086956, "learning_rate": 8.406593406593407e-06, "loss": 0.5509, "step": 460 }, { "epoch": 0.08446317332356175, "grad_norm": 0.4559821107303101, "learning_rate": 8.424908424908426e-06, "loss": 0.5118, "step": 461 }, { "epoch": 0.08464639061927445, "grad_norm": 0.48323545775926524, "learning_rate": 8.443223443223444e-06, "loss": 0.5049, "step": 462 }, { "epoch": 0.08482960791498717, "grad_norm": 0.4812859935236779, "learning_rate": 8.461538461538462e-06, "loss": 0.5204, "step": 463 }, { "epoch": 0.08501282521069989, "grad_norm": 0.454474341410942, "learning_rate": 8.47985347985348e-06, "loss": 0.5387, "step": 464 }, { "epoch": 0.0851960425064126, "grad_norm": 0.6615650181808158, "learning_rate": 8.498168498168498e-06, "loss": 0.5393, "step": 465 }, { "epoch": 0.08537925980212532, "grad_norm": 0.4355325124699567, "learning_rate": 8.516483516483517e-06, "loss": 0.5027, "step": 466 }, { "epoch": 0.08556247709783804, "grad_norm": 0.43977659231205646, "learning_rate": 8.534798534798535e-06, "loss": 0.4851, "step": 467 }, { "epoch": 0.08574569439355074, "grad_norm": 0.5490378411527629, "learning_rate": 8.553113553113553e-06, "loss": 0.5087, "step": 468 }, { "epoch": 0.08592891168926346, "grad_norm": 0.4334113007101785, "learning_rate": 8.571428571428571e-06, "loss": 0.4874, "step": 469 }, { "epoch": 0.08611212898497618, "grad_norm": 0.48093462345535853, "learning_rate": 8.58974358974359e-06, "loss": 0.5103, "step": 470 }, { "epoch": 0.0862953462806889, "grad_norm": 0.509104778628553, "learning_rate": 8.60805860805861e-06, "loss": 0.5125, "step": 471 }, { "epoch": 0.08647856357640162, "grad_norm": 0.46329459466937095, "learning_rate": 8.626373626373628e-06, "loss": 0.4862, "step": 472 }, { "epoch": 0.08666178087211433, "grad_norm": 0.41221910536829426, "learning_rate": 8.644688644688646e-06, "loss": 0.5058, "step": 473 }, { "epoch": 0.08684499816782704, "grad_norm": 0.4965957595147397, "learning_rate": 8.663003663003664e-06, "loss": 0.5045, "step": 474 }, { "epoch": 0.08702821546353975, "grad_norm": 0.6831024596734047, "learning_rate": 8.681318681318681e-06, "loss": 0.533, "step": 475 }, { "epoch": 0.08721143275925247, "grad_norm": 0.44117689667679433, "learning_rate": 8.699633699633701e-06, "loss": 0.494, "step": 476 }, { "epoch": 0.08739465005496519, "grad_norm": 0.49552902332130444, "learning_rate": 8.717948717948719e-06, "loss": 0.5295, "step": 477 }, { "epoch": 0.08757786735067791, "grad_norm": 0.5202864223018397, "learning_rate": 8.736263736263737e-06, "loss": 0.5315, "step": 478 }, { "epoch": 0.08776108464639062, "grad_norm": 0.4531456591891388, "learning_rate": 8.754578754578755e-06, "loss": 0.5177, "step": 479 }, { "epoch": 0.08794430194210333, "grad_norm": 0.5060531781359298, "learning_rate": 8.772893772893774e-06, "loss": 0.5422, "step": 480 }, { "epoch": 0.08812751923781605, "grad_norm": 0.44222377285427994, "learning_rate": 8.791208791208792e-06, "loss": 0.4679, "step": 481 }, { "epoch": 0.08831073653352876, "grad_norm": 0.4308135949373974, "learning_rate": 8.80952380952381e-06, "loss": 0.5057, "step": 482 }, { "epoch": 0.08849395382924148, "grad_norm": 0.5107147323097014, "learning_rate": 8.827838827838828e-06, "loss": 0.5147, "step": 483 }, { "epoch": 0.0886771711249542, "grad_norm": 0.4722359038239213, "learning_rate": 8.846153846153847e-06, "loss": 0.5069, "step": 484 }, { "epoch": 0.08886038842066692, "grad_norm": 0.4913307139634766, "learning_rate": 8.864468864468865e-06, "loss": 0.4943, "step": 485 }, { "epoch": 0.08904360571637962, "grad_norm": 0.47381589638250404, "learning_rate": 8.882783882783883e-06, "loss": 0.5432, "step": 486 }, { "epoch": 0.08922682301209234, "grad_norm": 0.4763727220823097, "learning_rate": 8.9010989010989e-06, "loss": 0.4765, "step": 487 }, { "epoch": 0.08941004030780506, "grad_norm": 0.48581942940784734, "learning_rate": 8.91941391941392e-06, "loss": 0.4918, "step": 488 }, { "epoch": 0.08959325760351777, "grad_norm": 0.4771435540271998, "learning_rate": 8.937728937728938e-06, "loss": 0.5146, "step": 489 }, { "epoch": 0.08977647489923049, "grad_norm": 0.5554558590150881, "learning_rate": 8.956043956043958e-06, "loss": 0.5139, "step": 490 }, { "epoch": 0.08995969219494321, "grad_norm": 0.4409852140988399, "learning_rate": 8.974358974358976e-06, "loss": 0.494, "step": 491 }, { "epoch": 0.09014290949065591, "grad_norm": 0.470617109732078, "learning_rate": 8.992673992673993e-06, "loss": 0.5293, "step": 492 }, { "epoch": 0.09032612678636863, "grad_norm": 0.48719044521659705, "learning_rate": 9.010989010989011e-06, "loss": 0.5111, "step": 493 }, { "epoch": 0.09050934408208135, "grad_norm": 0.46876439756978366, "learning_rate": 9.02930402930403e-06, "loss": 0.4942, "step": 494 }, { "epoch": 0.09069256137779406, "grad_norm": 0.46671821534033575, "learning_rate": 9.047619047619049e-06, "loss": 0.5617, "step": 495 }, { "epoch": 0.09087577867350678, "grad_norm": 0.4868372988815388, "learning_rate": 9.065934065934067e-06, "loss": 0.5195, "step": 496 }, { "epoch": 0.0910589959692195, "grad_norm": 0.47526685976460153, "learning_rate": 9.084249084249084e-06, "loss": 0.4561, "step": 497 }, { "epoch": 0.0912422132649322, "grad_norm": 0.526795639153496, "learning_rate": 9.102564102564104e-06, "loss": 0.4905, "step": 498 }, { "epoch": 0.09142543056064492, "grad_norm": 0.47935527575760656, "learning_rate": 9.120879120879122e-06, "loss": 0.5129, "step": 499 }, { "epoch": 0.09160864785635764, "grad_norm": 0.47020582683735346, "learning_rate": 9.13919413919414e-06, "loss": 0.511, "step": 500 }, { "epoch": 0.09179186515207036, "grad_norm": 0.46760420885699033, "learning_rate": 9.157509157509158e-06, "loss": 0.4789, "step": 501 }, { "epoch": 0.09197508244778307, "grad_norm": 0.4695251283934706, "learning_rate": 9.175824175824175e-06, "loss": 0.5351, "step": 502 }, { "epoch": 0.09215829974349579, "grad_norm": 0.5153360795669553, "learning_rate": 9.194139194139195e-06, "loss": 0.5522, "step": 503 }, { "epoch": 0.0923415170392085, "grad_norm": 0.44806081150815513, "learning_rate": 9.212454212454213e-06, "loss": 0.5173, "step": 504 }, { "epoch": 0.09252473433492121, "grad_norm": 0.451536504838503, "learning_rate": 9.230769230769232e-06, "loss": 0.5082, "step": 505 }, { "epoch": 0.09270795163063393, "grad_norm": 0.48436384147084766, "learning_rate": 9.24908424908425e-06, "loss": 0.5388, "step": 506 }, { "epoch": 0.09289116892634665, "grad_norm": 0.4609113225702784, "learning_rate": 9.267399267399268e-06, "loss": 0.4892, "step": 507 }, { "epoch": 0.09307438622205937, "grad_norm": 0.4098711492197863, "learning_rate": 9.285714285714288e-06, "loss": 0.4856, "step": 508 }, { "epoch": 0.09325760351777208, "grad_norm": 0.4715675907322317, "learning_rate": 9.304029304029305e-06, "loss": 0.53, "step": 509 }, { "epoch": 0.09344082081348479, "grad_norm": 0.5026574189450673, "learning_rate": 9.322344322344323e-06, "loss": 0.5314, "step": 510 }, { "epoch": 0.0936240381091975, "grad_norm": 0.4253199666429807, "learning_rate": 9.340659340659341e-06, "loss": 0.5068, "step": 511 }, { "epoch": 0.09380725540491022, "grad_norm": 0.45894485399853974, "learning_rate": 9.358974358974359e-06, "loss": 0.5235, "step": 512 }, { "epoch": 0.09399047270062294, "grad_norm": 0.7513225377427756, "learning_rate": 9.377289377289379e-06, "loss": 0.5071, "step": 513 }, { "epoch": 0.09417368999633566, "grad_norm": 0.5231017027547264, "learning_rate": 9.395604395604396e-06, "loss": 0.5339, "step": 514 }, { "epoch": 0.09435690729204838, "grad_norm": 0.5146997140217462, "learning_rate": 9.413919413919414e-06, "loss": 0.5052, "step": 515 }, { "epoch": 0.09454012458776108, "grad_norm": 0.47760493192836756, "learning_rate": 9.432234432234432e-06, "loss": 0.5102, "step": 516 }, { "epoch": 0.0947233418834738, "grad_norm": 0.44117590454697736, "learning_rate": 9.450549450549452e-06, "loss": 0.5237, "step": 517 }, { "epoch": 0.09490655917918651, "grad_norm": 0.5004680185091399, "learning_rate": 9.46886446886447e-06, "loss": 0.5243, "step": 518 }, { "epoch": 0.09508977647489923, "grad_norm": 0.4549800276149351, "learning_rate": 9.487179487179487e-06, "loss": 0.5139, "step": 519 }, { "epoch": 0.09527299377061195, "grad_norm": 0.4263327588755031, "learning_rate": 9.505494505494505e-06, "loss": 0.4877, "step": 520 }, { "epoch": 0.09545621106632467, "grad_norm": 0.4484968491218323, "learning_rate": 9.523809523809525e-06, "loss": 0.4986, "step": 521 }, { "epoch": 0.09563942836203737, "grad_norm": 0.39390345192807547, "learning_rate": 9.542124542124543e-06, "loss": 0.5116, "step": 522 }, { "epoch": 0.09582264565775009, "grad_norm": 0.4529147798983982, "learning_rate": 9.560439560439562e-06, "loss": 0.5312, "step": 523 }, { "epoch": 0.0960058629534628, "grad_norm": 0.49701803503982106, "learning_rate": 9.57875457875458e-06, "loss": 0.529, "step": 524 }, { "epoch": 0.09618908024917552, "grad_norm": 0.4097214793479841, "learning_rate": 9.597069597069598e-06, "loss": 0.5005, "step": 525 }, { "epoch": 0.09637229754488824, "grad_norm": 0.5148151884655592, "learning_rate": 9.615384615384616e-06, "loss": 0.53, "step": 526 }, { "epoch": 0.09655551484060096, "grad_norm": 0.4264857545219357, "learning_rate": 9.633699633699635e-06, "loss": 0.5136, "step": 527 }, { "epoch": 0.09673873213631366, "grad_norm": 0.45426464252638443, "learning_rate": 9.652014652014653e-06, "loss": 0.5117, "step": 528 }, { "epoch": 0.09692194943202638, "grad_norm": 0.47571176071645493, "learning_rate": 9.670329670329671e-06, "loss": 0.5185, "step": 529 }, { "epoch": 0.0971051667277391, "grad_norm": 0.42418752426373113, "learning_rate": 9.688644688644689e-06, "loss": 0.5285, "step": 530 }, { "epoch": 0.09728838402345182, "grad_norm": 0.48590348616099827, "learning_rate": 9.706959706959708e-06, "loss": 0.5339, "step": 531 }, { "epoch": 0.09747160131916453, "grad_norm": 0.463774513888958, "learning_rate": 9.725274725274726e-06, "loss": 0.5386, "step": 532 }, { "epoch": 0.09765481861487725, "grad_norm": 0.4891808396070614, "learning_rate": 9.743589743589744e-06, "loss": 0.5326, "step": 533 }, { "epoch": 0.09783803591058995, "grad_norm": 0.40841004082693305, "learning_rate": 9.761904761904762e-06, "loss": 0.5085, "step": 534 }, { "epoch": 0.09802125320630267, "grad_norm": 0.42494728872814735, "learning_rate": 9.780219780219781e-06, "loss": 0.4751, "step": 535 }, { "epoch": 0.09820447050201539, "grad_norm": 0.4594283822364781, "learning_rate": 9.7985347985348e-06, "loss": 0.5275, "step": 536 }, { "epoch": 0.09838768779772811, "grad_norm": 0.4064302259117676, "learning_rate": 9.816849816849817e-06, "loss": 0.5089, "step": 537 }, { "epoch": 0.09857090509344082, "grad_norm": 0.4860027724973198, "learning_rate": 9.835164835164835e-06, "loss": 0.505, "step": 538 }, { "epoch": 0.09875412238915354, "grad_norm": 0.47625362870716265, "learning_rate": 9.853479853479855e-06, "loss": 0.4954, "step": 539 }, { "epoch": 0.09893733968486625, "grad_norm": 0.4592940212317861, "learning_rate": 9.871794871794872e-06, "loss": 0.4871, "step": 540 }, { "epoch": 0.09912055698057896, "grad_norm": 0.42717127053329573, "learning_rate": 9.890109890109892e-06, "loss": 0.5261, "step": 541 }, { "epoch": 0.09930377427629168, "grad_norm": 0.43970535809042904, "learning_rate": 9.90842490842491e-06, "loss": 0.4866, "step": 542 }, { "epoch": 0.0994869915720044, "grad_norm": 0.42143360797263807, "learning_rate": 9.926739926739928e-06, "loss": 0.5113, "step": 543 }, { "epoch": 0.09967020886771712, "grad_norm": 0.4416711016318173, "learning_rate": 9.945054945054946e-06, "loss": 0.5116, "step": 544 }, { "epoch": 0.09985342616342983, "grad_norm": 0.45325267149265236, "learning_rate": 9.963369963369965e-06, "loss": 0.5283, "step": 545 }, { "epoch": 0.10003664345914254, "grad_norm": 0.3948707773420248, "learning_rate": 9.981684981684983e-06, "loss": 0.5114, "step": 546 }, { "epoch": 0.10021986075485526, "grad_norm": 0.42983426021717475, "learning_rate": 1e-05, "loss": 0.5169, "step": 547 }, { "epoch": 0.10040307805056797, "grad_norm": 0.5119731461361727, "learning_rate": 9.999998977359419e-06, "loss": 0.518, "step": 548 }, { "epoch": 0.10058629534628069, "grad_norm": 0.48479775231872485, "learning_rate": 9.999995909438092e-06, "loss": 0.5207, "step": 549 }, { "epoch": 0.10076951264199341, "grad_norm": 0.45005420614602115, "learning_rate": 9.999990796237274e-06, "loss": 0.4878, "step": 550 }, { "epoch": 0.10095272993770613, "grad_norm": 0.40595587220813967, "learning_rate": 9.999983637759059e-06, "loss": 0.4712, "step": 551 }, { "epoch": 0.10113594723341883, "grad_norm": 0.43893870910937327, "learning_rate": 9.999974434006372e-06, "loss": 0.4964, "step": 552 }, { "epoch": 0.10131916452913155, "grad_norm": 0.4638079942473788, "learning_rate": 9.99996318498298e-06, "loss": 0.4834, "step": 553 }, { "epoch": 0.10150238182484426, "grad_norm": 0.45078724585599195, "learning_rate": 9.999949890693484e-06, "loss": 0.4828, "step": 554 }, { "epoch": 0.10168559912055698, "grad_norm": 0.4968230440871285, "learning_rate": 9.999934551143319e-06, "loss": 0.5146, "step": 555 }, { "epoch": 0.1018688164162697, "grad_norm": 0.4431824586567993, "learning_rate": 9.999917166338767e-06, "loss": 0.5117, "step": 556 }, { "epoch": 0.10205203371198242, "grad_norm": 0.40809319706771474, "learning_rate": 9.999897736286932e-06, "loss": 0.4942, "step": 557 }, { "epoch": 0.10223525100769512, "grad_norm": 0.46904649406632837, "learning_rate": 9.999876260995767e-06, "loss": 0.5223, "step": 558 }, { "epoch": 0.10241846830340784, "grad_norm": 0.45993616230767786, "learning_rate": 9.999852740474054e-06, "loss": 0.498, "step": 559 }, { "epoch": 0.10260168559912056, "grad_norm": 0.41328723379038895, "learning_rate": 9.999827174731414e-06, "loss": 0.5294, "step": 560 }, { "epoch": 0.10278490289483327, "grad_norm": 0.4584774156426854, "learning_rate": 9.999799563778307e-06, "loss": 0.5138, "step": 561 }, { "epoch": 0.10296812019054599, "grad_norm": 0.4626480547701558, "learning_rate": 9.999769907626024e-06, "loss": 0.5122, "step": 562 }, { "epoch": 0.10315133748625871, "grad_norm": 0.4647555669122866, "learning_rate": 9.9997382062867e-06, "loss": 0.5276, "step": 563 }, { "epoch": 0.10333455478197141, "grad_norm": 0.41824885855037686, "learning_rate": 9.9997044597733e-06, "loss": 0.4976, "step": 564 }, { "epoch": 0.10351777207768413, "grad_norm": 0.41187305924650414, "learning_rate": 9.999668668099628e-06, "loss": 0.4953, "step": 565 }, { "epoch": 0.10370098937339685, "grad_norm": 0.42660595756816655, "learning_rate": 9.999630831280329e-06, "loss": 0.514, "step": 566 }, { "epoch": 0.10388420666910957, "grad_norm": 0.5099786508796531, "learning_rate": 9.999590949330876e-06, "loss": 0.5038, "step": 567 }, { "epoch": 0.10406742396482228, "grad_norm": 0.4926825490754442, "learning_rate": 9.999549022267582e-06, "loss": 0.5535, "step": 568 }, { "epoch": 0.104250641260535, "grad_norm": 0.4578490443823955, "learning_rate": 9.999505050107601e-06, "loss": 0.5075, "step": 569 }, { "epoch": 0.1044338585562477, "grad_norm": 0.4733323605035433, "learning_rate": 9.99945903286892e-06, "loss": 0.4747, "step": 570 }, { "epoch": 0.10461707585196042, "grad_norm": 0.44893796479946446, "learning_rate": 9.999410970570358e-06, "loss": 0.4711, "step": 571 }, { "epoch": 0.10480029314767314, "grad_norm": 0.5254914900315981, "learning_rate": 9.99936086323158e-06, "loss": 0.5358, "step": 572 }, { "epoch": 0.10498351044338586, "grad_norm": 0.537923540220477, "learning_rate": 9.999308710873083e-06, "loss": 0.5079, "step": 573 }, { "epoch": 0.10516672773909858, "grad_norm": 0.45536567675754, "learning_rate": 9.999254513516196e-06, "loss": 0.4908, "step": 574 }, { "epoch": 0.10534994503481128, "grad_norm": 0.38323869490085266, "learning_rate": 9.999198271183094e-06, "loss": 0.4983, "step": 575 }, { "epoch": 0.105533162330524, "grad_norm": 0.43580085679521197, "learning_rate": 9.999139983896779e-06, "loss": 0.5031, "step": 576 }, { "epoch": 0.10571637962623671, "grad_norm": 0.4193926406628939, "learning_rate": 9.999079651681096e-06, "loss": 0.5026, "step": 577 }, { "epoch": 0.10589959692194943, "grad_norm": 0.5149327291476714, "learning_rate": 9.999017274560722e-06, "loss": 0.5414, "step": 578 }, { "epoch": 0.10608281421766215, "grad_norm": 0.44098660936462153, "learning_rate": 9.998952852561176e-06, "loss": 0.5032, "step": 579 }, { "epoch": 0.10626603151337487, "grad_norm": 0.4413320145497695, "learning_rate": 9.998886385708807e-06, "loss": 0.497, "step": 580 }, { "epoch": 0.10644924880908757, "grad_norm": 0.45575842795199323, "learning_rate": 9.998817874030808e-06, "loss": 0.5207, "step": 581 }, { "epoch": 0.10663246610480029, "grad_norm": 0.532654166955093, "learning_rate": 9.9987473175552e-06, "loss": 0.5212, "step": 582 }, { "epoch": 0.106815683400513, "grad_norm": 0.4157619606008366, "learning_rate": 9.998674716310846e-06, "loss": 0.5085, "step": 583 }, { "epoch": 0.10699890069622572, "grad_norm": 0.4868388397165033, "learning_rate": 9.998600070327444e-06, "loss": 0.5306, "step": 584 }, { "epoch": 0.10718211799193844, "grad_norm": 0.49401937562080056, "learning_rate": 9.998523379635527e-06, "loss": 0.5429, "step": 585 }, { "epoch": 0.10736533528765116, "grad_norm": 0.3986626914030523, "learning_rate": 9.99844464426647e-06, "loss": 0.4695, "step": 586 }, { "epoch": 0.10754855258336386, "grad_norm": 0.49524254606268836, "learning_rate": 9.998363864252474e-06, "loss": 0.5226, "step": 587 }, { "epoch": 0.10773176987907658, "grad_norm": 0.46186447850660334, "learning_rate": 9.998281039626588e-06, "loss": 0.5294, "step": 588 }, { "epoch": 0.1079149871747893, "grad_norm": 0.4207082230286084, "learning_rate": 9.99819617042269e-06, "loss": 0.5096, "step": 589 }, { "epoch": 0.10809820447050202, "grad_norm": 0.3955565788721964, "learning_rate": 9.998109256675496e-06, "loss": 0.4931, "step": 590 }, { "epoch": 0.10828142176621473, "grad_norm": 0.36723747707336185, "learning_rate": 9.998020298420559e-06, "loss": 0.4964, "step": 591 }, { "epoch": 0.10846463906192745, "grad_norm": 0.4185544526751163, "learning_rate": 9.997929295694266e-06, "loss": 0.4801, "step": 592 }, { "epoch": 0.10864785635764015, "grad_norm": 0.42948593775837285, "learning_rate": 9.997836248533845e-06, "loss": 0.5197, "step": 593 }, { "epoch": 0.10883107365335287, "grad_norm": 0.43966759837587943, "learning_rate": 9.997741156977356e-06, "loss": 0.5254, "step": 594 }, { "epoch": 0.10901429094906559, "grad_norm": 0.4632445570564666, "learning_rate": 9.997644021063698e-06, "loss": 0.4802, "step": 595 }, { "epoch": 0.10919750824477831, "grad_norm": 0.46480621716730125, "learning_rate": 9.997544840832604e-06, "loss": 0.5219, "step": 596 }, { "epoch": 0.10938072554049103, "grad_norm": 0.44567173311984926, "learning_rate": 9.997443616324645e-06, "loss": 0.5105, "step": 597 }, { "epoch": 0.10956394283620374, "grad_norm": 0.4534263401515789, "learning_rate": 9.997340347581226e-06, "loss": 0.5212, "step": 598 }, { "epoch": 0.10974716013191645, "grad_norm": 0.40359525147680225, "learning_rate": 9.99723503464459e-06, "loss": 0.5261, "step": 599 }, { "epoch": 0.10993037742762916, "grad_norm": 0.8305073137318505, "learning_rate": 9.99712767755782e-06, "loss": 0.5173, "step": 600 }, { "epoch": 0.11011359472334188, "grad_norm": 0.5039765167575108, "learning_rate": 9.997018276364825e-06, "loss": 0.4983, "step": 601 }, { "epoch": 0.1102968120190546, "grad_norm": 0.4000444070826362, "learning_rate": 9.99690683111036e-06, "loss": 0.4876, "step": 602 }, { "epoch": 0.11048002931476732, "grad_norm": 0.45535979366815893, "learning_rate": 9.99679334184001e-06, "loss": 0.5116, "step": 603 }, { "epoch": 0.11066324661048003, "grad_norm": 0.5208092576973751, "learning_rate": 9.996677808600202e-06, "loss": 0.5079, "step": 604 }, { "epoch": 0.11084646390619274, "grad_norm": 0.47721138790862244, "learning_rate": 9.996560231438193e-06, "loss": 0.5031, "step": 605 }, { "epoch": 0.11102968120190546, "grad_norm": 0.4106250945716929, "learning_rate": 9.996440610402078e-06, "loss": 0.4789, "step": 606 }, { "epoch": 0.11121289849761817, "grad_norm": 0.4981443340009068, "learning_rate": 9.996318945540792e-06, "loss": 0.4749, "step": 607 }, { "epoch": 0.11139611579333089, "grad_norm": 0.4365748987869686, "learning_rate": 9.996195236904098e-06, "loss": 0.503, "step": 608 }, { "epoch": 0.11157933308904361, "grad_norm": 0.42769357659836243, "learning_rate": 9.996069484542605e-06, "loss": 0.4883, "step": 609 }, { "epoch": 0.11176255038475633, "grad_norm": 0.4468425715852285, "learning_rate": 9.995941688507749e-06, "loss": 0.528, "step": 610 }, { "epoch": 0.11194576768046903, "grad_norm": 0.44305501127316244, "learning_rate": 9.995811848851807e-06, "loss": 0.5192, "step": 611 }, { "epoch": 0.11212898497618175, "grad_norm": 0.4430509247761743, "learning_rate": 9.995679965627891e-06, "loss": 0.4879, "step": 612 }, { "epoch": 0.11231220227189447, "grad_norm": 0.4330736232909904, "learning_rate": 9.995546038889948e-06, "loss": 0.5315, "step": 613 }, { "epoch": 0.11249541956760718, "grad_norm": 0.38536887827486094, "learning_rate": 9.995410068692763e-06, "loss": 0.4789, "step": 614 }, { "epoch": 0.1126786368633199, "grad_norm": 0.5420414741691238, "learning_rate": 9.995272055091954e-06, "loss": 0.4863, "step": 615 }, { "epoch": 0.11286185415903262, "grad_norm": 0.4166977498062619, "learning_rate": 9.995131998143976e-06, "loss": 0.4863, "step": 616 }, { "epoch": 0.11304507145474532, "grad_norm": 0.40681536551456327, "learning_rate": 9.99498989790612e-06, "loss": 0.5158, "step": 617 }, { "epoch": 0.11322828875045804, "grad_norm": 0.4164498167643854, "learning_rate": 9.994845754436516e-06, "loss": 0.5287, "step": 618 }, { "epoch": 0.11341150604617076, "grad_norm": 0.430875770675555, "learning_rate": 9.994699567794123e-06, "loss": 0.4949, "step": 619 }, { "epoch": 0.11359472334188347, "grad_norm": 0.47209089186963143, "learning_rate": 9.994551338038742e-06, "loss": 0.5212, "step": 620 }, { "epoch": 0.11377794063759619, "grad_norm": 0.4693460118731517, "learning_rate": 9.994401065231008e-06, "loss": 0.5218, "step": 621 }, { "epoch": 0.11396115793330891, "grad_norm": 0.43645941737097255, "learning_rate": 9.994248749432388e-06, "loss": 0.5355, "step": 622 }, { "epoch": 0.11414437522902161, "grad_norm": 0.45174829183811943, "learning_rate": 9.994094390705189e-06, "loss": 0.4941, "step": 623 }, { "epoch": 0.11432759252473433, "grad_norm": 0.4609885421062077, "learning_rate": 9.993937989112554e-06, "loss": 0.5424, "step": 624 }, { "epoch": 0.11451080982044705, "grad_norm": 0.4711205796072519, "learning_rate": 9.993779544718459e-06, "loss": 0.5054, "step": 625 }, { "epoch": 0.11469402711615977, "grad_norm": 0.4692458224302816, "learning_rate": 9.993619057587714e-06, "loss": 0.5013, "step": 626 }, { "epoch": 0.11487724441187248, "grad_norm": 0.44543459536907126, "learning_rate": 9.99345652778597e-06, "loss": 0.4882, "step": 627 }, { "epoch": 0.1150604617075852, "grad_norm": 0.43461884283611496, "learning_rate": 9.993291955379713e-06, "loss": 0.5352, "step": 628 }, { "epoch": 0.1152436790032979, "grad_norm": 0.41500076390959145, "learning_rate": 9.993125340436258e-06, "loss": 0.4895, "step": 629 }, { "epoch": 0.11542689629901062, "grad_norm": 0.4307377636082482, "learning_rate": 9.992956683023762e-06, "loss": 0.5149, "step": 630 }, { "epoch": 0.11561011359472334, "grad_norm": 0.45085961488644083, "learning_rate": 9.992785983211214e-06, "loss": 0.5061, "step": 631 }, { "epoch": 0.11579333089043606, "grad_norm": 0.45702457348234676, "learning_rate": 9.992613241068444e-06, "loss": 0.4947, "step": 632 }, { "epoch": 0.11597654818614878, "grad_norm": 0.46046219980195896, "learning_rate": 9.992438456666108e-06, "loss": 0.4727, "step": 633 }, { "epoch": 0.1161597654818615, "grad_norm": 0.5109084834664002, "learning_rate": 9.992261630075704e-06, "loss": 0.5015, "step": 634 }, { "epoch": 0.1163429827775742, "grad_norm": 0.4605028589793491, "learning_rate": 9.992082761369567e-06, "loss": 0.518, "step": 635 }, { "epoch": 0.11652620007328691, "grad_norm": 0.4135254020313275, "learning_rate": 9.991901850620861e-06, "loss": 0.4833, "step": 636 }, { "epoch": 0.11670941736899963, "grad_norm": 0.5226822189268372, "learning_rate": 9.99171889790359e-06, "loss": 0.5118, "step": 637 }, { "epoch": 0.11689263466471235, "grad_norm": 0.4061164579366039, "learning_rate": 9.991533903292592e-06, "loss": 0.4953, "step": 638 }, { "epoch": 0.11707585196042507, "grad_norm": 2.3765281557608486, "learning_rate": 9.99134686686354e-06, "loss": 0.4708, "step": 639 }, { "epoch": 0.11725906925613779, "grad_norm": 0.4675538148921928, "learning_rate": 9.991157788692942e-06, "loss": 0.4994, "step": 640 }, { "epoch": 0.11744228655185049, "grad_norm": 0.4467597960418954, "learning_rate": 9.990966668858144e-06, "loss": 0.4982, "step": 641 }, { "epoch": 0.1176255038475632, "grad_norm": 0.46222076234046117, "learning_rate": 9.99077350743732e-06, "loss": 0.5208, "step": 642 }, { "epoch": 0.11780872114327592, "grad_norm": 0.45856899421216063, "learning_rate": 9.990578304509488e-06, "loss": 0.5109, "step": 643 }, { "epoch": 0.11799193843898864, "grad_norm": 0.5006843127431095, "learning_rate": 9.990381060154496e-06, "loss": 0.4941, "step": 644 }, { "epoch": 0.11817515573470136, "grad_norm": 0.4638055815431298, "learning_rate": 9.990181774453028e-06, "loss": 0.509, "step": 645 }, { "epoch": 0.11835837303041408, "grad_norm": 0.4429725714841303, "learning_rate": 9.989980447486601e-06, "loss": 0.5196, "step": 646 }, { "epoch": 0.11854159032612678, "grad_norm": 0.44051330732830757, "learning_rate": 9.989777079337572e-06, "loss": 0.5316, "step": 647 }, { "epoch": 0.1187248076218395, "grad_norm": 0.41555588890083767, "learning_rate": 9.989571670089129e-06, "loss": 0.5126, "step": 648 }, { "epoch": 0.11890802491755222, "grad_norm": 0.46239837428682623, "learning_rate": 9.989364219825295e-06, "loss": 0.506, "step": 649 }, { "epoch": 0.11909124221326493, "grad_norm": 0.416093035128269, "learning_rate": 9.98915472863093e-06, "loss": 0.5194, "step": 650 }, { "epoch": 0.11927445950897765, "grad_norm": 0.43847114193541153, "learning_rate": 9.988943196591727e-06, "loss": 0.4994, "step": 651 }, { "epoch": 0.11945767680469037, "grad_norm": 0.44451525143619486, "learning_rate": 9.988729623794215e-06, "loss": 0.5283, "step": 652 }, { "epoch": 0.11964089410040307, "grad_norm": 0.4041297930697904, "learning_rate": 9.988514010325758e-06, "loss": 0.5049, "step": 653 }, { "epoch": 0.11982411139611579, "grad_norm": 0.4075716530480146, "learning_rate": 9.988296356274551e-06, "loss": 0.4921, "step": 654 }, { "epoch": 0.12000732869182851, "grad_norm": 0.40736195443308915, "learning_rate": 9.988076661729631e-06, "loss": 0.4805, "step": 655 }, { "epoch": 0.12019054598754123, "grad_norm": 0.4545077167148167, "learning_rate": 9.987854926780863e-06, "loss": 0.4789, "step": 656 }, { "epoch": 0.12037376328325394, "grad_norm": 0.5314832542562551, "learning_rate": 9.987631151518948e-06, "loss": 0.5263, "step": 657 }, { "epoch": 0.12055698057896666, "grad_norm": 0.4967524623940482, "learning_rate": 9.987405336035425e-06, "loss": 0.5106, "step": 658 }, { "epoch": 0.12074019787467936, "grad_norm": 0.4306664623952609, "learning_rate": 9.987177480422663e-06, "loss": 0.5376, "step": 659 }, { "epoch": 0.12092341517039208, "grad_norm": 0.4158801663482348, "learning_rate": 9.98694758477387e-06, "loss": 0.4838, "step": 660 }, { "epoch": 0.1211066324661048, "grad_norm": 0.40605820771792417, "learning_rate": 9.986715649183084e-06, "loss": 0.505, "step": 661 }, { "epoch": 0.12128984976181752, "grad_norm": 0.4085721106839356, "learning_rate": 9.986481673745183e-06, "loss": 0.4993, "step": 662 }, { "epoch": 0.12147306705753023, "grad_norm": 0.4785768918832099, "learning_rate": 9.986245658555873e-06, "loss": 0.5387, "step": 663 }, { "epoch": 0.12165628435324295, "grad_norm": 0.4175528791505364, "learning_rate": 9.986007603711698e-06, "loss": 0.4976, "step": 664 }, { "epoch": 0.12183950164895566, "grad_norm": 0.4323331836919794, "learning_rate": 9.985767509310035e-06, "loss": 0.5346, "step": 665 }, { "epoch": 0.12202271894466837, "grad_norm": 0.4868011056584651, "learning_rate": 9.9855253754491e-06, "loss": 0.5309, "step": 666 }, { "epoch": 0.12220593624038109, "grad_norm": 0.46372133085485157, "learning_rate": 9.985281202227936e-06, "loss": 0.5053, "step": 667 }, { "epoch": 0.12238915353609381, "grad_norm": 0.48326269099644514, "learning_rate": 9.985034989746423e-06, "loss": 0.4941, "step": 668 }, { "epoch": 0.12257237083180653, "grad_norm": 0.4557552947362274, "learning_rate": 9.984786738105279e-06, "loss": 0.5121, "step": 669 }, { "epoch": 0.12275558812751924, "grad_norm": 0.39704238565295197, "learning_rate": 9.98453644740605e-06, "loss": 0.4962, "step": 670 }, { "epoch": 0.12293880542323195, "grad_norm": 0.40645243334504044, "learning_rate": 9.98428411775112e-06, "loss": 0.5046, "step": 671 }, { "epoch": 0.12312202271894467, "grad_norm": 0.42629021258457467, "learning_rate": 9.984029749243707e-06, "loss": 0.5084, "step": 672 }, { "epoch": 0.12330524001465738, "grad_norm": 0.45965819318406503, "learning_rate": 9.98377334198786e-06, "loss": 0.4759, "step": 673 }, { "epoch": 0.1234884573103701, "grad_norm": 0.41967629717267835, "learning_rate": 9.983514896088466e-06, "loss": 0.498, "step": 674 }, { "epoch": 0.12367167460608282, "grad_norm": 0.4295939596150219, "learning_rate": 9.983254411651242e-06, "loss": 0.5243, "step": 675 }, { "epoch": 0.12385489190179554, "grad_norm": 0.4808089074794197, "learning_rate": 9.982991888782742e-06, "loss": 0.5, "step": 676 }, { "epoch": 0.12403810919750824, "grad_norm": 0.48728725704379316, "learning_rate": 9.982727327590352e-06, "loss": 0.5459, "step": 677 }, { "epoch": 0.12422132649322096, "grad_norm": 0.42799437772375803, "learning_rate": 9.982460728182292e-06, "loss": 0.5488, "step": 678 }, { "epoch": 0.12440454378893367, "grad_norm": 0.4334349975707624, "learning_rate": 9.982192090667618e-06, "loss": 0.501, "step": 679 }, { "epoch": 0.12458776108464639, "grad_norm": 0.45671378695636383, "learning_rate": 9.981921415156217e-06, "loss": 0.5236, "step": 680 }, { "epoch": 0.12477097838035911, "grad_norm": 0.49531406109254905, "learning_rate": 9.98164870175881e-06, "loss": 0.5224, "step": 681 }, { "epoch": 0.12495419567607183, "grad_norm": 0.3902067301024602, "learning_rate": 9.981373950586952e-06, "loss": 0.4835, "step": 682 }, { "epoch": 0.12513741297178455, "grad_norm": 0.46995820446689096, "learning_rate": 9.981097161753032e-06, "loss": 0.527, "step": 683 }, { "epoch": 0.12532063026749726, "grad_norm": 0.43642001976560174, "learning_rate": 9.980818335370273e-06, "loss": 0.514, "step": 684 }, { "epoch": 0.12550384756320998, "grad_norm": 0.5074521053811396, "learning_rate": 9.980537471552728e-06, "loss": 0.4897, "step": 685 }, { "epoch": 0.12568706485892267, "grad_norm": 0.43625566798430576, "learning_rate": 9.98025457041529e-06, "loss": 0.5174, "step": 686 }, { "epoch": 0.1258702821546354, "grad_norm": 0.4129578447344159, "learning_rate": 9.979969632073678e-06, "loss": 0.5135, "step": 687 }, { "epoch": 0.1260534994503481, "grad_norm": 0.4553566655837448, "learning_rate": 9.97968265664445e-06, "loss": 0.5228, "step": 688 }, { "epoch": 0.12623671674606082, "grad_norm": 0.4682536997157449, "learning_rate": 9.979393644244992e-06, "loss": 0.5039, "step": 689 }, { "epoch": 0.12641993404177354, "grad_norm": 0.42839595687214543, "learning_rate": 9.979102594993533e-06, "loss": 0.5128, "step": 690 }, { "epoch": 0.12660315133748626, "grad_norm": 0.46092632925079513, "learning_rate": 9.978809509009121e-06, "loss": 0.5011, "step": 691 }, { "epoch": 0.12678636863319898, "grad_norm": 0.4486176186332823, "learning_rate": 9.97851438641165e-06, "loss": 0.5169, "step": 692 }, { "epoch": 0.1269695859289117, "grad_norm": 0.47419826995019465, "learning_rate": 9.978217227321837e-06, "loss": 0.5131, "step": 693 }, { "epoch": 0.1271528032246244, "grad_norm": 0.4004091408746546, "learning_rate": 9.97791803186124e-06, "loss": 0.4743, "step": 694 }, { "epoch": 0.12733602052033713, "grad_norm": 0.38849935615066844, "learning_rate": 9.977616800152248e-06, "loss": 0.4944, "step": 695 }, { "epoch": 0.12751923781604985, "grad_norm": 0.44216099670793213, "learning_rate": 9.977313532318078e-06, "loss": 0.5072, "step": 696 }, { "epoch": 0.12770245511176256, "grad_norm": 0.4462646919434802, "learning_rate": 9.977008228482785e-06, "loss": 0.4901, "step": 697 }, { "epoch": 0.12788567240747525, "grad_norm": 0.44769942596462214, "learning_rate": 9.976700888771259e-06, "loss": 0.5263, "step": 698 }, { "epoch": 0.12806888970318797, "grad_norm": 0.5113829977752261, "learning_rate": 9.976391513309212e-06, "loss": 0.5202, "step": 699 }, { "epoch": 0.1282521069989007, "grad_norm": 0.4245344457429005, "learning_rate": 9.976080102223202e-06, "loss": 0.5182, "step": 700 }, { "epoch": 0.1284353242946134, "grad_norm": 0.4305381223047699, "learning_rate": 9.97576665564061e-06, "loss": 0.507, "step": 701 }, { "epoch": 0.12861854159032612, "grad_norm": 0.47072785404283757, "learning_rate": 9.975451173689658e-06, "loss": 0.5166, "step": 702 }, { "epoch": 0.12880175888603884, "grad_norm": 0.6675845713373886, "learning_rate": 9.975133656499392e-06, "loss": 0.47, "step": 703 }, { "epoch": 0.12898497618175156, "grad_norm": 0.4533261672443661, "learning_rate": 9.974814104199694e-06, "loss": 0.5362, "step": 704 }, { "epoch": 0.12916819347746428, "grad_norm": 0.4503943455070683, "learning_rate": 9.97449251692128e-06, "loss": 0.5134, "step": 705 }, { "epoch": 0.129351410773177, "grad_norm": 0.44545859946020006, "learning_rate": 9.974168894795698e-06, "loss": 0.4981, "step": 706 }, { "epoch": 0.1295346280688897, "grad_norm": 0.430864283858029, "learning_rate": 9.973843237955328e-06, "loss": 0.5329, "step": 707 }, { "epoch": 0.12971784536460243, "grad_norm": 0.7865986917626206, "learning_rate": 9.973515546533379e-06, "loss": 0.504, "step": 708 }, { "epoch": 0.12990106266031515, "grad_norm": 0.44976211133712124, "learning_rate": 9.973185820663897e-06, "loss": 0.4737, "step": 709 }, { "epoch": 0.13008427995602784, "grad_norm": 0.4549304394815992, "learning_rate": 9.97285406048176e-06, "loss": 0.5425, "step": 710 }, { "epoch": 0.13026749725174055, "grad_norm": 0.5049990473394447, "learning_rate": 9.972520266122676e-06, "loss": 0.5117, "step": 711 }, { "epoch": 0.13045071454745327, "grad_norm": 0.42657669286902156, "learning_rate": 9.972184437723182e-06, "loss": 0.5057, "step": 712 }, { "epoch": 0.130633931843166, "grad_norm": 0.5299863700854671, "learning_rate": 9.971846575420656e-06, "loss": 0.5262, "step": 713 }, { "epoch": 0.1308171491388787, "grad_norm": 0.45345740582998295, "learning_rate": 9.9715066793533e-06, "loss": 0.4997, "step": 714 }, { "epoch": 0.13100036643459143, "grad_norm": 0.39393086616308026, "learning_rate": 9.971164749660149e-06, "loss": 0.4996, "step": 715 }, { "epoch": 0.13118358373030414, "grad_norm": 0.4577611248522203, "learning_rate": 9.970820786481075e-06, "loss": 0.5018, "step": 716 }, { "epoch": 0.13136680102601686, "grad_norm": 0.43279352533859367, "learning_rate": 9.970474789956775e-06, "loss": 0.4914, "step": 717 }, { "epoch": 0.13155001832172958, "grad_norm": 0.4366448132250816, "learning_rate": 9.970126760228785e-06, "loss": 0.4864, "step": 718 }, { "epoch": 0.1317332356174423, "grad_norm": 0.43998158412052296, "learning_rate": 9.969776697439463e-06, "loss": 0.4834, "step": 719 }, { "epoch": 0.131916452913155, "grad_norm": 0.46940772392002544, "learning_rate": 9.969424601732011e-06, "loss": 0.501, "step": 720 }, { "epoch": 0.1320996702088677, "grad_norm": 0.43117110167391715, "learning_rate": 9.96907047325045e-06, "loss": 0.5562, "step": 721 }, { "epoch": 0.13228288750458042, "grad_norm": 0.4119352967972661, "learning_rate": 9.968714312139642e-06, "loss": 0.4748, "step": 722 }, { "epoch": 0.13246610480029314, "grad_norm": 0.5028443530663289, "learning_rate": 9.968356118545277e-06, "loss": 0.5344, "step": 723 }, { "epoch": 0.13264932209600586, "grad_norm": 0.43089419048593447, "learning_rate": 9.967995892613875e-06, "loss": 0.4704, "step": 724 }, { "epoch": 0.13283253939171857, "grad_norm": 0.5044933769783081, "learning_rate": 9.967633634492788e-06, "loss": 0.4996, "step": 725 }, { "epoch": 0.1330157566874313, "grad_norm": 0.442554097189932, "learning_rate": 9.967269344330201e-06, "loss": 0.5278, "step": 726 }, { "epoch": 0.133198973983144, "grad_norm": 0.431094495901828, "learning_rate": 9.966903022275131e-06, "loss": 0.4943, "step": 727 }, { "epoch": 0.13338219127885673, "grad_norm": 0.6222105275898999, "learning_rate": 9.966534668477421e-06, "loss": 0.5215, "step": 728 }, { "epoch": 0.13356540857456944, "grad_norm": 0.4562449049230116, "learning_rate": 9.96616428308775e-06, "loss": 0.5112, "step": 729 }, { "epoch": 0.13374862587028216, "grad_norm": 0.4249667668518143, "learning_rate": 9.965791866257626e-06, "loss": 0.5083, "step": 730 }, { "epoch": 0.13393184316599488, "grad_norm": 0.42736565388331654, "learning_rate": 9.96541741813939e-06, "loss": 0.5078, "step": 731 }, { "epoch": 0.1341150604617076, "grad_norm": 0.42789645341508464, "learning_rate": 9.96504093888621e-06, "loss": 0.5148, "step": 732 }, { "epoch": 0.1342982777574203, "grad_norm": 0.5018533252539279, "learning_rate": 9.964662428652088e-06, "loss": 0.4994, "step": 733 }, { "epoch": 0.134481495053133, "grad_norm": 0.4561559749816332, "learning_rate": 9.964281887591856e-06, "loss": 0.5036, "step": 734 }, { "epoch": 0.13466471234884572, "grad_norm": 0.43874828747054045, "learning_rate": 9.963899315861176e-06, "loss": 0.4976, "step": 735 }, { "epoch": 0.13484792964455844, "grad_norm": 0.44683273370213666, "learning_rate": 9.963514713616544e-06, "loss": 0.4837, "step": 736 }, { "epoch": 0.13503114694027116, "grad_norm": 0.4575940082218496, "learning_rate": 9.963128081015282e-06, "loss": 0.531, "step": 737 }, { "epoch": 0.13521436423598387, "grad_norm": 0.4190671986873532, "learning_rate": 9.962739418215545e-06, "loss": 0.525, "step": 738 }, { "epoch": 0.1353975815316966, "grad_norm": 0.41284775197321, "learning_rate": 9.962348725376318e-06, "loss": 0.519, "step": 739 }, { "epoch": 0.1355807988274093, "grad_norm": 0.41220938724190687, "learning_rate": 9.961956002657414e-06, "loss": 0.535, "step": 740 }, { "epoch": 0.13576401612312203, "grad_norm": 0.4428272636340964, "learning_rate": 9.961561250219482e-06, "loss": 0.5287, "step": 741 }, { "epoch": 0.13594723341883475, "grad_norm": 0.42565025293852765, "learning_rate": 9.961164468223996e-06, "loss": 0.5329, "step": 742 }, { "epoch": 0.13613045071454746, "grad_norm": 0.5963216630193675, "learning_rate": 9.960765656833263e-06, "loss": 0.5036, "step": 743 }, { "epoch": 0.13631366801026018, "grad_norm": 0.42494573499796934, "learning_rate": 9.96036481621042e-06, "loss": 0.4936, "step": 744 }, { "epoch": 0.13649688530597287, "grad_norm": 0.39224716002216886, "learning_rate": 9.959961946519431e-06, "loss": 0.4704, "step": 745 }, { "epoch": 0.1366801026016856, "grad_norm": 0.42574265198042266, "learning_rate": 9.959557047925095e-06, "loss": 0.4964, "step": 746 }, { "epoch": 0.1368633198973983, "grad_norm": 0.457270415221301, "learning_rate": 9.959150120593035e-06, "loss": 0.5322, "step": 747 }, { "epoch": 0.13704653719311102, "grad_norm": 0.4245718410150897, "learning_rate": 9.95874116468971e-06, "loss": 0.4952, "step": 748 }, { "epoch": 0.13722975448882374, "grad_norm": 0.47589556246220993, "learning_rate": 9.958330180382405e-06, "loss": 0.5124, "step": 749 }, { "epoch": 0.13741297178453646, "grad_norm": 0.45440270321018955, "learning_rate": 9.957917167839238e-06, "loss": 0.5109, "step": 750 }, { "epoch": 0.13759618908024918, "grad_norm": 0.5232685566183007, "learning_rate": 9.95750212722915e-06, "loss": 0.5174, "step": 751 }, { "epoch": 0.1377794063759619, "grad_norm": 0.538012343254673, "learning_rate": 9.957085058721918e-06, "loss": 0.5105, "step": 752 }, { "epoch": 0.1379626236716746, "grad_norm": 0.38959097348139504, "learning_rate": 9.956665962488148e-06, "loss": 0.4533, "step": 753 }, { "epoch": 0.13814584096738733, "grad_norm": 0.4404610898092818, "learning_rate": 9.956244838699271e-06, "loss": 0.4978, "step": 754 }, { "epoch": 0.13832905826310005, "grad_norm": 0.47384922621236214, "learning_rate": 9.955821687527554e-06, "loss": 0.5309, "step": 755 }, { "epoch": 0.13851227555881276, "grad_norm": 0.45266354792605684, "learning_rate": 9.955396509146084e-06, "loss": 0.5151, "step": 756 }, { "epoch": 0.13869549285452545, "grad_norm": 0.4098036061670406, "learning_rate": 9.95496930372879e-06, "loss": 0.5, "step": 757 }, { "epoch": 0.13887871015023817, "grad_norm": 0.4520393902857722, "learning_rate": 9.954540071450418e-06, "loss": 0.4766, "step": 758 }, { "epoch": 0.1390619274459509, "grad_norm": 0.41592431105518124, "learning_rate": 9.95410881248655e-06, "loss": 0.5279, "step": 759 }, { "epoch": 0.1392451447416636, "grad_norm": 0.9055926932881602, "learning_rate": 9.953675527013594e-06, "loss": 0.4544, "step": 760 }, { "epoch": 0.13942836203737632, "grad_norm": 0.4230083797544526, "learning_rate": 9.953240215208787e-06, "loss": 0.5174, "step": 761 }, { "epoch": 0.13961157933308904, "grad_norm": 0.46167105853029766, "learning_rate": 9.9528028772502e-06, "loss": 0.5087, "step": 762 }, { "epoch": 0.13979479662880176, "grad_norm": 0.4614988799168283, "learning_rate": 9.952363513316727e-06, "loss": 0.5035, "step": 763 }, { "epoch": 0.13997801392451448, "grad_norm": 0.4599378559524754, "learning_rate": 9.951922123588091e-06, "loss": 0.4982, "step": 764 }, { "epoch": 0.1401612312202272, "grad_norm": 0.4169187563233499, "learning_rate": 9.951478708244847e-06, "loss": 0.5205, "step": 765 }, { "epoch": 0.1403444485159399, "grad_norm": 1.4346541301623688, "learning_rate": 9.951033267468375e-06, "loss": 0.4704, "step": 766 }, { "epoch": 0.14052766581165263, "grad_norm": 0.4477016515011828, "learning_rate": 9.950585801440889e-06, "loss": 0.489, "step": 767 }, { "epoch": 0.14071088310736535, "grad_norm": 0.4178701034776549, "learning_rate": 9.950136310345425e-06, "loss": 0.5147, "step": 768 }, { "epoch": 0.14089410040307804, "grad_norm": 0.4893104037393506, "learning_rate": 9.949684794365848e-06, "loss": 0.4913, "step": 769 }, { "epoch": 0.14107731769879076, "grad_norm": 0.4018838415277579, "learning_rate": 9.949231253686857e-06, "loss": 0.464, "step": 770 }, { "epoch": 0.14126053499450347, "grad_norm": 0.38608379924790914, "learning_rate": 9.948775688493974e-06, "loss": 0.5173, "step": 771 }, { "epoch": 0.1414437522902162, "grad_norm": 0.4162388206396457, "learning_rate": 9.948318098973552e-06, "loss": 0.4917, "step": 772 }, { "epoch": 0.1416269695859289, "grad_norm": 0.3781950328417496, "learning_rate": 9.947858485312772e-06, "loss": 0.4446, "step": 773 }, { "epoch": 0.14181018688164163, "grad_norm": 0.45959051057998834, "learning_rate": 9.947396847699638e-06, "loss": 0.4918, "step": 774 }, { "epoch": 0.14199340417735434, "grad_norm": 0.480317102541819, "learning_rate": 9.946933186322988e-06, "loss": 0.5369, "step": 775 }, { "epoch": 0.14217662147306706, "grad_norm": 0.5008985493720574, "learning_rate": 9.946467501372485e-06, "loss": 0.4992, "step": 776 }, { "epoch": 0.14235983876877978, "grad_norm": 0.4117591563183844, "learning_rate": 9.94599979303862e-06, "loss": 0.4818, "step": 777 }, { "epoch": 0.1425430560644925, "grad_norm": 0.531108956987235, "learning_rate": 9.945530061512714e-06, "loss": 0.517, "step": 778 }, { "epoch": 0.1427262733602052, "grad_norm": 0.43817523582155865, "learning_rate": 9.945058306986911e-06, "loss": 0.5473, "step": 779 }, { "epoch": 0.14290949065591793, "grad_norm": 0.4747080227792923, "learning_rate": 9.944584529654187e-06, "loss": 0.4819, "step": 780 }, { "epoch": 0.14309270795163062, "grad_norm": 0.4249215515551791, "learning_rate": 9.944108729708342e-06, "loss": 0.4979, "step": 781 }, { "epoch": 0.14327592524734334, "grad_norm": 0.43558257752578505, "learning_rate": 9.943630907344008e-06, "loss": 0.484, "step": 782 }, { "epoch": 0.14345914254305606, "grad_norm": 0.4750219883418816, "learning_rate": 9.943151062756638e-06, "loss": 0.4825, "step": 783 }, { "epoch": 0.14364235983876877, "grad_norm": 0.4332476417990451, "learning_rate": 9.942669196142516e-06, "loss": 0.4929, "step": 784 }, { "epoch": 0.1438255771344815, "grad_norm": 0.48307720479460303, "learning_rate": 9.942185307698754e-06, "loss": 0.5099, "step": 785 }, { "epoch": 0.1440087944301942, "grad_norm": 0.43004630110605213, "learning_rate": 9.941699397623289e-06, "loss": 0.4435, "step": 786 }, { "epoch": 0.14419201172590693, "grad_norm": 0.5526090978976514, "learning_rate": 9.941211466114883e-06, "loss": 0.4831, "step": 787 }, { "epoch": 0.14437522902161964, "grad_norm": 0.4299548315161719, "learning_rate": 9.94072151337313e-06, "loss": 0.5055, "step": 788 }, { "epoch": 0.14455844631733236, "grad_norm": 0.46488795405005734, "learning_rate": 9.940229539598449e-06, "loss": 0.5247, "step": 789 }, { "epoch": 0.14474166361304508, "grad_norm": 0.45711196453956754, "learning_rate": 9.939735544992084e-06, "loss": 0.5155, "step": 790 }, { "epoch": 0.1449248809087578, "grad_norm": 0.39377908902960373, "learning_rate": 9.939239529756106e-06, "loss": 0.4861, "step": 791 }, { "epoch": 0.14510809820447051, "grad_norm": 0.3780973854580167, "learning_rate": 9.938741494093413e-06, "loss": 0.4743, "step": 792 }, { "epoch": 0.1452913155001832, "grad_norm": 0.4442667746426281, "learning_rate": 9.93824143820773e-06, "loss": 0.5124, "step": 793 }, { "epoch": 0.14547453279589592, "grad_norm": 0.47139079982459514, "learning_rate": 9.93773936230361e-06, "loss": 0.5095, "step": 794 }, { "epoch": 0.14565775009160864, "grad_norm": 0.47478113201389704, "learning_rate": 9.937235266586425e-06, "loss": 0.5242, "step": 795 }, { "epoch": 0.14584096738732136, "grad_norm": 0.4253566981856297, "learning_rate": 9.936729151262383e-06, "loss": 0.4816, "step": 796 }, { "epoch": 0.14602418468303408, "grad_norm": 0.3995444614750483, "learning_rate": 9.936221016538514e-06, "loss": 0.5114, "step": 797 }, { "epoch": 0.1462074019787468, "grad_norm": 0.3958644449360531, "learning_rate": 9.935710862622671e-06, "loss": 0.5044, "step": 798 }, { "epoch": 0.1463906192744595, "grad_norm": 0.39672728856696343, "learning_rate": 9.935198689723537e-06, "loss": 0.5079, "step": 799 }, { "epoch": 0.14657383657017223, "grad_norm": 0.4401216989939036, "learning_rate": 9.934684498050619e-06, "loss": 0.5169, "step": 800 }, { "epoch": 0.14675705386588495, "grad_norm": 0.42531985045283716, "learning_rate": 9.93416828781425e-06, "loss": 0.5039, "step": 801 }, { "epoch": 0.14694027116159766, "grad_norm": 0.43641771981673, "learning_rate": 9.93365005922559e-06, "loss": 0.4705, "step": 802 }, { "epoch": 0.14712348845731038, "grad_norm": 0.4154214727248315, "learning_rate": 9.933129812496623e-06, "loss": 0.4801, "step": 803 }, { "epoch": 0.1473067057530231, "grad_norm": 0.4668765888893056, "learning_rate": 9.932607547840161e-06, "loss": 0.5262, "step": 804 }, { "epoch": 0.1474899230487358, "grad_norm": 0.45163409733644266, "learning_rate": 9.932083265469836e-06, "loss": 0.5009, "step": 805 }, { "epoch": 0.1476731403444485, "grad_norm": 0.44602336917824525, "learning_rate": 9.93155696560011e-06, "loss": 0.5071, "step": 806 }, { "epoch": 0.14785635764016122, "grad_norm": 0.40571563214546336, "learning_rate": 9.931028648446273e-06, "loss": 0.4867, "step": 807 }, { "epoch": 0.14803957493587394, "grad_norm": 0.412866633306789, "learning_rate": 9.930498314224433e-06, "loss": 0.5204, "step": 808 }, { "epoch": 0.14822279223158666, "grad_norm": 0.4491417286505596, "learning_rate": 9.929965963151526e-06, "loss": 0.4959, "step": 809 }, { "epoch": 0.14840600952729938, "grad_norm": 0.483698524685917, "learning_rate": 9.929431595445315e-06, "loss": 0.4945, "step": 810 }, { "epoch": 0.1485892268230121, "grad_norm": 0.45809625895317874, "learning_rate": 9.928895211324387e-06, "loss": 0.4795, "step": 811 }, { "epoch": 0.1487724441187248, "grad_norm": 0.4671855759216309, "learning_rate": 9.928356811008153e-06, "loss": 0.502, "step": 812 }, { "epoch": 0.14895566141443753, "grad_norm": 0.49991252806247843, "learning_rate": 9.927816394716847e-06, "loss": 0.4989, "step": 813 }, { "epoch": 0.14913887871015025, "grad_norm": 0.4930888660574426, "learning_rate": 9.92727396267153e-06, "loss": 0.5256, "step": 814 }, { "epoch": 0.14932209600586296, "grad_norm": 0.4301245288190715, "learning_rate": 9.926729515094092e-06, "loss": 0.5062, "step": 815 }, { "epoch": 0.14950531330157568, "grad_norm": 0.4254345495226345, "learning_rate": 9.926183052207235e-06, "loss": 0.4885, "step": 816 }, { "epoch": 0.14968853059728837, "grad_norm": 0.40854658960332646, "learning_rate": 9.925634574234499e-06, "loss": 0.4909, "step": 817 }, { "epoch": 0.1498717478930011, "grad_norm": 0.4307966660913799, "learning_rate": 9.925084081400241e-06, "loss": 0.4974, "step": 818 }, { "epoch": 0.1500549651887138, "grad_norm": 0.44710983697734263, "learning_rate": 9.924531573929641e-06, "loss": 0.5107, "step": 819 }, { "epoch": 0.15023818248442652, "grad_norm": 0.47247983608563787, "learning_rate": 9.923977052048708e-06, "loss": 0.5185, "step": 820 }, { "epoch": 0.15042139978013924, "grad_norm": 0.43408666873611645, "learning_rate": 9.923420515984272e-06, "loss": 0.5256, "step": 821 }, { "epoch": 0.15060461707585196, "grad_norm": 0.3618361050633503, "learning_rate": 9.922861965963987e-06, "loss": 0.4835, "step": 822 }, { "epoch": 0.15078783437156468, "grad_norm": 0.478572088704784, "learning_rate": 9.922301402216334e-06, "loss": 0.4971, "step": 823 }, { "epoch": 0.1509710516672774, "grad_norm": 0.4956723559981853, "learning_rate": 9.921738824970611e-06, "loss": 0.4664, "step": 824 }, { "epoch": 0.1511542689629901, "grad_norm": 0.43104833444065105, "learning_rate": 9.921174234456947e-06, "loss": 0.5056, "step": 825 }, { "epoch": 0.15133748625870283, "grad_norm": 0.38636058698792763, "learning_rate": 9.920607630906289e-06, "loss": 0.509, "step": 826 }, { "epoch": 0.15152070355441555, "grad_norm": 0.4638108326048624, "learning_rate": 9.920039014550413e-06, "loss": 0.4928, "step": 827 }, { "epoch": 0.15170392085012827, "grad_norm": 0.38962500919500936, "learning_rate": 9.919468385621912e-06, "loss": 0.5025, "step": 828 }, { "epoch": 0.15188713814584096, "grad_norm": 0.423466480397551, "learning_rate": 9.918895744354204e-06, "loss": 0.4903, "step": 829 }, { "epoch": 0.15207035544155367, "grad_norm": 0.5945183810824295, "learning_rate": 9.918321090981537e-06, "loss": 0.4961, "step": 830 }, { "epoch": 0.1522535727372664, "grad_norm": 0.3857346095853992, "learning_rate": 9.917744425738971e-06, "loss": 0.4903, "step": 831 }, { "epoch": 0.1524367900329791, "grad_norm": 0.4368019305898112, "learning_rate": 9.917165748862398e-06, "loss": 0.4985, "step": 832 }, { "epoch": 0.15262000732869183, "grad_norm": 0.4622792781214721, "learning_rate": 9.916585060588526e-06, "loss": 0.5298, "step": 833 }, { "epoch": 0.15280322462440454, "grad_norm": 0.43464073300377387, "learning_rate": 9.916002361154894e-06, "loss": 0.4956, "step": 834 }, { "epoch": 0.15298644192011726, "grad_norm": 0.37914154472744427, "learning_rate": 9.915417650799855e-06, "loss": 0.4914, "step": 835 }, { "epoch": 0.15316965921582998, "grad_norm": 0.41110768627195465, "learning_rate": 9.914830929762588e-06, "loss": 0.4933, "step": 836 }, { "epoch": 0.1533528765115427, "grad_norm": 0.3922989216974703, "learning_rate": 9.914242198283099e-06, "loss": 0.5, "step": 837 }, { "epoch": 0.1535360938072554, "grad_norm": 0.4310098182316404, "learning_rate": 9.91365145660221e-06, "loss": 0.4817, "step": 838 }, { "epoch": 0.15371931110296813, "grad_norm": 0.40690680117071937, "learning_rate": 9.913058704961565e-06, "loss": 0.4778, "step": 839 }, { "epoch": 0.15390252839868085, "grad_norm": 0.4584415585967629, "learning_rate": 9.912463943603635e-06, "loss": 0.4585, "step": 840 }, { "epoch": 0.15408574569439354, "grad_norm": 0.48734349658323495, "learning_rate": 9.911867172771711e-06, "loss": 0.5345, "step": 841 }, { "epoch": 0.15426896299010626, "grad_norm": 0.442686833828428, "learning_rate": 9.911268392709908e-06, "loss": 0.5134, "step": 842 }, { "epoch": 0.15445218028581897, "grad_norm": 0.44506293028939, "learning_rate": 9.910667603663156e-06, "loss": 0.5158, "step": 843 }, { "epoch": 0.1546353975815317, "grad_norm": 0.4311992637968541, "learning_rate": 9.910064805877214e-06, "loss": 0.4837, "step": 844 }, { "epoch": 0.1548186148772444, "grad_norm": 0.4493099076015729, "learning_rate": 9.90945999959866e-06, "loss": 0.5108, "step": 845 }, { "epoch": 0.15500183217295713, "grad_norm": 0.45478365363321205, "learning_rate": 9.908853185074896e-06, "loss": 0.4998, "step": 846 }, { "epoch": 0.15518504946866984, "grad_norm": 0.46301414888311854, "learning_rate": 9.90824436255414e-06, "loss": 0.5225, "step": 847 }, { "epoch": 0.15536826676438256, "grad_norm": 0.40118617155548936, "learning_rate": 9.907633532285435e-06, "loss": 0.4927, "step": 848 }, { "epoch": 0.15555148406009528, "grad_norm": 0.49481295874709574, "learning_rate": 9.907020694518646e-06, "loss": 0.4933, "step": 849 }, { "epoch": 0.155734701355808, "grad_norm": 0.4102626498510104, "learning_rate": 9.90640584950446e-06, "loss": 0.5111, "step": 850 }, { "epoch": 0.15591791865152071, "grad_norm": 0.46514690031713113, "learning_rate": 9.905788997494377e-06, "loss": 0.4728, "step": 851 }, { "epoch": 0.15610113594723343, "grad_norm": 0.4054441718781369, "learning_rate": 9.905170138740732e-06, "loss": 0.4828, "step": 852 }, { "epoch": 0.15628435324294612, "grad_norm": 0.4400252577977126, "learning_rate": 9.904549273496666e-06, "loss": 0.501, "step": 853 }, { "epoch": 0.15646757053865884, "grad_norm": 0.42167618939817325, "learning_rate": 9.903926402016153e-06, "loss": 0.4922, "step": 854 }, { "epoch": 0.15665078783437156, "grad_norm": 0.4093707700968775, "learning_rate": 9.90330152455398e-06, "loss": 0.5014, "step": 855 }, { "epoch": 0.15683400513008428, "grad_norm": 0.4523522809487888, "learning_rate": 9.902674641365756e-06, "loss": 0.4829, "step": 856 }, { "epoch": 0.157017222425797, "grad_norm": 0.49143897815029397, "learning_rate": 9.902045752707916e-06, "loss": 0.5074, "step": 857 }, { "epoch": 0.1572004397215097, "grad_norm": 0.43621740114283203, "learning_rate": 9.901414858837706e-06, "loss": 0.5182, "step": 858 }, { "epoch": 0.15738365701722243, "grad_norm": 0.40732498942379025, "learning_rate": 9.9007819600132e-06, "loss": 0.4674, "step": 859 }, { "epoch": 0.15756687431293515, "grad_norm": 0.43675476681733244, "learning_rate": 9.900147056493285e-06, "loss": 0.5374, "step": 860 }, { "epoch": 0.15775009160864786, "grad_norm": 0.43060074322101793, "learning_rate": 9.89951014853768e-06, "loss": 0.4635, "step": 861 }, { "epoch": 0.15793330890436058, "grad_norm": 0.41179202097945616, "learning_rate": 9.898871236406907e-06, "loss": 0.4951, "step": 862 }, { "epoch": 0.1581165262000733, "grad_norm": 0.371370018002044, "learning_rate": 9.898230320362323e-06, "loss": 0.479, "step": 863 }, { "epoch": 0.158299743495786, "grad_norm": 0.4300134296591519, "learning_rate": 9.897587400666097e-06, "loss": 0.4868, "step": 864 }, { "epoch": 0.1584829607914987, "grad_norm": 0.43589209027697184, "learning_rate": 9.896942477581221e-06, "loss": 0.5124, "step": 865 }, { "epoch": 0.15866617808721142, "grad_norm": 0.575247632832188, "learning_rate": 9.896295551371503e-06, "loss": 0.5053, "step": 866 }, { "epoch": 0.15884939538292414, "grad_norm": 0.4273872012355264, "learning_rate": 9.89564662230157e-06, "loss": 0.5138, "step": 867 }, { "epoch": 0.15903261267863686, "grad_norm": 0.4543163505662005, "learning_rate": 9.894995690636874e-06, "loss": 0.5229, "step": 868 }, { "epoch": 0.15921582997434958, "grad_norm": 0.4491410597219586, "learning_rate": 9.894342756643684e-06, "loss": 0.5045, "step": 869 }, { "epoch": 0.1593990472700623, "grad_norm": 0.47087708750396473, "learning_rate": 9.893687820589081e-06, "loss": 0.5087, "step": 870 }, { "epoch": 0.159582264565775, "grad_norm": 0.4240236693967998, "learning_rate": 9.893030882740977e-06, "loss": 0.5052, "step": 871 }, { "epoch": 0.15976548186148773, "grad_norm": 0.4287525963131784, "learning_rate": 9.892371943368092e-06, "loss": 0.527, "step": 872 }, { "epoch": 0.15994869915720045, "grad_norm": 0.4131563755872441, "learning_rate": 9.891711002739971e-06, "loss": 0.5074, "step": 873 }, { "epoch": 0.16013191645291316, "grad_norm": 0.39231143806918656, "learning_rate": 9.891048061126975e-06, "loss": 0.5018, "step": 874 }, { "epoch": 0.16031513374862588, "grad_norm": 0.4391493046066672, "learning_rate": 9.890383118800287e-06, "loss": 0.518, "step": 875 }, { "epoch": 0.16049835104433857, "grad_norm": 0.468016398258625, "learning_rate": 9.889716176031903e-06, "loss": 0.5191, "step": 876 }, { "epoch": 0.1606815683400513, "grad_norm": 0.4619425954928644, "learning_rate": 9.88904723309464e-06, "loss": 0.5114, "step": 877 }, { "epoch": 0.160864785635764, "grad_norm": 0.4581492770476195, "learning_rate": 9.888376290262134e-06, "loss": 0.5349, "step": 878 }, { "epoch": 0.16104800293147672, "grad_norm": 0.390022001202663, "learning_rate": 9.887703347808838e-06, "loss": 0.507, "step": 879 }, { "epoch": 0.16123122022718944, "grad_norm": 0.38271876665722593, "learning_rate": 9.887028406010026e-06, "loss": 0.4987, "step": 880 }, { "epoch": 0.16141443752290216, "grad_norm": 0.39991523006604696, "learning_rate": 9.886351465141785e-06, "loss": 0.4807, "step": 881 }, { "epoch": 0.16159765481861488, "grad_norm": 0.3867195675970219, "learning_rate": 9.88567252548102e-06, "loss": 0.515, "step": 882 }, { "epoch": 0.1617808721143276, "grad_norm": 0.47566001886735476, "learning_rate": 9.884991587305459e-06, "loss": 0.4461, "step": 883 }, { "epoch": 0.1619640894100403, "grad_norm": 0.4453653597243864, "learning_rate": 9.884308650893642e-06, "loss": 0.5493, "step": 884 }, { "epoch": 0.16214730670575303, "grad_norm": 0.38787040868110323, "learning_rate": 9.883623716524929e-06, "loss": 0.4998, "step": 885 }, { "epoch": 0.16233052400146575, "grad_norm": 0.4570258592705414, "learning_rate": 9.882936784479498e-06, "loss": 0.5226, "step": 886 }, { "epoch": 0.16251374129717847, "grad_norm": 0.4438200453897428, "learning_rate": 9.882247855038339e-06, "loss": 0.5073, "step": 887 }, { "epoch": 0.16269695859289116, "grad_norm": 0.3887912773501902, "learning_rate": 9.881556928483266e-06, "loss": 0.5213, "step": 888 }, { "epoch": 0.16288017588860387, "grad_norm": 0.44048355093202696, "learning_rate": 9.880864005096906e-06, "loss": 0.511, "step": 889 }, { "epoch": 0.1630633931843166, "grad_norm": 0.44235079469227967, "learning_rate": 9.880169085162703e-06, "loss": 0.5223, "step": 890 }, { "epoch": 0.1632466104800293, "grad_norm": 0.3877832289424, "learning_rate": 9.87947216896492e-06, "loss": 0.4537, "step": 891 }, { "epoch": 0.16342982777574203, "grad_norm": 0.44869938145307137, "learning_rate": 9.878773256788635e-06, "loss": 0.5013, "step": 892 }, { "epoch": 0.16361304507145474, "grad_norm": 0.46375171983760244, "learning_rate": 9.878072348919738e-06, "loss": 0.489, "step": 893 }, { "epoch": 0.16379626236716746, "grad_norm": 0.49608852880970994, "learning_rate": 9.877369445644945e-06, "loss": 0.5177, "step": 894 }, { "epoch": 0.16397947966288018, "grad_norm": 0.4082183549684993, "learning_rate": 9.876664547251781e-06, "loss": 0.5091, "step": 895 }, { "epoch": 0.1641626969585929, "grad_norm": 0.4347968628628821, "learning_rate": 9.875957654028588e-06, "loss": 0.5242, "step": 896 }, { "epoch": 0.16434591425430561, "grad_norm": 0.45545255399303036, "learning_rate": 9.875248766264527e-06, "loss": 0.5406, "step": 897 }, { "epoch": 0.16452913155001833, "grad_norm": 0.41358644063630506, "learning_rate": 9.874537884249574e-06, "loss": 0.511, "step": 898 }, { "epoch": 0.16471234884573105, "grad_norm": 0.4576667619206321, "learning_rate": 9.873825008274514e-06, "loss": 0.5073, "step": 899 }, { "epoch": 0.16489556614144374, "grad_norm": 0.47721781865779744, "learning_rate": 9.87311013863096e-06, "loss": 0.5044, "step": 900 }, { "epoch": 0.16507878343715646, "grad_norm": 0.4622288619794578, "learning_rate": 9.872393275611329e-06, "loss": 0.5096, "step": 901 }, { "epoch": 0.16526200073286917, "grad_norm": 0.41177688817899083, "learning_rate": 9.871674419508864e-06, "loss": 0.523, "step": 902 }, { "epoch": 0.1654452180285819, "grad_norm": 0.371054810469729, "learning_rate": 9.87095357061761e-06, "loss": 0.4897, "step": 903 }, { "epoch": 0.1656284353242946, "grad_norm": 0.41574608839188226, "learning_rate": 9.87023072923244e-06, "loss": 0.4823, "step": 904 }, { "epoch": 0.16581165262000733, "grad_norm": 0.4218915251343208, "learning_rate": 9.869505895649036e-06, "loss": 0.5079, "step": 905 }, { "epoch": 0.16599486991572004, "grad_norm": 0.5549831381708766, "learning_rate": 9.868779070163895e-06, "loss": 0.484, "step": 906 }, { "epoch": 0.16617808721143276, "grad_norm": 0.3891473564827362, "learning_rate": 9.868050253074328e-06, "loss": 0.4903, "step": 907 }, { "epoch": 0.16636130450714548, "grad_norm": 0.4268945705632106, "learning_rate": 9.867319444678465e-06, "loss": 0.5024, "step": 908 }, { "epoch": 0.1665445218028582, "grad_norm": 0.40332206359603545, "learning_rate": 9.866586645275247e-06, "loss": 0.4477, "step": 909 }, { "epoch": 0.16672773909857092, "grad_norm": 0.3720747589583317, "learning_rate": 9.86585185516443e-06, "loss": 0.469, "step": 910 }, { "epoch": 0.16691095639428363, "grad_norm": 0.451113552639029, "learning_rate": 9.865115074646583e-06, "loss": 0.5221, "step": 911 }, { "epoch": 0.16709417368999632, "grad_norm": 0.4454998332020719, "learning_rate": 9.864376304023092e-06, "loss": 0.495, "step": 912 }, { "epoch": 0.16727739098570904, "grad_norm": 0.5363910151131063, "learning_rate": 9.863635543596156e-06, "loss": 0.5233, "step": 913 }, { "epoch": 0.16746060828142176, "grad_norm": 0.4392403856880424, "learning_rate": 9.862892793668787e-06, "loss": 0.4998, "step": 914 }, { "epoch": 0.16764382557713448, "grad_norm": 0.42093143672832734, "learning_rate": 9.862148054544812e-06, "loss": 0.4977, "step": 915 }, { "epoch": 0.1678270428728472, "grad_norm": 0.39275787265485074, "learning_rate": 9.86140132652887e-06, "loss": 0.4694, "step": 916 }, { "epoch": 0.1680102601685599, "grad_norm": 0.4159908861548446, "learning_rate": 9.860652609926417e-06, "loss": 0.4886, "step": 917 }, { "epoch": 0.16819347746427263, "grad_norm": 0.38579180273261776, "learning_rate": 9.859901905043718e-06, "loss": 0.48, "step": 918 }, { "epoch": 0.16837669475998535, "grad_norm": 0.4273621763370794, "learning_rate": 9.859149212187855e-06, "loss": 0.512, "step": 919 }, { "epoch": 0.16855991205569806, "grad_norm": 0.45913559298162226, "learning_rate": 9.85839453166672e-06, "loss": 0.4877, "step": 920 }, { "epoch": 0.16874312935141078, "grad_norm": 0.3847363797846655, "learning_rate": 9.85763786378902e-06, "loss": 0.4586, "step": 921 }, { "epoch": 0.1689263466471235, "grad_norm": 0.4872256895625686, "learning_rate": 9.856879208864277e-06, "loss": 0.5035, "step": 922 }, { "epoch": 0.16910956394283622, "grad_norm": 0.47756329412725707, "learning_rate": 9.85611856720282e-06, "loss": 0.5031, "step": 923 }, { "epoch": 0.1692927812385489, "grad_norm": 0.474431940864528, "learning_rate": 9.8553559391158e-06, "loss": 0.512, "step": 924 }, { "epoch": 0.16947599853426162, "grad_norm": 0.41972238109665655, "learning_rate": 9.85459132491517e-06, "loss": 0.5386, "step": 925 }, { "epoch": 0.16965921582997434, "grad_norm": 0.4369771091332424, "learning_rate": 9.8538247249137e-06, "loss": 0.4909, "step": 926 }, { "epoch": 0.16984243312568706, "grad_norm": 0.4523772832392917, "learning_rate": 9.853056139424974e-06, "loss": 0.4951, "step": 927 }, { "epoch": 0.17002565042139978, "grad_norm": 0.40235086982371876, "learning_rate": 9.852285568763387e-06, "loss": 0.4828, "step": 928 }, { "epoch": 0.1702088677171125, "grad_norm": 0.4150844402567988, "learning_rate": 9.851513013244144e-06, "loss": 0.4933, "step": 929 }, { "epoch": 0.1703920850128252, "grad_norm": 0.3941602147072389, "learning_rate": 9.850738473183266e-06, "loss": 0.4855, "step": 930 }, { "epoch": 0.17057530230853793, "grad_norm": 0.47118394280764797, "learning_rate": 9.849961948897582e-06, "loss": 0.5024, "step": 931 }, { "epoch": 0.17075851960425065, "grad_norm": 0.3912878663413955, "learning_rate": 9.849183440704735e-06, "loss": 0.5041, "step": 932 }, { "epoch": 0.17094173689996336, "grad_norm": 0.4001961484144372, "learning_rate": 9.848402948923177e-06, "loss": 0.4753, "step": 933 }, { "epoch": 0.17112495419567608, "grad_norm": 0.4441137892348275, "learning_rate": 9.847620473872172e-06, "loss": 0.5102, "step": 934 }, { "epoch": 0.1713081714913888, "grad_norm": 0.40101758764349804, "learning_rate": 9.846836015871802e-06, "loss": 0.4824, "step": 935 }, { "epoch": 0.1714913887871015, "grad_norm": 0.41075944565503664, "learning_rate": 9.846049575242949e-06, "loss": 0.5235, "step": 936 }, { "epoch": 0.1716746060828142, "grad_norm": 0.3780837940291423, "learning_rate": 9.845261152307312e-06, "loss": 0.5211, "step": 937 }, { "epoch": 0.17185782337852692, "grad_norm": 0.45070991069476113, "learning_rate": 9.844470747387403e-06, "loss": 0.4926, "step": 938 }, { "epoch": 0.17204104067423964, "grad_norm": 0.40836878453605646, "learning_rate": 9.843678360806542e-06, "loss": 0.5173, "step": 939 }, { "epoch": 0.17222425796995236, "grad_norm": 0.36676941914948374, "learning_rate": 9.842883992888855e-06, "loss": 0.5101, "step": 940 }, { "epoch": 0.17240747526566508, "grad_norm": 0.42894488109200807, "learning_rate": 9.842087643959288e-06, "loss": 0.5398, "step": 941 }, { "epoch": 0.1725906925613778, "grad_norm": 0.4004768449690441, "learning_rate": 9.841289314343591e-06, "loss": 0.4967, "step": 942 }, { "epoch": 0.1727739098570905, "grad_norm": 0.3781076276341992, "learning_rate": 9.840489004368325e-06, "loss": 0.4731, "step": 943 }, { "epoch": 0.17295712715280323, "grad_norm": 0.45307945315267256, "learning_rate": 9.839686714360864e-06, "loss": 0.4737, "step": 944 }, { "epoch": 0.17314034444851595, "grad_norm": 0.4079247794136399, "learning_rate": 9.838882444649387e-06, "loss": 0.4927, "step": 945 }, { "epoch": 0.17332356174422867, "grad_norm": 0.47171018664243136, "learning_rate": 9.838076195562886e-06, "loss": 0.5009, "step": 946 }, { "epoch": 0.17350677903994138, "grad_norm": 0.44572108382734166, "learning_rate": 9.837267967431164e-06, "loss": 0.4858, "step": 947 }, { "epoch": 0.17368999633565407, "grad_norm": 0.40649730883742624, "learning_rate": 9.83645776058483e-06, "loss": 0.4943, "step": 948 }, { "epoch": 0.1738732136313668, "grad_norm": 0.43002759908477306, "learning_rate": 9.835645575355304e-06, "loss": 0.4872, "step": 949 }, { "epoch": 0.1740564309270795, "grad_norm": 0.39064612647120167, "learning_rate": 9.834831412074816e-06, "loss": 0.4409, "step": 950 }, { "epoch": 0.17423964822279223, "grad_norm": 0.3761573603114394, "learning_rate": 9.834015271076405e-06, "loss": 0.459, "step": 951 }, { "epoch": 0.17442286551850494, "grad_norm": 0.42796519720566795, "learning_rate": 9.83319715269392e-06, "loss": 0.4906, "step": 952 }, { "epoch": 0.17460608281421766, "grad_norm": 0.5091807084893948, "learning_rate": 9.832377057262015e-06, "loss": 0.5516, "step": 953 }, { "epoch": 0.17478930010993038, "grad_norm": 0.48065067696295527, "learning_rate": 9.831554985116155e-06, "loss": 0.5097, "step": 954 }, { "epoch": 0.1749725174056431, "grad_norm": 0.4310364985277951, "learning_rate": 9.830730936592615e-06, "loss": 0.5025, "step": 955 }, { "epoch": 0.17515573470135581, "grad_norm": 0.39964086093662415, "learning_rate": 9.829904912028477e-06, "loss": 0.494, "step": 956 }, { "epoch": 0.17533895199706853, "grad_norm": 0.40935380850458175, "learning_rate": 9.829076911761631e-06, "loss": 0.5182, "step": 957 }, { "epoch": 0.17552216929278125, "grad_norm": 0.454766087870644, "learning_rate": 9.828246936130777e-06, "loss": 0.5113, "step": 958 }, { "epoch": 0.17570538658849397, "grad_norm": 0.5087972352515657, "learning_rate": 9.827414985475419e-06, "loss": 0.4797, "step": 959 }, { "epoch": 0.17588860388420666, "grad_norm": 0.4255562501722622, "learning_rate": 9.826581060135873e-06, "loss": 0.5045, "step": 960 }, { "epoch": 0.17607182117991937, "grad_norm": 0.45583787528367603, "learning_rate": 9.825745160453264e-06, "loss": 0.4755, "step": 961 }, { "epoch": 0.1762550384756321, "grad_norm": 0.45648883029913284, "learning_rate": 9.824907286769519e-06, "loss": 0.4736, "step": 962 }, { "epoch": 0.1764382557713448, "grad_norm": 0.436071698669034, "learning_rate": 9.824067439427374e-06, "loss": 0.5036, "step": 963 }, { "epoch": 0.17662147306705753, "grad_norm": 0.41502000197263017, "learning_rate": 9.823225618770378e-06, "loss": 0.5121, "step": 964 }, { "epoch": 0.17680469036277024, "grad_norm": 0.450252173902017, "learning_rate": 9.822381825142879e-06, "loss": 0.506, "step": 965 }, { "epoch": 0.17698790765848296, "grad_norm": 0.43561778776659554, "learning_rate": 9.82153605889004e-06, "loss": 0.5304, "step": 966 }, { "epoch": 0.17717112495419568, "grad_norm": 0.4306520452468297, "learning_rate": 9.820688320357823e-06, "loss": 0.5112, "step": 967 }, { "epoch": 0.1773543422499084, "grad_norm": 0.39680977809710444, "learning_rate": 9.819838609893005e-06, "loss": 0.4904, "step": 968 }, { "epoch": 0.17753755954562112, "grad_norm": 0.4225898008101114, "learning_rate": 9.81898692784316e-06, "loss": 0.5137, "step": 969 }, { "epoch": 0.17772077684133383, "grad_norm": 0.4267062187327317, "learning_rate": 9.818133274556679e-06, "loss": 0.517, "step": 970 }, { "epoch": 0.17790399413704655, "grad_norm": 0.4450452397955529, "learning_rate": 9.81727765038275e-06, "loss": 0.4949, "step": 971 }, { "epoch": 0.17808721143275924, "grad_norm": 0.42730228271929965, "learning_rate": 9.816420055671374e-06, "loss": 0.5013, "step": 972 }, { "epoch": 0.17827042872847196, "grad_norm": 0.4421383626296405, "learning_rate": 9.815560490773356e-06, "loss": 0.4679, "step": 973 }, { "epoch": 0.17845364602418468, "grad_norm": 0.43307265096730885, "learning_rate": 9.814698956040305e-06, "loss": 0.4976, "step": 974 }, { "epoch": 0.1786368633198974, "grad_norm": 0.40376464412932705, "learning_rate": 9.813835451824636e-06, "loss": 0.5021, "step": 975 }, { "epoch": 0.1788200806156101, "grad_norm": 0.44775776339538953, "learning_rate": 9.812969978479573e-06, "loss": 0.4824, "step": 976 }, { "epoch": 0.17900329791132283, "grad_norm": 0.468036097947629, "learning_rate": 9.812102536359142e-06, "loss": 0.4966, "step": 977 }, { "epoch": 0.17918651520703555, "grad_norm": 0.49832149697786493, "learning_rate": 9.811233125818176e-06, "loss": 0.4853, "step": 978 }, { "epoch": 0.17936973250274826, "grad_norm": 0.44754723833693727, "learning_rate": 9.810361747212313e-06, "loss": 0.4999, "step": 979 }, { "epoch": 0.17955294979846098, "grad_norm": 0.4747011696315986, "learning_rate": 9.809488400897996e-06, "loss": 0.5108, "step": 980 }, { "epoch": 0.1797361670941737, "grad_norm": 0.4055842519238836, "learning_rate": 9.808613087232473e-06, "loss": 0.4786, "step": 981 }, { "epoch": 0.17991938438988642, "grad_norm": 0.3971701443501286, "learning_rate": 9.807735806573795e-06, "loss": 0.4925, "step": 982 }, { "epoch": 0.18010260168559913, "grad_norm": 0.3784480421793096, "learning_rate": 9.806856559280819e-06, "loss": 0.4762, "step": 983 }, { "epoch": 0.18028581898131182, "grad_norm": 0.41465183173286063, "learning_rate": 9.80597534571321e-06, "loss": 0.5037, "step": 984 }, { "epoch": 0.18046903627702454, "grad_norm": 0.4713550462839489, "learning_rate": 9.80509216623143e-06, "loss": 0.4816, "step": 985 }, { "epoch": 0.18065225357273726, "grad_norm": 0.34923786309805893, "learning_rate": 9.804207021196751e-06, "loss": 0.4903, "step": 986 }, { "epoch": 0.18083547086844998, "grad_norm": 0.4359203385963395, "learning_rate": 9.803319910971248e-06, "loss": 0.5252, "step": 987 }, { "epoch": 0.1810186881641627, "grad_norm": 0.40286824933567683, "learning_rate": 9.802430835917796e-06, "loss": 0.5248, "step": 988 }, { "epoch": 0.1812019054598754, "grad_norm": 0.44314680454994215, "learning_rate": 9.801539796400078e-06, "loss": 0.5071, "step": 989 }, { "epoch": 0.18138512275558813, "grad_norm": 0.44074612784921385, "learning_rate": 9.80064679278258e-06, "loss": 0.5057, "step": 990 }, { "epoch": 0.18156834005130085, "grad_norm": 0.40968360376086815, "learning_rate": 9.799751825430592e-06, "loss": 0.4919, "step": 991 }, { "epoch": 0.18175155734701356, "grad_norm": 0.4204735967781832, "learning_rate": 9.798854894710202e-06, "loss": 0.546, "step": 992 }, { "epoch": 0.18193477464272628, "grad_norm": 0.36210624498761007, "learning_rate": 9.79795600098831e-06, "loss": 0.4904, "step": 993 }, { "epoch": 0.182117991938439, "grad_norm": 0.45193717949202633, "learning_rate": 9.797055144632609e-06, "loss": 0.5292, "step": 994 }, { "epoch": 0.18230120923415172, "grad_norm": 0.415630496229026, "learning_rate": 9.796152326011604e-06, "loss": 0.533, "step": 995 }, { "epoch": 0.1824844265298644, "grad_norm": 0.39372729416385926, "learning_rate": 9.795247545494594e-06, "loss": 0.5147, "step": 996 }, { "epoch": 0.18266764382557713, "grad_norm": 0.43540226077312505, "learning_rate": 9.794340803451692e-06, "loss": 0.49, "step": 997 }, { "epoch": 0.18285086112128984, "grad_norm": 0.39372791981341043, "learning_rate": 9.7934321002538e-06, "loss": 0.4776, "step": 998 }, { "epoch": 0.18303407841700256, "grad_norm": 0.40381508343023165, "learning_rate": 9.792521436272633e-06, "loss": 0.4566, "step": 999 }, { "epoch": 0.18321729571271528, "grad_norm": 0.47729065322343966, "learning_rate": 9.791608811880702e-06, "loss": 0.5013, "step": 1000 } ], "logging_steps": 1, "max_steps": 5458, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 178891845599232.0, "train_batch_size": 2, "trial_name": null, "trial_params": null }