diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,35034 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 0.9160864785635764, + "eval_steps": 500, + "global_step": 5000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00018321729571271528, + "grad_norm": 7.481241625973833, + "learning_rate": 0.0, + "loss": 0.764, + "step": 1 + }, + { + "epoch": 0.00036643459142543056, + "grad_norm": 7.936839352385908, + "learning_rate": 1.8315018315018317e-08, + "loss": 0.7505, + "step": 2 + }, + { + "epoch": 0.0005496518871381459, + "grad_norm": 8.381229660678587, + "learning_rate": 3.6630036630036635e-08, + "loss": 0.7767, + "step": 3 + }, + { + "epoch": 0.0007328691828508611, + "grad_norm": 7.533434041856819, + "learning_rate": 5.494505494505495e-08, + "loss": 0.7545, + "step": 4 + }, + { + "epoch": 0.0009160864785635764, + "grad_norm": 6.4876435204860545, + "learning_rate": 7.326007326007327e-08, + "loss": 0.7254, + "step": 5 + }, + { + "epoch": 0.0010993037742762918, + "grad_norm": 6.984439562746247, + "learning_rate": 9.157509157509159e-08, + "loss": 0.7238, + "step": 6 + }, + { + "epoch": 0.001282521069989007, + "grad_norm": 8.432795226889532, + "learning_rate": 1.098901098901099e-07, + "loss": 0.7667, + "step": 7 + }, + { + "epoch": 0.0014657383657017222, + "grad_norm": 6.615389400475257, + "learning_rate": 1.282051282051282e-07, + "loss": 0.739, + "step": 8 + }, + { + "epoch": 0.0016489556614144375, + "grad_norm": 7.043438048491804, + "learning_rate": 1.4652014652014654e-07, + "loss": 0.7412, + "step": 9 + }, + { + "epoch": 0.001832172957127153, + "grad_norm": 7.763888342467084, + "learning_rate": 1.6483516483516484e-07, + "loss": 0.739, + "step": 10 + }, + { + "epoch": 0.002015390252839868, + "grad_norm": 7.115680896600135, + "learning_rate": 1.8315018315018317e-07, + "loss": 0.6814, + "step": 11 + }, + { + "epoch": 0.0021986075485525836, + "grad_norm": 9.472751553447468, + "learning_rate": 2.014652014652015e-07, + "loss": 0.7694, + "step": 12 + }, + { + "epoch": 0.0023818248442652986, + "grad_norm": 9.164365397122241, + "learning_rate": 2.197802197802198e-07, + "loss": 0.8101, + "step": 13 + }, + { + "epoch": 0.002565042139978014, + "grad_norm": 7.831934204465113, + "learning_rate": 2.3809523809523811e-07, + "loss": 0.7504, + "step": 14 + }, + { + "epoch": 0.002748259435690729, + "grad_norm": 6.7423320112288865, + "learning_rate": 2.564102564102564e-07, + "loss": 0.7442, + "step": 15 + }, + { + "epoch": 0.0029314767314034445, + "grad_norm": 7.8103949840893145, + "learning_rate": 2.7472527472527475e-07, + "loss": 0.7782, + "step": 16 + }, + { + "epoch": 0.00311469402711616, + "grad_norm": 6.1519268993583465, + "learning_rate": 2.930402930402931e-07, + "loss": 0.663, + "step": 17 + }, + { + "epoch": 0.003297911322828875, + "grad_norm": 8.049047835270914, + "learning_rate": 3.113553113553114e-07, + "loss": 0.7934, + "step": 18 + }, + { + "epoch": 0.0034811286185415903, + "grad_norm": 6.388709624292669, + "learning_rate": 3.296703296703297e-07, + "loss": 0.7241, + "step": 19 + }, + { + "epoch": 0.003664345914254306, + "grad_norm": 7.039170133500172, + "learning_rate": 3.47985347985348e-07, + "loss": 0.7133, + "step": 20 + }, + { + "epoch": 0.003847563209967021, + "grad_norm": 8.416142106876258, + "learning_rate": 3.6630036630036635e-07, + "loss": 0.7787, + "step": 21 + }, + { + "epoch": 0.004030780505679736, + "grad_norm": 5.942542945633484, + "learning_rate": 3.846153846153847e-07, + "loss": 0.692, + "step": 22 + }, + { + "epoch": 0.004213997801392451, + "grad_norm": 6.397620608740201, + "learning_rate": 4.02930402930403e-07, + "loss": 0.6885, + "step": 23 + }, + { + "epoch": 0.004397215097105167, + "grad_norm": 4.789506080162222, + "learning_rate": 4.212454212454213e-07, + "loss": 0.6818, + "step": 24 + }, + { + "epoch": 0.004580432392817882, + "grad_norm": 5.4323837663846275, + "learning_rate": 4.395604395604396e-07, + "loss": 0.7202, + "step": 25 + }, + { + "epoch": 0.004763649688530597, + "grad_norm": 6.001418955741802, + "learning_rate": 4.578754578754579e-07, + "loss": 0.6915, + "step": 26 + }, + { + "epoch": 0.004946866984243312, + "grad_norm": 6.931290264891086, + "learning_rate": 4.7619047619047623e-07, + "loss": 0.7214, + "step": 27 + }, + { + "epoch": 0.005130084279956028, + "grad_norm": 5.423177642103934, + "learning_rate": 4.945054945054946e-07, + "loss": 0.6627, + "step": 28 + }, + { + "epoch": 0.005313301575668743, + "grad_norm": 5.824082478920574, + "learning_rate": 5.128205128205128e-07, + "loss": 0.7128, + "step": 29 + }, + { + "epoch": 0.005496518871381458, + "grad_norm": 4.837515053542817, + "learning_rate": 5.311355311355311e-07, + "loss": 0.7063, + "step": 30 + }, + { + "epoch": 0.005679736167094174, + "grad_norm": 4.2178492878732, + "learning_rate": 5.494505494505495e-07, + "loss": 0.6667, + "step": 31 + }, + { + "epoch": 0.005862953462806889, + "grad_norm": 3.7310273500960283, + "learning_rate": 5.677655677655678e-07, + "loss": 0.6724, + "step": 32 + }, + { + "epoch": 0.006046170758519604, + "grad_norm": 3.4436166557785013, + "learning_rate": 5.860805860805862e-07, + "loss": 0.6158, + "step": 33 + }, + { + "epoch": 0.00622938805423232, + "grad_norm": 2.6914219359650264, + "learning_rate": 6.043956043956044e-07, + "loss": 0.6384, + "step": 34 + }, + { + "epoch": 0.006412605349945035, + "grad_norm": 2.8611154351289803, + "learning_rate": 6.227106227106228e-07, + "loss": 0.6564, + "step": 35 + }, + { + "epoch": 0.00659582264565775, + "grad_norm": 2.915608387835296, + "learning_rate": 6.41025641025641e-07, + "loss": 0.6556, + "step": 36 + }, + { + "epoch": 0.006779039941370466, + "grad_norm": 3.104210670394775, + "learning_rate": 6.593406593406594e-07, + "loss": 0.6571, + "step": 37 + }, + { + "epoch": 0.006962257237083181, + "grad_norm": 2.416024185643152, + "learning_rate": 6.776556776556777e-07, + "loss": 0.6118, + "step": 38 + }, + { + "epoch": 0.007145474532795896, + "grad_norm": 2.2840327527366187, + "learning_rate": 6.95970695970696e-07, + "loss": 0.5773, + "step": 39 + }, + { + "epoch": 0.007328691828508612, + "grad_norm": 2.994955066139323, + "learning_rate": 7.142857142857143e-07, + "loss": 0.6621, + "step": 40 + }, + { + "epoch": 0.007511909124221327, + "grad_norm": 2.4977146662969956, + "learning_rate": 7.326007326007327e-07, + "loss": 0.5878, + "step": 41 + }, + { + "epoch": 0.007695126419934042, + "grad_norm": 2.2697538551105443, + "learning_rate": 7.50915750915751e-07, + "loss": 0.6161, + "step": 42 + }, + { + "epoch": 0.007878343715646757, + "grad_norm": 2.040084870467292, + "learning_rate": 7.692307692307694e-07, + "loss": 0.625, + "step": 43 + }, + { + "epoch": 0.008061561011359472, + "grad_norm": 1.4264410515550054, + "learning_rate": 7.875457875457876e-07, + "loss": 0.5991, + "step": 44 + }, + { + "epoch": 0.008244778307072188, + "grad_norm": 1.3307329745484617, + "learning_rate": 8.05860805860806e-07, + "loss": 0.5802, + "step": 45 + }, + { + "epoch": 0.008427995602784902, + "grad_norm": 1.3186002922814006, + "learning_rate": 8.241758241758242e-07, + "loss": 0.5971, + "step": 46 + }, + { + "epoch": 0.008611212898497618, + "grad_norm": 1.4454419288227918, + "learning_rate": 8.424908424908426e-07, + "loss": 0.6152, + "step": 47 + }, + { + "epoch": 0.008794430194210334, + "grad_norm": 1.2728865809977838, + "learning_rate": 8.608058608058609e-07, + "loss": 0.6176, + "step": 48 + }, + { + "epoch": 0.008977647489923048, + "grad_norm": 1.2387163683388445, + "learning_rate": 8.791208791208792e-07, + "loss": 0.6233, + "step": 49 + }, + { + "epoch": 0.009160864785635764, + "grad_norm": 1.1465107255795282, + "learning_rate": 8.974358974358975e-07, + "loss": 0.5795, + "step": 50 + }, + { + "epoch": 0.00934408208134848, + "grad_norm": 1.264265250343769, + "learning_rate": 9.157509157509158e-07, + "loss": 0.6262, + "step": 51 + }, + { + "epoch": 0.009527299377061194, + "grad_norm": 0.919312835052228, + "learning_rate": 9.340659340659342e-07, + "loss": 0.5535, + "step": 52 + }, + { + "epoch": 0.00971051667277391, + "grad_norm": 0.8672735287012353, + "learning_rate": 9.523809523809525e-07, + "loss": 0.5327, + "step": 53 + }, + { + "epoch": 0.009893733968486624, + "grad_norm": 0.8288169490471791, + "learning_rate": 9.706959706959708e-07, + "loss": 0.5673, + "step": 54 + }, + { + "epoch": 0.01007695126419934, + "grad_norm": 0.8603030677007734, + "learning_rate": 9.890109890109891e-07, + "loss": 0.5547, + "step": 55 + }, + { + "epoch": 0.010260168559912056, + "grad_norm": 0.7482006424960487, + "learning_rate": 1.0073260073260074e-06, + "loss": 0.6012, + "step": 56 + }, + { + "epoch": 0.01044338585562477, + "grad_norm": 0.6144742944660406, + "learning_rate": 1.0256410256410257e-06, + "loss": 0.5231, + "step": 57 + }, + { + "epoch": 0.010626603151337486, + "grad_norm": 0.634897676727318, + "learning_rate": 1.0439560439560442e-06, + "loss": 0.5726, + "step": 58 + }, + { + "epoch": 0.010809820447050202, + "grad_norm": 0.6410545094813134, + "learning_rate": 1.0622710622710622e-06, + "loss": 0.5695, + "step": 59 + }, + { + "epoch": 0.010993037742762916, + "grad_norm": 0.6609456014192859, + "learning_rate": 1.0805860805860807e-06, + "loss": 0.5805, + "step": 60 + }, + { + "epoch": 0.011176255038475632, + "grad_norm": 0.7416419125176478, + "learning_rate": 1.098901098901099e-06, + "loss": 0.634, + "step": 61 + }, + { + "epoch": 0.011359472334188348, + "grad_norm": 0.6851293138177442, + "learning_rate": 1.1172161172161173e-06, + "loss": 0.5544, + "step": 62 + }, + { + "epoch": 0.011542689629901062, + "grad_norm": 0.6383351643137376, + "learning_rate": 1.1355311355311355e-06, + "loss": 0.557, + "step": 63 + }, + { + "epoch": 0.011725906925613778, + "grad_norm": 0.6348085314283121, + "learning_rate": 1.153846153846154e-06, + "loss": 0.5632, + "step": 64 + }, + { + "epoch": 0.011909124221326494, + "grad_norm": 0.7024363928163538, + "learning_rate": 1.1721611721611723e-06, + "loss": 0.5839, + "step": 65 + }, + { + "epoch": 0.012092341517039208, + "grad_norm": 0.5900911656210056, + "learning_rate": 1.1904761904761906e-06, + "loss": 0.5344, + "step": 66 + }, + { + "epoch": 0.012275558812751924, + "grad_norm": 0.6275963867777459, + "learning_rate": 1.2087912087912089e-06, + "loss": 0.5584, + "step": 67 + }, + { + "epoch": 0.01245877610846464, + "grad_norm": 0.5865171487667508, + "learning_rate": 1.2271062271062271e-06, + "loss": 0.575, + "step": 68 + }, + { + "epoch": 0.012641993404177354, + "grad_norm": 0.674046870317235, + "learning_rate": 1.2454212454212456e-06, + "loss": 0.5744, + "step": 69 + }, + { + "epoch": 0.01282521069989007, + "grad_norm": 0.60918911566989, + "learning_rate": 1.2637362637362637e-06, + "loss": 0.5423, + "step": 70 + }, + { + "epoch": 0.013008427995602785, + "grad_norm": 0.6221860528344966, + "learning_rate": 1.282051282051282e-06, + "loss": 0.5652, + "step": 71 + }, + { + "epoch": 0.0131916452913155, + "grad_norm": 0.606665778835579, + "learning_rate": 1.3003663003663005e-06, + "loss": 0.6123, + "step": 72 + }, + { + "epoch": 0.013374862587028215, + "grad_norm": 0.5884723188789052, + "learning_rate": 1.3186813186813187e-06, + "loss": 0.5863, + "step": 73 + }, + { + "epoch": 0.013558079882740931, + "grad_norm": 0.5824062487652404, + "learning_rate": 1.336996336996337e-06, + "loss": 0.5918, + "step": 74 + }, + { + "epoch": 0.013741297178453645, + "grad_norm": 0.5460196146561194, + "learning_rate": 1.3553113553113553e-06, + "loss": 0.5912, + "step": 75 + }, + { + "epoch": 0.013924514474166361, + "grad_norm": 0.49213553293102813, + "learning_rate": 1.3736263736263738e-06, + "loss": 0.5421, + "step": 76 + }, + { + "epoch": 0.014107731769879077, + "grad_norm": 0.5594168254817149, + "learning_rate": 1.391941391941392e-06, + "loss": 0.5122, + "step": 77 + }, + { + "epoch": 0.014290949065591791, + "grad_norm": 0.5277088034821339, + "learning_rate": 1.4102564102564104e-06, + "loss": 0.5678, + "step": 78 + }, + { + "epoch": 0.014474166361304507, + "grad_norm": 0.49574136870511754, + "learning_rate": 1.4285714285714286e-06, + "loss": 0.5498, + "step": 79 + }, + { + "epoch": 0.014657383657017223, + "grad_norm": 0.5944104688680958, + "learning_rate": 1.4468864468864471e-06, + "loss": 0.5469, + "step": 80 + }, + { + "epoch": 0.014840600952729937, + "grad_norm": 0.4732970613815555, + "learning_rate": 1.4652014652014654e-06, + "loss": 0.5233, + "step": 81 + }, + { + "epoch": 0.015023818248442653, + "grad_norm": 0.49591541086638596, + "learning_rate": 1.4835164835164837e-06, + "loss": 0.5367, + "step": 82 + }, + { + "epoch": 0.015207035544155369, + "grad_norm": 0.4883648219378977, + "learning_rate": 1.501831501831502e-06, + "loss": 0.5782, + "step": 83 + }, + { + "epoch": 0.015390252839868083, + "grad_norm": 0.48559293967760114, + "learning_rate": 1.5201465201465202e-06, + "loss": 0.5578, + "step": 84 + }, + { + "epoch": 0.015573470135580799, + "grad_norm": 0.5802435990379926, + "learning_rate": 1.5384615384615387e-06, + "loss": 0.5562, + "step": 85 + }, + { + "epoch": 0.015756687431293513, + "grad_norm": 0.6655689083082568, + "learning_rate": 1.556776556776557e-06, + "loss": 0.5916, + "step": 86 + }, + { + "epoch": 0.01593990472700623, + "grad_norm": 0.5272919501805459, + "learning_rate": 1.5750915750915753e-06, + "loss": 0.5462, + "step": 87 + }, + { + "epoch": 0.016123122022718945, + "grad_norm": 0.5025768009972991, + "learning_rate": 1.5934065934065933e-06, + "loss": 0.564, + "step": 88 + }, + { + "epoch": 0.01630633931843166, + "grad_norm": 0.4868371012830415, + "learning_rate": 1.611721611721612e-06, + "loss": 0.557, + "step": 89 + }, + { + "epoch": 0.016489556614144377, + "grad_norm": 0.4724914851279723, + "learning_rate": 1.6300366300366301e-06, + "loss": 0.534, + "step": 90 + }, + { + "epoch": 0.01667277390985709, + "grad_norm": 0.4933114328584066, + "learning_rate": 1.6483516483516484e-06, + "loss": 0.5628, + "step": 91 + }, + { + "epoch": 0.016855991205569805, + "grad_norm": 0.5429724547645147, + "learning_rate": 1.6666666666666667e-06, + "loss": 0.5693, + "step": 92 + }, + { + "epoch": 0.017039208501282523, + "grad_norm": 0.46970029586030615, + "learning_rate": 1.6849816849816852e-06, + "loss": 0.5761, + "step": 93 + }, + { + "epoch": 0.017222425796995237, + "grad_norm": 0.5466204682716642, + "learning_rate": 1.7032967032967034e-06, + "loss": 0.5407, + "step": 94 + }, + { + "epoch": 0.01740564309270795, + "grad_norm": 0.46448768993410167, + "learning_rate": 1.7216117216117217e-06, + "loss": 0.5784, + "step": 95 + }, + { + "epoch": 0.01758886038842067, + "grad_norm": 0.46769210989227256, + "learning_rate": 1.73992673992674e-06, + "loss": 0.5262, + "step": 96 + }, + { + "epoch": 0.017772077684133383, + "grad_norm": 0.4412027593954725, + "learning_rate": 1.7582417582417585e-06, + "loss": 0.5608, + "step": 97 + }, + { + "epoch": 0.017955294979846097, + "grad_norm": 0.5975733130220022, + "learning_rate": 1.7765567765567768e-06, + "loss": 0.5633, + "step": 98 + }, + { + "epoch": 0.018138512275558814, + "grad_norm": 0.46156861116006753, + "learning_rate": 1.794871794871795e-06, + "loss": 0.5576, + "step": 99 + }, + { + "epoch": 0.01832172957127153, + "grad_norm": 0.5193644526534718, + "learning_rate": 1.8131868131868133e-06, + "loss": 0.5533, + "step": 100 + }, + { + "epoch": 0.018504946866984243, + "grad_norm": 0.479596247036775, + "learning_rate": 1.8315018315018316e-06, + "loss": 0.5337, + "step": 101 + }, + { + "epoch": 0.01868816416269696, + "grad_norm": 0.4610500867184236, + "learning_rate": 1.84981684981685e-06, + "loss": 0.5314, + "step": 102 + }, + { + "epoch": 0.018871381458409674, + "grad_norm": 0.481950984787821, + "learning_rate": 1.8681318681318684e-06, + "loss": 0.5865, + "step": 103 + }, + { + "epoch": 0.01905459875412239, + "grad_norm": 0.5225151739123198, + "learning_rate": 1.8864468864468866e-06, + "loss": 0.5694, + "step": 104 + }, + { + "epoch": 0.019237816049835106, + "grad_norm": 0.46381468108353424, + "learning_rate": 1.904761904761905e-06, + "loss": 0.557, + "step": 105 + }, + { + "epoch": 0.01942103334554782, + "grad_norm": 0.45411743556679485, + "learning_rate": 1.9230769230769234e-06, + "loss": 0.5648, + "step": 106 + }, + { + "epoch": 0.019604250641260534, + "grad_norm": 0.4859210831259208, + "learning_rate": 1.9413919413919417e-06, + "loss": 0.5685, + "step": 107 + }, + { + "epoch": 0.01978746793697325, + "grad_norm": 0.4205701080898951, + "learning_rate": 1.95970695970696e-06, + "loss": 0.5426, + "step": 108 + }, + { + "epoch": 0.019970685232685966, + "grad_norm": 0.5044757429436201, + "learning_rate": 1.9780219780219782e-06, + "loss": 0.5509, + "step": 109 + }, + { + "epoch": 0.02015390252839868, + "grad_norm": 0.47571345015861244, + "learning_rate": 1.9963369963369965e-06, + "loss": 0.5281, + "step": 110 + }, + { + "epoch": 0.020337119824111394, + "grad_norm": 0.5681283820477757, + "learning_rate": 2.0146520146520148e-06, + "loss": 0.5398, + "step": 111 + }, + { + "epoch": 0.020520337119824112, + "grad_norm": 0.48199138528425167, + "learning_rate": 2.032967032967033e-06, + "loss": 0.56, + "step": 112 + }, + { + "epoch": 0.020703554415536826, + "grad_norm": 0.4841461198775233, + "learning_rate": 2.0512820512820513e-06, + "loss": 0.5454, + "step": 113 + }, + { + "epoch": 0.02088677171124954, + "grad_norm": 0.4680411378560794, + "learning_rate": 2.0695970695970696e-06, + "loss": 0.516, + "step": 114 + }, + { + "epoch": 0.021069989006962258, + "grad_norm": 0.4426302566203345, + "learning_rate": 2.0879120879120883e-06, + "loss": 0.5493, + "step": 115 + }, + { + "epoch": 0.021253206302674972, + "grad_norm": 0.5379521772056074, + "learning_rate": 2.1062271062271066e-06, + "loss": 0.5642, + "step": 116 + }, + { + "epoch": 0.021436423598387686, + "grad_norm": 0.42644965401865687, + "learning_rate": 2.1245421245421245e-06, + "loss": 0.564, + "step": 117 + }, + { + "epoch": 0.021619640894100404, + "grad_norm": 1.3341683924694292, + "learning_rate": 2.1428571428571427e-06, + "loss": 0.5831, + "step": 118 + }, + { + "epoch": 0.021802858189813118, + "grad_norm": 0.5613781478585862, + "learning_rate": 2.1611721611721614e-06, + "loss": 0.5598, + "step": 119 + }, + { + "epoch": 0.021986075485525832, + "grad_norm": 0.4389899980262906, + "learning_rate": 2.1794871794871797e-06, + "loss": 0.5498, + "step": 120 + }, + { + "epoch": 0.02216929278123855, + "grad_norm": 0.5433793107490897, + "learning_rate": 2.197802197802198e-06, + "loss": 0.5838, + "step": 121 + }, + { + "epoch": 0.022352510076951264, + "grad_norm": 0.49992126724035435, + "learning_rate": 2.2161172161172163e-06, + "loss": 0.5581, + "step": 122 + }, + { + "epoch": 0.022535727372663978, + "grad_norm": 0.5192253683114394, + "learning_rate": 2.2344322344322345e-06, + "loss": 0.5963, + "step": 123 + }, + { + "epoch": 0.022718944668376696, + "grad_norm": 0.5218720529273816, + "learning_rate": 2.252747252747253e-06, + "loss": 0.5568, + "step": 124 + }, + { + "epoch": 0.02290216196408941, + "grad_norm": 0.41135006043138017, + "learning_rate": 2.271062271062271e-06, + "loss": 0.5256, + "step": 125 + }, + { + "epoch": 0.023085379259802124, + "grad_norm": 0.4960132808732071, + "learning_rate": 2.2893772893772894e-06, + "loss": 0.5636, + "step": 126 + }, + { + "epoch": 0.02326859655551484, + "grad_norm": 0.40294874528313, + "learning_rate": 2.307692307692308e-06, + "loss": 0.5373, + "step": 127 + }, + { + "epoch": 0.023451813851227556, + "grad_norm": 0.49129951716432685, + "learning_rate": 2.3260073260073264e-06, + "loss": 0.5571, + "step": 128 + }, + { + "epoch": 0.02363503114694027, + "grad_norm": 0.4541626724950761, + "learning_rate": 2.3443223443223446e-06, + "loss": 0.5407, + "step": 129 + }, + { + "epoch": 0.023818248442652987, + "grad_norm": 0.44854257290105054, + "learning_rate": 2.362637362637363e-06, + "loss": 0.5516, + "step": 130 + }, + { + "epoch": 0.0240014657383657, + "grad_norm": 0.4439039076319426, + "learning_rate": 2.380952380952381e-06, + "loss": 0.5192, + "step": 131 + }, + { + "epoch": 0.024184683034078416, + "grad_norm": 0.4410959524536652, + "learning_rate": 2.3992673992673995e-06, + "loss": 0.5678, + "step": 132 + }, + { + "epoch": 0.024367900329791133, + "grad_norm": 0.49361868947998544, + "learning_rate": 2.4175824175824177e-06, + "loss": 0.5591, + "step": 133 + }, + { + "epoch": 0.024551117625503847, + "grad_norm": 0.5023022664764156, + "learning_rate": 2.435897435897436e-06, + "loss": 0.5818, + "step": 134 + }, + { + "epoch": 0.02473433492121656, + "grad_norm": 0.4287249843742935, + "learning_rate": 2.4542124542124543e-06, + "loss": 0.5532, + "step": 135 + }, + { + "epoch": 0.02491755221692928, + "grad_norm": 0.43595844702277614, + "learning_rate": 2.472527472527473e-06, + "loss": 0.534, + "step": 136 + }, + { + "epoch": 0.025100769512641993, + "grad_norm": 0.5546618594341888, + "learning_rate": 2.4908424908424913e-06, + "loss": 0.5058, + "step": 137 + }, + { + "epoch": 0.025283986808354707, + "grad_norm": 0.5017549598118495, + "learning_rate": 2.509157509157509e-06, + "loss": 0.5416, + "step": 138 + }, + { + "epoch": 0.025467204104067425, + "grad_norm": 0.47047331776441903, + "learning_rate": 2.5274725274725274e-06, + "loss": 0.5557, + "step": 139 + }, + { + "epoch": 0.02565042139978014, + "grad_norm": 0.4881037742451912, + "learning_rate": 2.5457875457875457e-06, + "loss": 0.5494, + "step": 140 + }, + { + "epoch": 0.025833638695492853, + "grad_norm": 0.41933017046045795, + "learning_rate": 2.564102564102564e-06, + "loss": 0.5729, + "step": 141 + }, + { + "epoch": 0.02601685599120557, + "grad_norm": 0.39974209370594543, + "learning_rate": 2.582417582417583e-06, + "loss": 0.5381, + "step": 142 + }, + { + "epoch": 0.026200073286918285, + "grad_norm": 0.4473481461077892, + "learning_rate": 2.600732600732601e-06, + "loss": 0.5336, + "step": 143 + }, + { + "epoch": 0.026383290582631, + "grad_norm": 0.45197895778035446, + "learning_rate": 2.6190476190476192e-06, + "loss": 0.5271, + "step": 144 + }, + { + "epoch": 0.026566507878343717, + "grad_norm": 0.49529182062519755, + "learning_rate": 2.6373626373626375e-06, + "loss": 0.5468, + "step": 145 + }, + { + "epoch": 0.02674972517405643, + "grad_norm": 0.5413416851523152, + "learning_rate": 2.6556776556776558e-06, + "loss": 0.5807, + "step": 146 + }, + { + "epoch": 0.026932942469769145, + "grad_norm": 0.4703582853460863, + "learning_rate": 2.673992673992674e-06, + "loss": 0.538, + "step": 147 + }, + { + "epoch": 0.027116159765481863, + "grad_norm": 0.4332295907365602, + "learning_rate": 2.6923076923076923e-06, + "loss": 0.5167, + "step": 148 + }, + { + "epoch": 0.027299377061194577, + "grad_norm": 0.458616583095158, + "learning_rate": 2.7106227106227106e-06, + "loss": 0.5226, + "step": 149 + }, + { + "epoch": 0.02748259435690729, + "grad_norm": 0.43740744109233864, + "learning_rate": 2.728937728937729e-06, + "loss": 0.541, + "step": 150 + }, + { + "epoch": 0.02766581165262001, + "grad_norm": 0.5735574364330706, + "learning_rate": 2.7472527472527476e-06, + "loss": 0.5835, + "step": 151 + }, + { + "epoch": 0.027849028948332723, + "grad_norm": 0.5055518766772779, + "learning_rate": 2.765567765567766e-06, + "loss": 0.5631, + "step": 152 + }, + { + "epoch": 0.028032246244045437, + "grad_norm": 0.43271616931082146, + "learning_rate": 2.783882783882784e-06, + "loss": 0.5426, + "step": 153 + }, + { + "epoch": 0.028215463539758154, + "grad_norm": 0.4971463373765023, + "learning_rate": 2.8021978021978024e-06, + "loss": 0.5389, + "step": 154 + }, + { + "epoch": 0.02839868083547087, + "grad_norm": 0.4591734918280654, + "learning_rate": 2.8205128205128207e-06, + "loss": 0.5309, + "step": 155 + }, + { + "epoch": 0.028581898131183583, + "grad_norm": 0.45247087926212365, + "learning_rate": 2.838827838827839e-06, + "loss": 0.5536, + "step": 156 + }, + { + "epoch": 0.0287651154268963, + "grad_norm": 0.46626015341742694, + "learning_rate": 2.8571428571428573e-06, + "loss": 0.559, + "step": 157 + }, + { + "epoch": 0.028948332722609015, + "grad_norm": 0.4834854871636071, + "learning_rate": 2.8754578754578755e-06, + "loss": 0.5476, + "step": 158 + }, + { + "epoch": 0.02913155001832173, + "grad_norm": 0.570546858013933, + "learning_rate": 2.8937728937728942e-06, + "loss": 0.5394, + "step": 159 + }, + { + "epoch": 0.029314767314034446, + "grad_norm": 0.45441480104130544, + "learning_rate": 2.9120879120879125e-06, + "loss": 0.5029, + "step": 160 + }, + { + "epoch": 0.02949798460974716, + "grad_norm": 0.5206871098196969, + "learning_rate": 2.930402930402931e-06, + "loss": 0.528, + "step": 161 + }, + { + "epoch": 0.029681201905459875, + "grad_norm": 0.43359542229542136, + "learning_rate": 2.948717948717949e-06, + "loss": 0.5502, + "step": 162 + }, + { + "epoch": 0.029864419201172592, + "grad_norm": 0.46302617065984364, + "learning_rate": 2.9670329670329673e-06, + "loss": 0.5447, + "step": 163 + }, + { + "epoch": 0.030047636496885306, + "grad_norm": 0.3915642271916536, + "learning_rate": 2.9853479853479856e-06, + "loss": 0.503, + "step": 164 + }, + { + "epoch": 0.03023085379259802, + "grad_norm": 0.42991246482419715, + "learning_rate": 3.003663003663004e-06, + "loss": 0.521, + "step": 165 + }, + { + "epoch": 0.030414071088310738, + "grad_norm": 0.468175424095518, + "learning_rate": 3.021978021978022e-06, + "loss": 0.5101, + "step": 166 + }, + { + "epoch": 0.030597288384023452, + "grad_norm": 0.6468735604295471, + "learning_rate": 3.0402930402930405e-06, + "loss": 0.5617, + "step": 167 + }, + { + "epoch": 0.030780505679736166, + "grad_norm": 0.5058923699848836, + "learning_rate": 3.058608058608059e-06, + "loss": 0.5154, + "step": 168 + }, + { + "epoch": 0.030963722975448884, + "grad_norm": 0.45437537978064513, + "learning_rate": 3.0769230769230774e-06, + "loss": 0.4993, + "step": 169 + }, + { + "epoch": 0.031146940271161598, + "grad_norm": 0.467183819204265, + "learning_rate": 3.0952380952380957e-06, + "loss": 0.5701, + "step": 170 + }, + { + "epoch": 0.031330157566874316, + "grad_norm": 0.3835674459614267, + "learning_rate": 3.113553113553114e-06, + "loss": 0.4902, + "step": 171 + }, + { + "epoch": 0.031513374862587026, + "grad_norm": 0.4292795112150875, + "learning_rate": 3.1318681318681323e-06, + "loss": 0.5237, + "step": 172 + }, + { + "epoch": 0.031696592158299744, + "grad_norm": 0.6593051731116806, + "learning_rate": 3.1501831501831505e-06, + "loss": 0.5361, + "step": 173 + }, + { + "epoch": 0.03187980945401246, + "grad_norm": 0.4718436483558741, + "learning_rate": 3.1684981684981684e-06, + "loss": 0.5324, + "step": 174 + }, + { + "epoch": 0.03206302674972517, + "grad_norm": 0.48018416222395494, + "learning_rate": 3.1868131868131867e-06, + "loss": 0.5273, + "step": 175 + }, + { + "epoch": 0.03224624404543789, + "grad_norm": 0.6316028239985647, + "learning_rate": 3.205128205128206e-06, + "loss": 0.5346, + "step": 176 + }, + { + "epoch": 0.03242946134115061, + "grad_norm": 0.4596646339699305, + "learning_rate": 3.223443223443224e-06, + "loss": 0.5263, + "step": 177 + }, + { + "epoch": 0.03261267863686332, + "grad_norm": 0.6529252419894329, + "learning_rate": 3.2417582417582424e-06, + "loss": 0.5442, + "step": 178 + }, + { + "epoch": 0.032795895932576036, + "grad_norm": 0.5538108341969676, + "learning_rate": 3.2600732600732602e-06, + "loss": 0.543, + "step": 179 + }, + { + "epoch": 0.03297911322828875, + "grad_norm": 0.46058359751530825, + "learning_rate": 3.2783882783882785e-06, + "loss": 0.5101, + "step": 180 + }, + { + "epoch": 0.033162330524001464, + "grad_norm": 0.42238032667898895, + "learning_rate": 3.2967032967032968e-06, + "loss": 0.5487, + "step": 181 + }, + { + "epoch": 0.03334554781971418, + "grad_norm": 0.49973876221605035, + "learning_rate": 3.315018315018315e-06, + "loss": 0.5333, + "step": 182 + }, + { + "epoch": 0.0335287651154269, + "grad_norm": 0.49185079032879564, + "learning_rate": 3.3333333333333333e-06, + "loss": 0.5733, + "step": 183 + }, + { + "epoch": 0.03371198241113961, + "grad_norm": 0.41357500203470565, + "learning_rate": 3.3516483516483516e-06, + "loss": 0.5108, + "step": 184 + }, + { + "epoch": 0.03389519970685233, + "grad_norm": 0.45597659979596383, + "learning_rate": 3.3699633699633703e-06, + "loss": 0.5316, + "step": 185 + }, + { + "epoch": 0.034078417002565045, + "grad_norm": 0.523873134306111, + "learning_rate": 3.3882783882783886e-06, + "loss": 0.5516, + "step": 186 + }, + { + "epoch": 0.034261634298277756, + "grad_norm": 0.4604705346503309, + "learning_rate": 3.406593406593407e-06, + "loss": 0.5463, + "step": 187 + }, + { + "epoch": 0.03444485159399047, + "grad_norm": 0.4494660179181277, + "learning_rate": 3.424908424908425e-06, + "loss": 0.5195, + "step": 188 + }, + { + "epoch": 0.03462806888970319, + "grad_norm": 0.47161535336220833, + "learning_rate": 3.4432234432234434e-06, + "loss": 0.5495, + "step": 189 + }, + { + "epoch": 0.0348112861854159, + "grad_norm": 0.41422931205807795, + "learning_rate": 3.4615384615384617e-06, + "loss": 0.5165, + "step": 190 + }, + { + "epoch": 0.03499450348112862, + "grad_norm": 0.48552414748511474, + "learning_rate": 3.47985347985348e-06, + "loss": 0.5207, + "step": 191 + }, + { + "epoch": 0.03517772077684134, + "grad_norm": 0.49918231460984896, + "learning_rate": 3.4981684981684982e-06, + "loss": 0.5055, + "step": 192 + }, + { + "epoch": 0.03536093807255405, + "grad_norm": 0.5047784031000427, + "learning_rate": 3.516483516483517e-06, + "loss": 0.5541, + "step": 193 + }, + { + "epoch": 0.035544155368266765, + "grad_norm": 0.45515393167769386, + "learning_rate": 3.5347985347985352e-06, + "loss": 0.5324, + "step": 194 + }, + { + "epoch": 0.03572737266397948, + "grad_norm": 0.4697755766255993, + "learning_rate": 3.5531135531135535e-06, + "loss": 0.5317, + "step": 195 + }, + { + "epoch": 0.035910589959692193, + "grad_norm": 0.47670842684148323, + "learning_rate": 3.5714285714285718e-06, + "loss": 0.5331, + "step": 196 + }, + { + "epoch": 0.03609380725540491, + "grad_norm": 0.46160363428859325, + "learning_rate": 3.58974358974359e-06, + "loss": 0.5645, + "step": 197 + }, + { + "epoch": 0.03627702455111763, + "grad_norm": 0.5456279444043086, + "learning_rate": 3.6080586080586083e-06, + "loss": 0.4923, + "step": 198 + }, + { + "epoch": 0.03646024184683034, + "grad_norm": 0.45679324633890633, + "learning_rate": 3.6263736263736266e-06, + "loss": 0.5408, + "step": 199 + }, + { + "epoch": 0.03664345914254306, + "grad_norm": 0.4038620329916062, + "learning_rate": 3.644688644688645e-06, + "loss": 0.5542, + "step": 200 + }, + { + "epoch": 0.036826676438255775, + "grad_norm": 0.42338470275172285, + "learning_rate": 3.663003663003663e-06, + "loss": 0.5509, + "step": 201 + }, + { + "epoch": 0.037009893733968485, + "grad_norm": 0.48862631901451187, + "learning_rate": 3.681318681318682e-06, + "loss": 0.533, + "step": 202 + }, + { + "epoch": 0.0371931110296812, + "grad_norm": 0.4339739485667441, + "learning_rate": 3.6996336996337e-06, + "loss": 0.5331, + "step": 203 + }, + { + "epoch": 0.03737632832539392, + "grad_norm": 0.4040593227125272, + "learning_rate": 3.7179487179487184e-06, + "loss": 0.568, + "step": 204 + }, + { + "epoch": 0.03755954562110663, + "grad_norm": 0.5025361443024049, + "learning_rate": 3.7362637362637367e-06, + "loss": 0.5616, + "step": 205 + }, + { + "epoch": 0.03774276291681935, + "grad_norm": 0.43671528407867277, + "learning_rate": 3.754578754578755e-06, + "loss": 0.5468, + "step": 206 + }, + { + "epoch": 0.037925980212532066, + "grad_norm": 0.5426409729355149, + "learning_rate": 3.7728937728937733e-06, + "loss": 0.5274, + "step": 207 + }, + { + "epoch": 0.03810919750824478, + "grad_norm": 0.4383066270524436, + "learning_rate": 3.7912087912087915e-06, + "loss": 0.5491, + "step": 208 + }, + { + "epoch": 0.038292414803957495, + "grad_norm": 0.4541571165503901, + "learning_rate": 3.80952380952381e-06, + "loss": 0.511, + "step": 209 + }, + { + "epoch": 0.03847563209967021, + "grad_norm": 0.42914694472434756, + "learning_rate": 3.827838827838828e-06, + "loss": 0.4891, + "step": 210 + }, + { + "epoch": 0.03865884939538292, + "grad_norm": 0.4583586099579229, + "learning_rate": 3.846153846153847e-06, + "loss": 0.5359, + "step": 211 + }, + { + "epoch": 0.03884206669109564, + "grad_norm": 0.46332307938408596, + "learning_rate": 3.864468864468865e-06, + "loss": 0.506, + "step": 212 + }, + { + "epoch": 0.03902528398680836, + "grad_norm": 0.4979093495563886, + "learning_rate": 3.882783882783883e-06, + "loss": 0.5523, + "step": 213 + }, + { + "epoch": 0.03920850128252107, + "grad_norm": 0.524225251704003, + "learning_rate": 3.901098901098901e-06, + "loss": 0.543, + "step": 214 + }, + { + "epoch": 0.039391718578233786, + "grad_norm": 0.5147235908704881, + "learning_rate": 3.91941391941392e-06, + "loss": 0.5465, + "step": 215 + }, + { + "epoch": 0.0395749358739465, + "grad_norm": 0.4170039710750516, + "learning_rate": 3.937728937728938e-06, + "loss": 0.557, + "step": 216 + }, + { + "epoch": 0.039758153169659215, + "grad_norm": 0.4459151957597281, + "learning_rate": 3.9560439560439565e-06, + "loss": 0.5176, + "step": 217 + }, + { + "epoch": 0.03994137046537193, + "grad_norm": 0.4387028825678643, + "learning_rate": 3.974358974358974e-06, + "loss": 0.5433, + "step": 218 + }, + { + "epoch": 0.04012458776108464, + "grad_norm": 0.4606220492260897, + "learning_rate": 3.992673992673993e-06, + "loss": 0.5558, + "step": 219 + }, + { + "epoch": 0.04030780505679736, + "grad_norm": 0.43876788524484817, + "learning_rate": 4.010989010989012e-06, + "loss": 0.496, + "step": 220 + }, + { + "epoch": 0.04049102235251008, + "grad_norm": 0.5354442317270937, + "learning_rate": 4.0293040293040296e-06, + "loss": 0.4883, + "step": 221 + }, + { + "epoch": 0.04067423964822279, + "grad_norm": 0.4471338769246311, + "learning_rate": 4.047619047619048e-06, + "loss": 0.5248, + "step": 222 + }, + { + "epoch": 0.040857456943935506, + "grad_norm": 0.4728450428012797, + "learning_rate": 4.065934065934066e-06, + "loss": 0.5412, + "step": 223 + }, + { + "epoch": 0.041040674239648224, + "grad_norm": 0.5021628515290991, + "learning_rate": 4.084249084249085e-06, + "loss": 0.5398, + "step": 224 + }, + { + "epoch": 0.041223891535360935, + "grad_norm": 0.379469323319607, + "learning_rate": 4.102564102564103e-06, + "loss": 0.5484, + "step": 225 + }, + { + "epoch": 0.04140710883107365, + "grad_norm": 0.4682923741442823, + "learning_rate": 4.120879120879121e-06, + "loss": 0.5221, + "step": 226 + }, + { + "epoch": 0.04159032612678637, + "grad_norm": 0.5298230539031403, + "learning_rate": 4.139194139194139e-06, + "loss": 0.5248, + "step": 227 + }, + { + "epoch": 0.04177354342249908, + "grad_norm": 0.448919135267925, + "learning_rate": 4.157509157509158e-06, + "loss": 0.5492, + "step": 228 + }, + { + "epoch": 0.0419567607182118, + "grad_norm": 0.4651550006247672, + "learning_rate": 4.175824175824177e-06, + "loss": 0.5244, + "step": 229 + }, + { + "epoch": 0.042139978013924516, + "grad_norm": 0.46885786974024124, + "learning_rate": 4.1941391941391945e-06, + "loss": 0.5145, + "step": 230 + }, + { + "epoch": 0.04232319530963723, + "grad_norm": 0.9939588576814024, + "learning_rate": 4.212454212454213e-06, + "loss": 0.507, + "step": 231 + }, + { + "epoch": 0.042506412605349944, + "grad_norm": 0.42445387506164906, + "learning_rate": 4.230769230769231e-06, + "loss": 0.528, + "step": 232 + }, + { + "epoch": 0.04268962990106266, + "grad_norm": 0.4386077325175301, + "learning_rate": 4.249084249084249e-06, + "loss": 0.5235, + "step": 233 + }, + { + "epoch": 0.04287284719677537, + "grad_norm": 0.4275787850644743, + "learning_rate": 4.267399267399268e-06, + "loss": 0.5268, + "step": 234 + }, + { + "epoch": 0.04305606449248809, + "grad_norm": 0.3819534900324145, + "learning_rate": 4.2857142857142855e-06, + "loss": 0.5108, + "step": 235 + }, + { + "epoch": 0.04323928178820081, + "grad_norm": 0.5069196499403993, + "learning_rate": 4.304029304029305e-06, + "loss": 0.4914, + "step": 236 + }, + { + "epoch": 0.04342249908391352, + "grad_norm": 0.4224671659155117, + "learning_rate": 4.322344322344323e-06, + "loss": 0.531, + "step": 237 + }, + { + "epoch": 0.043605716379626236, + "grad_norm": 0.49892435842001814, + "learning_rate": 4.340659340659341e-06, + "loss": 0.5309, + "step": 238 + }, + { + "epoch": 0.043788933675338954, + "grad_norm": 0.4435906661481072, + "learning_rate": 4.358974358974359e-06, + "loss": 0.5386, + "step": 239 + }, + { + "epoch": 0.043972150971051664, + "grad_norm": 0.4646701721197805, + "learning_rate": 4.377289377289377e-06, + "loss": 0.5445, + "step": 240 + }, + { + "epoch": 0.04415536826676438, + "grad_norm": 0.42345412346700445, + "learning_rate": 4.395604395604396e-06, + "loss": 0.5171, + "step": 241 + }, + { + "epoch": 0.0443385855624771, + "grad_norm": 0.4664513677280782, + "learning_rate": 4.413919413919414e-06, + "loss": 0.5471, + "step": 242 + }, + { + "epoch": 0.04452180285818981, + "grad_norm": 0.5007403600133091, + "learning_rate": 4.4322344322344325e-06, + "loss": 0.5224, + "step": 243 + }, + { + "epoch": 0.04470502015390253, + "grad_norm": 0.4202775330369337, + "learning_rate": 4.45054945054945e-06, + "loss": 0.5283, + "step": 244 + }, + { + "epoch": 0.044888237449615245, + "grad_norm": 0.485312343325994, + "learning_rate": 4.468864468864469e-06, + "loss": 0.5365, + "step": 245 + }, + { + "epoch": 0.045071454745327956, + "grad_norm": 0.4796744192562041, + "learning_rate": 4.487179487179488e-06, + "loss": 0.5481, + "step": 246 + }, + { + "epoch": 0.045254672041040674, + "grad_norm": 0.4699525411145978, + "learning_rate": 4.505494505494506e-06, + "loss": 0.526, + "step": 247 + }, + { + "epoch": 0.04543788933675339, + "grad_norm": 0.37235134296143163, + "learning_rate": 4.523809523809524e-06, + "loss": 0.4929, + "step": 248 + }, + { + "epoch": 0.0456211066324661, + "grad_norm": 0.4594065464998237, + "learning_rate": 4.542124542124542e-06, + "loss": 0.5435, + "step": 249 + }, + { + "epoch": 0.04580432392817882, + "grad_norm": 0.4812140084395847, + "learning_rate": 4.560439560439561e-06, + "loss": 0.4714, + "step": 250 + }, + { + "epoch": 0.04598754122389154, + "grad_norm": 0.44275937617791644, + "learning_rate": 4.578754578754579e-06, + "loss": 0.5289, + "step": 251 + }, + { + "epoch": 0.04617075851960425, + "grad_norm": 0.543477861772032, + "learning_rate": 4.5970695970695975e-06, + "loss": 0.5176, + "step": 252 + }, + { + "epoch": 0.046353975815316965, + "grad_norm": 0.9195336625704912, + "learning_rate": 4.615384615384616e-06, + "loss": 0.5315, + "step": 253 + }, + { + "epoch": 0.04653719311102968, + "grad_norm": 0.4922315670719196, + "learning_rate": 4.633699633699634e-06, + "loss": 0.5364, + "step": 254 + }, + { + "epoch": 0.046720410406742394, + "grad_norm": 0.47361690353516367, + "learning_rate": 4.652014652014653e-06, + "loss": 0.5591, + "step": 255 + }, + { + "epoch": 0.04690362770245511, + "grad_norm": 0.5164779414047217, + "learning_rate": 4.6703296703296706e-06, + "loss": 0.5538, + "step": 256 + }, + { + "epoch": 0.04708684499816783, + "grad_norm": 0.421397036777767, + "learning_rate": 4.688644688644689e-06, + "loss": 0.5338, + "step": 257 + }, + { + "epoch": 0.04727006229388054, + "grad_norm": 0.5053871259325204, + "learning_rate": 4.706959706959707e-06, + "loss": 0.544, + "step": 258 + }, + { + "epoch": 0.04745327958959326, + "grad_norm": 0.4607447877406368, + "learning_rate": 4.725274725274726e-06, + "loss": 0.4812, + "step": 259 + }, + { + "epoch": 0.047636496885305975, + "grad_norm": 0.4875422302168998, + "learning_rate": 4.743589743589744e-06, + "loss": 0.5614, + "step": 260 + }, + { + "epoch": 0.047819714181018685, + "grad_norm": 0.41579068200919733, + "learning_rate": 4.761904761904762e-06, + "loss": 0.4729, + "step": 261 + }, + { + "epoch": 0.0480029314767314, + "grad_norm": 0.46282242693556186, + "learning_rate": 4.780219780219781e-06, + "loss": 0.5224, + "step": 262 + }, + { + "epoch": 0.04818614877244412, + "grad_norm": 0.4482830307148575, + "learning_rate": 4.798534798534799e-06, + "loss": 0.5326, + "step": 263 + }, + { + "epoch": 0.04836936606815683, + "grad_norm": 0.4342637761169385, + "learning_rate": 4.816849816849818e-06, + "loss": 0.5056, + "step": 264 + }, + { + "epoch": 0.04855258336386955, + "grad_norm": 0.42342307710917526, + "learning_rate": 4.8351648351648355e-06, + "loss": 0.5049, + "step": 265 + }, + { + "epoch": 0.04873580065958227, + "grad_norm": 0.41567729548709964, + "learning_rate": 4.853479853479854e-06, + "loss": 0.5313, + "step": 266 + }, + { + "epoch": 0.04891901795529498, + "grad_norm": 0.44877730848158315, + "learning_rate": 4.871794871794872e-06, + "loss": 0.5027, + "step": 267 + }, + { + "epoch": 0.049102235251007695, + "grad_norm": 0.46298601715996757, + "learning_rate": 4.890109890109891e-06, + "loss": 0.5418, + "step": 268 + }, + { + "epoch": 0.04928545254672041, + "grad_norm": 0.4233152082129357, + "learning_rate": 4.908424908424909e-06, + "loss": 0.534, + "step": 269 + }, + { + "epoch": 0.04946866984243312, + "grad_norm": 0.42327618076780654, + "learning_rate": 4.926739926739927e-06, + "loss": 0.5073, + "step": 270 + }, + { + "epoch": 0.04965188713814584, + "grad_norm": 0.47132160003804374, + "learning_rate": 4.945054945054946e-06, + "loss": 0.564, + "step": 271 + }, + { + "epoch": 0.04983510443385856, + "grad_norm": 0.467395758671848, + "learning_rate": 4.963369963369964e-06, + "loss": 0.5194, + "step": 272 + }, + { + "epoch": 0.05001832172957127, + "grad_norm": 0.4377241688268797, + "learning_rate": 4.9816849816849826e-06, + "loss": 0.5473, + "step": 273 + }, + { + "epoch": 0.05020153902528399, + "grad_norm": 0.4344310773487788, + "learning_rate": 5e-06, + "loss": 0.5198, + "step": 274 + }, + { + "epoch": 0.050384756320996704, + "grad_norm": 0.505691771505538, + "learning_rate": 5.018315018315018e-06, + "loss": 0.5001, + "step": 275 + }, + { + "epoch": 0.050567973616709415, + "grad_norm": 0.4741702696479342, + "learning_rate": 5.036630036630037e-06, + "loss": 0.5269, + "step": 276 + }, + { + "epoch": 0.05075119091242213, + "grad_norm": 0.4123760853657366, + "learning_rate": 5.054945054945055e-06, + "loss": 0.5366, + "step": 277 + }, + { + "epoch": 0.05093440820813485, + "grad_norm": 0.5165952852785715, + "learning_rate": 5.0732600732600735e-06, + "loss": 0.5629, + "step": 278 + }, + { + "epoch": 0.05111762550384756, + "grad_norm": 0.6018743338441076, + "learning_rate": 5.091575091575091e-06, + "loss": 0.5268, + "step": 279 + }, + { + "epoch": 0.05130084279956028, + "grad_norm": 0.4647905088113548, + "learning_rate": 5.10989010989011e-06, + "loss": 0.5353, + "step": 280 + }, + { + "epoch": 0.051484060095272996, + "grad_norm": 0.4666862939383661, + "learning_rate": 5.128205128205128e-06, + "loss": 0.5582, + "step": 281 + }, + { + "epoch": 0.05166727739098571, + "grad_norm": 0.42295571485014016, + "learning_rate": 5.146520146520147e-06, + "loss": 0.5389, + "step": 282 + }, + { + "epoch": 0.051850494686698424, + "grad_norm": 0.5088982589681916, + "learning_rate": 5.164835164835166e-06, + "loss": 0.5114, + "step": 283 + }, + { + "epoch": 0.05203371198241114, + "grad_norm": 0.39252264391052066, + "learning_rate": 5.183150183150184e-06, + "loss": 0.5078, + "step": 284 + }, + { + "epoch": 0.05221692927812385, + "grad_norm": 0.47464881831711925, + "learning_rate": 5.201465201465202e-06, + "loss": 0.5719, + "step": 285 + }, + { + "epoch": 0.05240014657383657, + "grad_norm": 0.42811181510690394, + "learning_rate": 5.219780219780221e-06, + "loss": 0.536, + "step": 286 + }, + { + "epoch": 0.05258336386954929, + "grad_norm": 0.4983761768332983, + "learning_rate": 5.2380952380952384e-06, + "loss": 0.5454, + "step": 287 + }, + { + "epoch": 0.052766581165262, + "grad_norm": 0.3871752405078846, + "learning_rate": 5.256410256410257e-06, + "loss": 0.5172, + "step": 288 + }, + { + "epoch": 0.052949798460974716, + "grad_norm": 0.44878294094458826, + "learning_rate": 5.274725274725275e-06, + "loss": 0.4965, + "step": 289 + }, + { + "epoch": 0.053133015756687434, + "grad_norm": 0.4193937449864018, + "learning_rate": 5.293040293040294e-06, + "loss": 0.5418, + "step": 290 + }, + { + "epoch": 0.053316233052400144, + "grad_norm": 0.47326184468203625, + "learning_rate": 5.3113553113553116e-06, + "loss": 0.5465, + "step": 291 + }, + { + "epoch": 0.05349945034811286, + "grad_norm": 0.46890596965933473, + "learning_rate": 5.32967032967033e-06, + "loss": 0.5427, + "step": 292 + }, + { + "epoch": 0.05368266764382558, + "grad_norm": 0.4718404244115825, + "learning_rate": 5.347985347985348e-06, + "loss": 0.5315, + "step": 293 + }, + { + "epoch": 0.05386588493953829, + "grad_norm": 0.5313479853203268, + "learning_rate": 5.366300366300367e-06, + "loss": 0.5261, + "step": 294 + }, + { + "epoch": 0.05404910223525101, + "grad_norm": 0.4619862699623299, + "learning_rate": 5.384615384615385e-06, + "loss": 0.5224, + "step": 295 + }, + { + "epoch": 0.054232319530963725, + "grad_norm": 0.4235637361399875, + "learning_rate": 5.402930402930403e-06, + "loss": 0.5484, + "step": 296 + }, + { + "epoch": 0.054415536826676436, + "grad_norm": 0.4826898937600368, + "learning_rate": 5.421245421245421e-06, + "loss": 0.524, + "step": 297 + }, + { + "epoch": 0.054598754122389154, + "grad_norm": 0.43904512964940123, + "learning_rate": 5.43956043956044e-06, + "loss": 0.5486, + "step": 298 + }, + { + "epoch": 0.05478197141810187, + "grad_norm": 0.5045757484045217, + "learning_rate": 5.457875457875458e-06, + "loss": 0.5407, + "step": 299 + }, + { + "epoch": 0.05496518871381458, + "grad_norm": 0.47829971819207484, + "learning_rate": 5.476190476190477e-06, + "loss": 0.5344, + "step": 300 + }, + { + "epoch": 0.0551484060095273, + "grad_norm": 0.416644246441645, + "learning_rate": 5.494505494505495e-06, + "loss": 0.5111, + "step": 301 + }, + { + "epoch": 0.05533162330524002, + "grad_norm": 0.488275746902462, + "learning_rate": 5.512820512820514e-06, + "loss": 0.54, + "step": 302 + }, + { + "epoch": 0.05551484060095273, + "grad_norm": 0.43082352297647686, + "learning_rate": 5.531135531135532e-06, + "loss": 0.5219, + "step": 303 + }, + { + "epoch": 0.055698057896665445, + "grad_norm": 0.41708996725660685, + "learning_rate": 5.5494505494505504e-06, + "loss": 0.5272, + "step": 304 + }, + { + "epoch": 0.05588127519237816, + "grad_norm": 0.4748217492221608, + "learning_rate": 5.567765567765568e-06, + "loss": 0.5439, + "step": 305 + }, + { + "epoch": 0.056064492488090874, + "grad_norm": 0.5257169187612324, + "learning_rate": 5.586080586080587e-06, + "loss": 0.5671, + "step": 306 + }, + { + "epoch": 0.05624770978380359, + "grad_norm": 0.4243472668028098, + "learning_rate": 5.604395604395605e-06, + "loss": 0.5044, + "step": 307 + }, + { + "epoch": 0.05643092707951631, + "grad_norm": 0.43876355547814727, + "learning_rate": 5.6227106227106235e-06, + "loss": 0.4815, + "step": 308 + }, + { + "epoch": 0.05661414437522902, + "grad_norm": 0.43342398130791976, + "learning_rate": 5.641025641025641e-06, + "loss": 0.5308, + "step": 309 + }, + { + "epoch": 0.05679736167094174, + "grad_norm": 0.4660705177035744, + "learning_rate": 5.65934065934066e-06, + "loss": 0.5686, + "step": 310 + }, + { + "epoch": 0.056980578966654455, + "grad_norm": 0.4170384834874546, + "learning_rate": 5.677655677655678e-06, + "loss": 0.4965, + "step": 311 + }, + { + "epoch": 0.057163796262367166, + "grad_norm": 0.4568771189697619, + "learning_rate": 5.695970695970696e-06, + "loss": 0.5074, + "step": 312 + }, + { + "epoch": 0.05734701355807988, + "grad_norm": 0.4666988354433752, + "learning_rate": 5.7142857142857145e-06, + "loss": 0.5225, + "step": 313 + }, + { + "epoch": 0.0575302308537926, + "grad_norm": 0.38305953336818827, + "learning_rate": 5.732600732600732e-06, + "loss": 0.5057, + "step": 314 + }, + { + "epoch": 0.05771344814950531, + "grad_norm": 0.8971752127635111, + "learning_rate": 5.750915750915751e-06, + "loss": 0.5556, + "step": 315 + }, + { + "epoch": 0.05789666544521803, + "grad_norm": 0.4580428245754137, + "learning_rate": 5.769230769230769e-06, + "loss": 0.5344, + "step": 316 + }, + { + "epoch": 0.05807988274093075, + "grad_norm": 0.414647309534276, + "learning_rate": 5.7875457875457885e-06, + "loss": 0.5177, + "step": 317 + }, + { + "epoch": 0.05826310003664346, + "grad_norm": 0.45495854796733415, + "learning_rate": 5.805860805860807e-06, + "loss": 0.5214, + "step": 318 + }, + { + "epoch": 0.058446317332356175, + "grad_norm": 0.47098300998769715, + "learning_rate": 5.824175824175825e-06, + "loss": 0.5519, + "step": 319 + }, + { + "epoch": 0.05862953462806889, + "grad_norm": 0.43335718768494447, + "learning_rate": 5.842490842490844e-06, + "loss": 0.5282, + "step": 320 + }, + { + "epoch": 0.0588127519237816, + "grad_norm": 0.4649559782288625, + "learning_rate": 5.860805860805862e-06, + "loss": 0.5163, + "step": 321 + }, + { + "epoch": 0.05899596921949432, + "grad_norm": 0.44623794630893965, + "learning_rate": 5.8791208791208794e-06, + "loss": 0.514, + "step": 322 + }, + { + "epoch": 0.05917918651520704, + "grad_norm": 0.4521538486948526, + "learning_rate": 5.897435897435898e-06, + "loss": 0.5248, + "step": 323 + }, + { + "epoch": 0.05936240381091975, + "grad_norm": 0.453853173278213, + "learning_rate": 5.915750915750916e-06, + "loss": 0.5105, + "step": 324 + }, + { + "epoch": 0.05954562110663247, + "grad_norm": 0.45615939787874893, + "learning_rate": 5.934065934065935e-06, + "loss": 0.5321, + "step": 325 + }, + { + "epoch": 0.059728838402345184, + "grad_norm": 0.5103507598238842, + "learning_rate": 5.9523809523809525e-06, + "loss": 0.4889, + "step": 326 + }, + { + "epoch": 0.059912055698057895, + "grad_norm": 0.4755868373014201, + "learning_rate": 5.970695970695971e-06, + "loss": 0.5233, + "step": 327 + }, + { + "epoch": 0.06009527299377061, + "grad_norm": 0.48186711326734255, + "learning_rate": 5.989010989010989e-06, + "loss": 0.5216, + "step": 328 + }, + { + "epoch": 0.06027849028948333, + "grad_norm": 0.406290487409726, + "learning_rate": 6.007326007326008e-06, + "loss": 0.5048, + "step": 329 + }, + { + "epoch": 0.06046170758519604, + "grad_norm": 0.47965200235867606, + "learning_rate": 6.025641025641026e-06, + "loss": 0.513, + "step": 330 + }, + { + "epoch": 0.06064492488090876, + "grad_norm": 0.492103246179344, + "learning_rate": 6.043956043956044e-06, + "loss": 0.5553, + "step": 331 + }, + { + "epoch": 0.060828142176621476, + "grad_norm": 0.49171785130479284, + "learning_rate": 6.062271062271062e-06, + "loss": 0.5289, + "step": 332 + }, + { + "epoch": 0.06101135947233419, + "grad_norm": 0.4659535600554528, + "learning_rate": 6.080586080586081e-06, + "loss": 0.5143, + "step": 333 + }, + { + "epoch": 0.061194576768046904, + "grad_norm": 0.43136677550682173, + "learning_rate": 6.0989010989011e-06, + "loss": 0.469, + "step": 334 + }, + { + "epoch": 0.06137779406375962, + "grad_norm": 0.6094192173431031, + "learning_rate": 6.117216117216118e-06, + "loss": 0.5121, + "step": 335 + }, + { + "epoch": 0.06156101135947233, + "grad_norm": 0.45352827270619606, + "learning_rate": 6.135531135531136e-06, + "loss": 0.5097, + "step": 336 + }, + { + "epoch": 0.06174422865518505, + "grad_norm": 0.42626268219917746, + "learning_rate": 6.153846153846155e-06, + "loss": 0.5234, + "step": 337 + }, + { + "epoch": 0.06192744595089777, + "grad_norm": 0.4438479757326601, + "learning_rate": 6.172161172161173e-06, + "loss": 0.4852, + "step": 338 + }, + { + "epoch": 0.06211066324661048, + "grad_norm": 0.688188614062373, + "learning_rate": 6.1904761904761914e-06, + "loss": 0.502, + "step": 339 + }, + { + "epoch": 0.062293880542323196, + "grad_norm": 0.5160627484540854, + "learning_rate": 6.208791208791209e-06, + "loss": 0.521, + "step": 340 + }, + { + "epoch": 0.062477097838035914, + "grad_norm": 0.4356067335955304, + "learning_rate": 6.227106227106228e-06, + "loss": 0.4918, + "step": 341 + }, + { + "epoch": 0.06266031513374863, + "grad_norm": 0.5001791021027777, + "learning_rate": 6.245421245421246e-06, + "loss": 0.5173, + "step": 342 + }, + { + "epoch": 0.06284353242946134, + "grad_norm": 0.45621279623031163, + "learning_rate": 6.2637362637362645e-06, + "loss": 0.5532, + "step": 343 + }, + { + "epoch": 0.06302674972517405, + "grad_norm": 0.4686583276600699, + "learning_rate": 6.282051282051282e-06, + "loss": 0.5544, + "step": 344 + }, + { + "epoch": 0.06320996702088677, + "grad_norm": 0.4500457038114704, + "learning_rate": 6.300366300366301e-06, + "loss": 0.5094, + "step": 345 + }, + { + "epoch": 0.06339318431659949, + "grad_norm": 0.4707435610591325, + "learning_rate": 6.318681318681319e-06, + "loss": 0.5317, + "step": 346 + }, + { + "epoch": 0.0635764016123122, + "grad_norm": 0.44910822533973516, + "learning_rate": 6.336996336996337e-06, + "loss": 0.5063, + "step": 347 + }, + { + "epoch": 0.06375961890802492, + "grad_norm": 0.4262957774336457, + "learning_rate": 6.3553113553113555e-06, + "loss": 0.5389, + "step": 348 + }, + { + "epoch": 0.06394283620373763, + "grad_norm": 0.4958715530578741, + "learning_rate": 6.373626373626373e-06, + "loss": 0.5253, + "step": 349 + }, + { + "epoch": 0.06412605349945034, + "grad_norm": 0.4706503863353741, + "learning_rate": 6.391941391941392e-06, + "loss": 0.5078, + "step": 350 + }, + { + "epoch": 0.06430927079516306, + "grad_norm": 0.5025375698465898, + "learning_rate": 6.410256410256412e-06, + "loss": 0.5366, + "step": 351 + }, + { + "epoch": 0.06449248809087578, + "grad_norm": 0.43307402952194485, + "learning_rate": 6.4285714285714295e-06, + "loss": 0.5053, + "step": 352 + }, + { + "epoch": 0.0646757053865885, + "grad_norm": 0.4145957842766705, + "learning_rate": 6.446886446886448e-06, + "loss": 0.5134, + "step": 353 + }, + { + "epoch": 0.06485892268230121, + "grad_norm": 0.42558146560441634, + "learning_rate": 6.465201465201466e-06, + "loss": 0.498, + "step": 354 + }, + { + "epoch": 0.06504213997801392, + "grad_norm": 0.6808736114735602, + "learning_rate": 6.483516483516485e-06, + "loss": 0.4865, + "step": 355 + }, + { + "epoch": 0.06522535727372664, + "grad_norm": 0.4858578720351213, + "learning_rate": 6.5018315018315026e-06, + "loss": 0.4977, + "step": 356 + }, + { + "epoch": 0.06540857456943935, + "grad_norm": 0.5523209851617275, + "learning_rate": 6.5201465201465204e-06, + "loss": 0.5278, + "step": 357 + }, + { + "epoch": 0.06559179186515207, + "grad_norm": 0.4756243917905379, + "learning_rate": 6.538461538461539e-06, + "loss": 0.5018, + "step": 358 + }, + { + "epoch": 0.06577500916086479, + "grad_norm": 0.5270789359516691, + "learning_rate": 6.556776556776557e-06, + "loss": 0.5127, + "step": 359 + }, + { + "epoch": 0.0659582264565775, + "grad_norm": 0.5213313487503423, + "learning_rate": 6.575091575091576e-06, + "loss": 0.504, + "step": 360 + }, + { + "epoch": 0.06614144375229021, + "grad_norm": 0.47699985237826076, + "learning_rate": 6.5934065934065935e-06, + "loss": 0.5356, + "step": 361 + }, + { + "epoch": 0.06632466104800293, + "grad_norm": 0.4398115760336562, + "learning_rate": 6.611721611721612e-06, + "loss": 0.5038, + "step": 362 + }, + { + "epoch": 0.06650787834371565, + "grad_norm": 0.5848808647247892, + "learning_rate": 6.63003663003663e-06, + "loss": 0.5393, + "step": 363 + }, + { + "epoch": 0.06669109563942836, + "grad_norm": 0.40297449884724584, + "learning_rate": 6.648351648351649e-06, + "loss": 0.5231, + "step": 364 + }, + { + "epoch": 0.06687431293514108, + "grad_norm": 0.4615954062532406, + "learning_rate": 6.666666666666667e-06, + "loss": 0.5017, + "step": 365 + }, + { + "epoch": 0.0670575302308538, + "grad_norm": 0.4204668889117905, + "learning_rate": 6.684981684981685e-06, + "loss": 0.5651, + "step": 366 + }, + { + "epoch": 0.0672407475265665, + "grad_norm": 0.4765525618556606, + "learning_rate": 6.703296703296703e-06, + "loss": 0.5253, + "step": 367 + }, + { + "epoch": 0.06742396482227922, + "grad_norm": 0.431663431335195, + "learning_rate": 6.721611721611723e-06, + "loss": 0.5398, + "step": 368 + }, + { + "epoch": 0.06760718211799194, + "grad_norm": 0.4403652862248983, + "learning_rate": 6.739926739926741e-06, + "loss": 0.5252, + "step": 369 + }, + { + "epoch": 0.06779039941370466, + "grad_norm": 0.483725795531489, + "learning_rate": 6.758241758241759e-06, + "loss": 0.5349, + "step": 370 + }, + { + "epoch": 0.06797361670941737, + "grad_norm": 0.8892216780264888, + "learning_rate": 6.776556776556777e-06, + "loss": 0.5184, + "step": 371 + }, + { + "epoch": 0.06815683400513009, + "grad_norm": 0.48775432933817636, + "learning_rate": 6.794871794871796e-06, + "loss": 0.5275, + "step": 372 + }, + { + "epoch": 0.0683400513008428, + "grad_norm": 0.4772597017011689, + "learning_rate": 6.813186813186814e-06, + "loss": 0.5517, + "step": 373 + }, + { + "epoch": 0.06852326859655551, + "grad_norm": 0.46418441159832313, + "learning_rate": 6.831501831501832e-06, + "loss": 0.5429, + "step": 374 + }, + { + "epoch": 0.06870648589226823, + "grad_norm": 0.45474945807496187, + "learning_rate": 6.84981684981685e-06, + "loss": 0.4999, + "step": 375 + }, + { + "epoch": 0.06888970318798095, + "grad_norm": 0.5576640330118253, + "learning_rate": 6.868131868131869e-06, + "loss": 0.5198, + "step": 376 + }, + { + "epoch": 0.06907292048369366, + "grad_norm": 0.4921371781399045, + "learning_rate": 6.886446886446887e-06, + "loss": 0.5379, + "step": 377 + }, + { + "epoch": 0.06925613777940638, + "grad_norm": 0.41639097386056484, + "learning_rate": 6.9047619047619055e-06, + "loss": 0.519, + "step": 378 + }, + { + "epoch": 0.06943935507511909, + "grad_norm": 0.4778909271586564, + "learning_rate": 6.923076923076923e-06, + "loss": 0.5336, + "step": 379 + }, + { + "epoch": 0.0696225723708318, + "grad_norm": 0.4735414706699025, + "learning_rate": 6.941391941391942e-06, + "loss": 0.5, + "step": 380 + }, + { + "epoch": 0.06980578966654452, + "grad_norm": 0.5646322357431656, + "learning_rate": 6.95970695970696e-06, + "loss": 0.5073, + "step": 381 + }, + { + "epoch": 0.06998900696225724, + "grad_norm": 0.48231469572609614, + "learning_rate": 6.978021978021979e-06, + "loss": 0.519, + "step": 382 + }, + { + "epoch": 0.07017222425796996, + "grad_norm": 0.5051213072897166, + "learning_rate": 6.9963369963369965e-06, + "loss": 0.5315, + "step": 383 + }, + { + "epoch": 0.07035544155368267, + "grad_norm": 0.5166250853237789, + "learning_rate": 7.014652014652014e-06, + "loss": 0.4978, + "step": 384 + }, + { + "epoch": 0.07053865884939538, + "grad_norm": 0.43900982727967125, + "learning_rate": 7.032967032967034e-06, + "loss": 0.4913, + "step": 385 + }, + { + "epoch": 0.0707218761451081, + "grad_norm": 0.5262819691743885, + "learning_rate": 7.051282051282053e-06, + "loss": 0.5189, + "step": 386 + }, + { + "epoch": 0.07090509344082081, + "grad_norm": 0.4906825009780104, + "learning_rate": 7.0695970695970705e-06, + "loss": 0.5586, + "step": 387 + }, + { + "epoch": 0.07108831073653353, + "grad_norm": 0.48825547184252527, + "learning_rate": 7.087912087912089e-06, + "loss": 0.5218, + "step": 388 + }, + { + "epoch": 0.07127152803224625, + "grad_norm": 0.5005437189224704, + "learning_rate": 7.106227106227107e-06, + "loss": 0.5088, + "step": 389 + }, + { + "epoch": 0.07145474532795897, + "grad_norm": 0.4820090004987099, + "learning_rate": 7.124542124542126e-06, + "loss": 0.5132, + "step": 390 + }, + { + "epoch": 0.07163796262367167, + "grad_norm": 0.46585258023856246, + "learning_rate": 7.1428571428571436e-06, + "loss": 0.5198, + "step": 391 + }, + { + "epoch": 0.07182117991938439, + "grad_norm": 0.5351346796257885, + "learning_rate": 7.161172161172162e-06, + "loss": 0.5174, + "step": 392 + }, + { + "epoch": 0.0720043972150971, + "grad_norm": 0.505980784032641, + "learning_rate": 7.17948717948718e-06, + "loss": 0.517, + "step": 393 + }, + { + "epoch": 0.07218761451080982, + "grad_norm": 0.3991493735258601, + "learning_rate": 7.197802197802198e-06, + "loss": 0.5193, + "step": 394 + }, + { + "epoch": 0.07237083180652254, + "grad_norm": 0.47050897190858826, + "learning_rate": 7.216117216117217e-06, + "loss": 0.5278, + "step": 395 + }, + { + "epoch": 0.07255404910223526, + "grad_norm": 0.42709223237684557, + "learning_rate": 7.2344322344322345e-06, + "loss": 0.5233, + "step": 396 + }, + { + "epoch": 0.07273726639794796, + "grad_norm": 0.40701893214503493, + "learning_rate": 7.252747252747253e-06, + "loss": 0.4594, + "step": 397 + }, + { + "epoch": 0.07292048369366068, + "grad_norm": 0.4793723507064191, + "learning_rate": 7.271062271062271e-06, + "loss": 0.5022, + "step": 398 + }, + { + "epoch": 0.0731037009893734, + "grad_norm": 0.4084855491847111, + "learning_rate": 7.28937728937729e-06, + "loss": 0.5177, + "step": 399 + }, + { + "epoch": 0.07328691828508611, + "grad_norm": 0.48791204362548757, + "learning_rate": 7.307692307692308e-06, + "loss": 0.5336, + "step": 400 + }, + { + "epoch": 0.07347013558079883, + "grad_norm": 0.4721917787188149, + "learning_rate": 7.326007326007326e-06, + "loss": 0.5291, + "step": 401 + }, + { + "epoch": 0.07365335287651155, + "grad_norm": 0.4482564665450733, + "learning_rate": 7.344322344322346e-06, + "loss": 0.516, + "step": 402 + }, + { + "epoch": 0.07383657017222425, + "grad_norm": 0.503350703275625, + "learning_rate": 7.362637362637364e-06, + "loss": 0.5232, + "step": 403 + }, + { + "epoch": 0.07401978746793697, + "grad_norm": 0.5152368271826678, + "learning_rate": 7.380952380952382e-06, + "loss": 0.5258, + "step": 404 + }, + { + "epoch": 0.07420300476364969, + "grad_norm": 0.4523690267979874, + "learning_rate": 7.3992673992674e-06, + "loss": 0.5007, + "step": 405 + }, + { + "epoch": 0.0743862220593624, + "grad_norm": 0.44122338596098665, + "learning_rate": 7.417582417582418e-06, + "loss": 0.5325, + "step": 406 + }, + { + "epoch": 0.07456943935507512, + "grad_norm": 0.418021168922302, + "learning_rate": 7.435897435897437e-06, + "loss": 0.5048, + "step": 407 + }, + { + "epoch": 0.07475265665078784, + "grad_norm": 0.44537098554090276, + "learning_rate": 7.454212454212455e-06, + "loss": 0.5294, + "step": 408 + }, + { + "epoch": 0.07493587394650054, + "grad_norm": 0.41384101894590325, + "learning_rate": 7.472527472527473e-06, + "loss": 0.5119, + "step": 409 + }, + { + "epoch": 0.07511909124221326, + "grad_norm": 0.4647470180215214, + "learning_rate": 7.490842490842491e-06, + "loss": 0.5202, + "step": 410 + }, + { + "epoch": 0.07530230853792598, + "grad_norm": 0.4836409216489063, + "learning_rate": 7.50915750915751e-06, + "loss": 0.5056, + "step": 411 + }, + { + "epoch": 0.0754855258336387, + "grad_norm": 0.45195196209865357, + "learning_rate": 7.527472527472528e-06, + "loss": 0.5213, + "step": 412 + }, + { + "epoch": 0.07566874312935142, + "grad_norm": 0.5036829282617328, + "learning_rate": 7.5457875457875465e-06, + "loss": 0.5347, + "step": 413 + }, + { + "epoch": 0.07585196042506413, + "grad_norm": 0.4666038283310089, + "learning_rate": 7.564102564102564e-06, + "loss": 0.5151, + "step": 414 + }, + { + "epoch": 0.07603517772077684, + "grad_norm": 0.5067769344340564, + "learning_rate": 7.582417582417583e-06, + "loss": 0.5171, + "step": 415 + }, + { + "epoch": 0.07621839501648955, + "grad_norm": 0.4093230258713508, + "learning_rate": 7.600732600732601e-06, + "loss": 0.5039, + "step": 416 + }, + { + "epoch": 0.07640161231220227, + "grad_norm": 0.6498006485071868, + "learning_rate": 7.61904761904762e-06, + "loss": 0.5232, + "step": 417 + }, + { + "epoch": 0.07658482960791499, + "grad_norm": 0.4558152414900924, + "learning_rate": 7.637362637362638e-06, + "loss": 0.5232, + "step": 418 + }, + { + "epoch": 0.0767680469036277, + "grad_norm": 0.4804930639066785, + "learning_rate": 7.655677655677656e-06, + "loss": 0.5088, + "step": 419 + }, + { + "epoch": 0.07695126419934042, + "grad_norm": 0.5081120095319347, + "learning_rate": 7.673992673992676e-06, + "loss": 0.5308, + "step": 420 + }, + { + "epoch": 0.07713448149505313, + "grad_norm": 0.42393212923875745, + "learning_rate": 7.692307692307694e-06, + "loss": 0.5084, + "step": 421 + }, + { + "epoch": 0.07731769879076585, + "grad_norm": 0.44672075335716155, + "learning_rate": 7.710622710622711e-06, + "loss": 0.483, + "step": 422 + }, + { + "epoch": 0.07750091608647856, + "grad_norm": 0.4942211324726324, + "learning_rate": 7.72893772893773e-06, + "loss": 0.5236, + "step": 423 + }, + { + "epoch": 0.07768413338219128, + "grad_norm": 0.39817279842841763, + "learning_rate": 7.747252747252749e-06, + "loss": 0.5332, + "step": 424 + }, + { + "epoch": 0.077867350677904, + "grad_norm": 0.4608671155196199, + "learning_rate": 7.765567765567767e-06, + "loss": 0.523, + "step": 425 + }, + { + "epoch": 0.07805056797361672, + "grad_norm": 0.5225991254379231, + "learning_rate": 7.783882783882785e-06, + "loss": 0.5171, + "step": 426 + }, + { + "epoch": 0.07823378526932942, + "grad_norm": 0.5423490196220466, + "learning_rate": 7.802197802197802e-06, + "loss": 0.4985, + "step": 427 + }, + { + "epoch": 0.07841700256504214, + "grad_norm": 0.502419506962013, + "learning_rate": 7.820512820512822e-06, + "loss": 0.5087, + "step": 428 + }, + { + "epoch": 0.07860021986075486, + "grad_norm": 0.5102984182578104, + "learning_rate": 7.83882783882784e-06, + "loss": 0.5358, + "step": 429 + }, + { + "epoch": 0.07878343715646757, + "grad_norm": 0.4742126281062651, + "learning_rate": 7.857142857142858e-06, + "loss": 0.5585, + "step": 430 + }, + { + "epoch": 0.07896665445218029, + "grad_norm": 0.508787409495283, + "learning_rate": 7.875457875457876e-06, + "loss": 0.5411, + "step": 431 + }, + { + "epoch": 0.079149871747893, + "grad_norm": 0.4559258370924022, + "learning_rate": 7.893772893772893e-06, + "loss": 0.5229, + "step": 432 + }, + { + "epoch": 0.07933308904360571, + "grad_norm": 0.6320276655690735, + "learning_rate": 7.912087912087913e-06, + "loss": 0.4688, + "step": 433 + }, + { + "epoch": 0.07951630633931843, + "grad_norm": 0.9797165625621964, + "learning_rate": 7.93040293040293e-06, + "loss": 0.5372, + "step": 434 + }, + { + "epoch": 0.07969952363503115, + "grad_norm": 0.43648526458810954, + "learning_rate": 7.948717948717949e-06, + "loss": 0.4528, + "step": 435 + }, + { + "epoch": 0.07988274093074386, + "grad_norm": 0.46101717931557235, + "learning_rate": 7.967032967032966e-06, + "loss": 0.4836, + "step": 436 + }, + { + "epoch": 0.08006595822645658, + "grad_norm": 0.46019219132505085, + "learning_rate": 7.985347985347986e-06, + "loss": 0.4874, + "step": 437 + }, + { + "epoch": 0.08024917552216929, + "grad_norm": 0.4228905699110607, + "learning_rate": 8.003663003663006e-06, + "loss": 0.483, + "step": 438 + }, + { + "epoch": 0.080432392817882, + "grad_norm": 0.4920935432281776, + "learning_rate": 8.021978021978023e-06, + "loss": 0.489, + "step": 439 + }, + { + "epoch": 0.08061561011359472, + "grad_norm": 0.45385267234753407, + "learning_rate": 8.040293040293041e-06, + "loss": 0.5262, + "step": 440 + }, + { + "epoch": 0.08079882740930744, + "grad_norm": 0.48390990285633756, + "learning_rate": 8.058608058608059e-06, + "loss": 0.4969, + "step": 441 + }, + { + "epoch": 0.08098204470502016, + "grad_norm": 0.4424118373260341, + "learning_rate": 8.076923076923077e-06, + "loss": 0.5217, + "step": 442 + }, + { + "epoch": 0.08116526200073287, + "grad_norm": 0.4647327426734162, + "learning_rate": 8.095238095238097e-06, + "loss": 0.5103, + "step": 443 + }, + { + "epoch": 0.08134847929644558, + "grad_norm": 0.42019112334691044, + "learning_rate": 8.113553113553114e-06, + "loss": 0.5378, + "step": 444 + }, + { + "epoch": 0.0815316965921583, + "grad_norm": 0.40127483204178854, + "learning_rate": 8.131868131868132e-06, + "loss": 0.4813, + "step": 445 + }, + { + "epoch": 0.08171491388787101, + "grad_norm": 0.4606579465800615, + "learning_rate": 8.15018315018315e-06, + "loss": 0.5067, + "step": 446 + }, + { + "epoch": 0.08189813118358373, + "grad_norm": 0.4715535246345655, + "learning_rate": 8.16849816849817e-06, + "loss": 0.5181, + "step": 447 + }, + { + "epoch": 0.08208134847929645, + "grad_norm": 0.5386430386856308, + "learning_rate": 8.186813186813188e-06, + "loss": 0.5188, + "step": 448 + }, + { + "epoch": 0.08226456577500917, + "grad_norm": 0.5389030998275688, + "learning_rate": 8.205128205128205e-06, + "loss": 0.516, + "step": 449 + }, + { + "epoch": 0.08244778307072187, + "grad_norm": 0.5181954163327931, + "learning_rate": 8.223443223443223e-06, + "loss": 0.5099, + "step": 450 + }, + { + "epoch": 0.08263100036643459, + "grad_norm": 0.5425277965290457, + "learning_rate": 8.241758241758243e-06, + "loss": 0.5508, + "step": 451 + }, + { + "epoch": 0.0828142176621473, + "grad_norm": 0.4366152129754919, + "learning_rate": 8.26007326007326e-06, + "loss": 0.4772, + "step": 452 + }, + { + "epoch": 0.08299743495786002, + "grad_norm": 0.45607475441878936, + "learning_rate": 8.278388278388278e-06, + "loss": 0.5319, + "step": 453 + }, + { + "epoch": 0.08318065225357274, + "grad_norm": 0.4712644997156548, + "learning_rate": 8.296703296703298e-06, + "loss": 0.5477, + "step": 454 + }, + { + "epoch": 0.08336386954928546, + "grad_norm": 0.6856852588986854, + "learning_rate": 8.315018315018316e-06, + "loss": 0.5235, + "step": 455 + }, + { + "epoch": 0.08354708684499816, + "grad_norm": 0.5019345602931549, + "learning_rate": 8.333333333333334e-06, + "loss": 0.5101, + "step": 456 + }, + { + "epoch": 0.08373030414071088, + "grad_norm": 0.4155972768165925, + "learning_rate": 8.351648351648353e-06, + "loss": 0.4986, + "step": 457 + }, + { + "epoch": 0.0839135214364236, + "grad_norm": 0.45725631484653695, + "learning_rate": 8.369963369963371e-06, + "loss": 0.5233, + "step": 458 + }, + { + "epoch": 0.08409673873213631, + "grad_norm": 0.5500464841633228, + "learning_rate": 8.388278388278389e-06, + "loss": 0.5225, + "step": 459 + }, + { + "epoch": 0.08427995602784903, + "grad_norm": 0.5210381864086956, + "learning_rate": 8.406593406593407e-06, + "loss": 0.5509, + "step": 460 + }, + { + "epoch": 0.08446317332356175, + "grad_norm": 0.4559821107303101, + "learning_rate": 8.424908424908426e-06, + "loss": 0.5118, + "step": 461 + }, + { + "epoch": 0.08464639061927445, + "grad_norm": 0.48323545775926524, + "learning_rate": 8.443223443223444e-06, + "loss": 0.5049, + "step": 462 + }, + { + "epoch": 0.08482960791498717, + "grad_norm": 0.4812859935236779, + "learning_rate": 8.461538461538462e-06, + "loss": 0.5204, + "step": 463 + }, + { + "epoch": 0.08501282521069989, + "grad_norm": 0.454474341410942, + "learning_rate": 8.47985347985348e-06, + "loss": 0.5387, + "step": 464 + }, + { + "epoch": 0.0851960425064126, + "grad_norm": 0.6615650181808158, + "learning_rate": 8.498168498168498e-06, + "loss": 0.5393, + "step": 465 + }, + { + "epoch": 0.08537925980212532, + "grad_norm": 0.4355325124699567, + "learning_rate": 8.516483516483517e-06, + "loss": 0.5027, + "step": 466 + }, + { + "epoch": 0.08556247709783804, + "grad_norm": 0.43977659231205646, + "learning_rate": 8.534798534798535e-06, + "loss": 0.4851, + "step": 467 + }, + { + "epoch": 0.08574569439355074, + "grad_norm": 0.5490378411527629, + "learning_rate": 8.553113553113553e-06, + "loss": 0.5087, + "step": 468 + }, + { + "epoch": 0.08592891168926346, + "grad_norm": 0.4334113007101785, + "learning_rate": 8.571428571428571e-06, + "loss": 0.4874, + "step": 469 + }, + { + "epoch": 0.08611212898497618, + "grad_norm": 0.48093462345535853, + "learning_rate": 8.58974358974359e-06, + "loss": 0.5103, + "step": 470 + }, + { + "epoch": 0.0862953462806889, + "grad_norm": 0.509104778628553, + "learning_rate": 8.60805860805861e-06, + "loss": 0.5125, + "step": 471 + }, + { + "epoch": 0.08647856357640162, + "grad_norm": 0.46329459466937095, + "learning_rate": 8.626373626373628e-06, + "loss": 0.4862, + "step": 472 + }, + { + "epoch": 0.08666178087211433, + "grad_norm": 0.41221910536829426, + "learning_rate": 8.644688644688646e-06, + "loss": 0.5058, + "step": 473 + }, + { + "epoch": 0.08684499816782704, + "grad_norm": 0.4965957595147397, + "learning_rate": 8.663003663003664e-06, + "loss": 0.5045, + "step": 474 + }, + { + "epoch": 0.08702821546353975, + "grad_norm": 0.6831024596734047, + "learning_rate": 8.681318681318681e-06, + "loss": 0.533, + "step": 475 + }, + { + "epoch": 0.08721143275925247, + "grad_norm": 0.44117689667679433, + "learning_rate": 8.699633699633701e-06, + "loss": 0.494, + "step": 476 + }, + { + "epoch": 0.08739465005496519, + "grad_norm": 0.49552902332130444, + "learning_rate": 8.717948717948719e-06, + "loss": 0.5295, + "step": 477 + }, + { + "epoch": 0.08757786735067791, + "grad_norm": 0.5202864223018397, + "learning_rate": 8.736263736263737e-06, + "loss": 0.5315, + "step": 478 + }, + { + "epoch": 0.08776108464639062, + "grad_norm": 0.4531456591891388, + "learning_rate": 8.754578754578755e-06, + "loss": 0.5177, + "step": 479 + }, + { + "epoch": 0.08794430194210333, + "grad_norm": 0.5060531781359298, + "learning_rate": 8.772893772893774e-06, + "loss": 0.5422, + "step": 480 + }, + { + "epoch": 0.08812751923781605, + "grad_norm": 0.44222377285427994, + "learning_rate": 8.791208791208792e-06, + "loss": 0.4679, + "step": 481 + }, + { + "epoch": 0.08831073653352876, + "grad_norm": 0.4308135949373974, + "learning_rate": 8.80952380952381e-06, + "loss": 0.5057, + "step": 482 + }, + { + "epoch": 0.08849395382924148, + "grad_norm": 0.5107147323097014, + "learning_rate": 8.827838827838828e-06, + "loss": 0.5147, + "step": 483 + }, + { + "epoch": 0.0886771711249542, + "grad_norm": 0.4722359038239213, + "learning_rate": 8.846153846153847e-06, + "loss": 0.5069, + "step": 484 + }, + { + "epoch": 0.08886038842066692, + "grad_norm": 0.4913307139634766, + "learning_rate": 8.864468864468865e-06, + "loss": 0.4943, + "step": 485 + }, + { + "epoch": 0.08904360571637962, + "grad_norm": 0.47381589638250404, + "learning_rate": 8.882783882783883e-06, + "loss": 0.5432, + "step": 486 + }, + { + "epoch": 0.08922682301209234, + "grad_norm": 0.4763727220823097, + "learning_rate": 8.9010989010989e-06, + "loss": 0.4765, + "step": 487 + }, + { + "epoch": 0.08941004030780506, + "grad_norm": 0.48581942940784734, + "learning_rate": 8.91941391941392e-06, + "loss": 0.4918, + "step": 488 + }, + { + "epoch": 0.08959325760351777, + "grad_norm": 0.4771435540271998, + "learning_rate": 8.937728937728938e-06, + "loss": 0.5146, + "step": 489 + }, + { + "epoch": 0.08977647489923049, + "grad_norm": 0.5554558590150881, + "learning_rate": 8.956043956043958e-06, + "loss": 0.5139, + "step": 490 + }, + { + "epoch": 0.08995969219494321, + "grad_norm": 0.4409852140988399, + "learning_rate": 8.974358974358976e-06, + "loss": 0.494, + "step": 491 + }, + { + "epoch": 0.09014290949065591, + "grad_norm": 0.470617109732078, + "learning_rate": 8.992673992673993e-06, + "loss": 0.5293, + "step": 492 + }, + { + "epoch": 0.09032612678636863, + "grad_norm": 0.48719044521659705, + "learning_rate": 9.010989010989011e-06, + "loss": 0.5111, + "step": 493 + }, + { + "epoch": 0.09050934408208135, + "grad_norm": 0.46876439756978366, + "learning_rate": 9.02930402930403e-06, + "loss": 0.4942, + "step": 494 + }, + { + "epoch": 0.09069256137779406, + "grad_norm": 0.46671821534033575, + "learning_rate": 9.047619047619049e-06, + "loss": 0.5617, + "step": 495 + }, + { + "epoch": 0.09087577867350678, + "grad_norm": 0.4868372988815388, + "learning_rate": 9.065934065934067e-06, + "loss": 0.5195, + "step": 496 + }, + { + "epoch": 0.0910589959692195, + "grad_norm": 0.47526685976460153, + "learning_rate": 9.084249084249084e-06, + "loss": 0.4561, + "step": 497 + }, + { + "epoch": 0.0912422132649322, + "grad_norm": 0.526795639153496, + "learning_rate": 9.102564102564104e-06, + "loss": 0.4905, + "step": 498 + }, + { + "epoch": 0.09142543056064492, + "grad_norm": 0.47935527575760656, + "learning_rate": 9.120879120879122e-06, + "loss": 0.5129, + "step": 499 + }, + { + "epoch": 0.09160864785635764, + "grad_norm": 0.47020582683735346, + "learning_rate": 9.13919413919414e-06, + "loss": 0.511, + "step": 500 + }, + { + "epoch": 0.09179186515207036, + "grad_norm": 0.46760420885699033, + "learning_rate": 9.157509157509158e-06, + "loss": 0.4789, + "step": 501 + }, + { + "epoch": 0.09197508244778307, + "grad_norm": 0.4695251283934706, + "learning_rate": 9.175824175824175e-06, + "loss": 0.5351, + "step": 502 + }, + { + "epoch": 0.09215829974349579, + "grad_norm": 0.5153360795669553, + "learning_rate": 9.194139194139195e-06, + "loss": 0.5522, + "step": 503 + }, + { + "epoch": 0.0923415170392085, + "grad_norm": 0.44806081150815513, + "learning_rate": 9.212454212454213e-06, + "loss": 0.5173, + "step": 504 + }, + { + "epoch": 0.09252473433492121, + "grad_norm": 0.451536504838503, + "learning_rate": 9.230769230769232e-06, + "loss": 0.5082, + "step": 505 + }, + { + "epoch": 0.09270795163063393, + "grad_norm": 0.48436384147084766, + "learning_rate": 9.24908424908425e-06, + "loss": 0.5388, + "step": 506 + }, + { + "epoch": 0.09289116892634665, + "grad_norm": 0.4609113225702784, + "learning_rate": 9.267399267399268e-06, + "loss": 0.4892, + "step": 507 + }, + { + "epoch": 0.09307438622205937, + "grad_norm": 0.4098711492197863, + "learning_rate": 9.285714285714288e-06, + "loss": 0.4856, + "step": 508 + }, + { + "epoch": 0.09325760351777208, + "grad_norm": 0.4715675907322317, + "learning_rate": 9.304029304029305e-06, + "loss": 0.53, + "step": 509 + }, + { + "epoch": 0.09344082081348479, + "grad_norm": 0.5026574189450673, + "learning_rate": 9.322344322344323e-06, + "loss": 0.5314, + "step": 510 + }, + { + "epoch": 0.0936240381091975, + "grad_norm": 0.4253199666429807, + "learning_rate": 9.340659340659341e-06, + "loss": 0.5068, + "step": 511 + }, + { + "epoch": 0.09380725540491022, + "grad_norm": 0.45894485399853974, + "learning_rate": 9.358974358974359e-06, + "loss": 0.5235, + "step": 512 + }, + { + "epoch": 0.09399047270062294, + "grad_norm": 0.7513225377427756, + "learning_rate": 9.377289377289379e-06, + "loss": 0.5071, + "step": 513 + }, + { + "epoch": 0.09417368999633566, + "grad_norm": 0.5231017027547264, + "learning_rate": 9.395604395604396e-06, + "loss": 0.5339, + "step": 514 + }, + { + "epoch": 0.09435690729204838, + "grad_norm": 0.5146997140217462, + "learning_rate": 9.413919413919414e-06, + "loss": 0.5052, + "step": 515 + }, + { + "epoch": 0.09454012458776108, + "grad_norm": 0.47760493192836756, + "learning_rate": 9.432234432234432e-06, + "loss": 0.5102, + "step": 516 + }, + { + "epoch": 0.0947233418834738, + "grad_norm": 0.44117590454697736, + "learning_rate": 9.450549450549452e-06, + "loss": 0.5237, + "step": 517 + }, + { + "epoch": 0.09490655917918651, + "grad_norm": 0.5004680185091399, + "learning_rate": 9.46886446886447e-06, + "loss": 0.5243, + "step": 518 + }, + { + "epoch": 0.09508977647489923, + "grad_norm": 0.4549800276149351, + "learning_rate": 9.487179487179487e-06, + "loss": 0.5139, + "step": 519 + }, + { + "epoch": 0.09527299377061195, + "grad_norm": 0.4263327588755031, + "learning_rate": 9.505494505494505e-06, + "loss": 0.4877, + "step": 520 + }, + { + "epoch": 0.09545621106632467, + "grad_norm": 0.4484968491218323, + "learning_rate": 9.523809523809525e-06, + "loss": 0.4986, + "step": 521 + }, + { + "epoch": 0.09563942836203737, + "grad_norm": 0.39390345192807547, + "learning_rate": 9.542124542124543e-06, + "loss": 0.5116, + "step": 522 + }, + { + "epoch": 0.09582264565775009, + "grad_norm": 0.4529147798983982, + "learning_rate": 9.560439560439562e-06, + "loss": 0.5312, + "step": 523 + }, + { + "epoch": 0.0960058629534628, + "grad_norm": 0.49701803503982106, + "learning_rate": 9.57875457875458e-06, + "loss": 0.529, + "step": 524 + }, + { + "epoch": 0.09618908024917552, + "grad_norm": 0.4097214793479841, + "learning_rate": 9.597069597069598e-06, + "loss": 0.5005, + "step": 525 + }, + { + "epoch": 0.09637229754488824, + "grad_norm": 0.5148151884655592, + "learning_rate": 9.615384615384616e-06, + "loss": 0.53, + "step": 526 + }, + { + "epoch": 0.09655551484060096, + "grad_norm": 0.4264857545219357, + "learning_rate": 9.633699633699635e-06, + "loss": 0.5136, + "step": 527 + }, + { + "epoch": 0.09673873213631366, + "grad_norm": 0.45426464252638443, + "learning_rate": 9.652014652014653e-06, + "loss": 0.5117, + "step": 528 + }, + { + "epoch": 0.09692194943202638, + "grad_norm": 0.47571176071645493, + "learning_rate": 9.670329670329671e-06, + "loss": 0.5185, + "step": 529 + }, + { + "epoch": 0.0971051667277391, + "grad_norm": 0.42418752426373113, + "learning_rate": 9.688644688644689e-06, + "loss": 0.5285, + "step": 530 + }, + { + "epoch": 0.09728838402345182, + "grad_norm": 0.48590348616099827, + "learning_rate": 9.706959706959708e-06, + "loss": 0.5339, + "step": 531 + }, + { + "epoch": 0.09747160131916453, + "grad_norm": 0.463774513888958, + "learning_rate": 9.725274725274726e-06, + "loss": 0.5386, + "step": 532 + }, + { + "epoch": 0.09765481861487725, + "grad_norm": 0.4891808396070614, + "learning_rate": 9.743589743589744e-06, + "loss": 0.5326, + "step": 533 + }, + { + "epoch": 0.09783803591058995, + "grad_norm": 0.40841004082693305, + "learning_rate": 9.761904761904762e-06, + "loss": 0.5085, + "step": 534 + }, + { + "epoch": 0.09802125320630267, + "grad_norm": 0.42494728872814735, + "learning_rate": 9.780219780219781e-06, + "loss": 0.4751, + "step": 535 + }, + { + "epoch": 0.09820447050201539, + "grad_norm": 0.4594283822364781, + "learning_rate": 9.7985347985348e-06, + "loss": 0.5275, + "step": 536 + }, + { + "epoch": 0.09838768779772811, + "grad_norm": 0.4064302259117676, + "learning_rate": 9.816849816849817e-06, + "loss": 0.5089, + "step": 537 + }, + { + "epoch": 0.09857090509344082, + "grad_norm": 0.4860027724973198, + "learning_rate": 9.835164835164835e-06, + "loss": 0.505, + "step": 538 + }, + { + "epoch": 0.09875412238915354, + "grad_norm": 0.47625362870716265, + "learning_rate": 9.853479853479855e-06, + "loss": 0.4954, + "step": 539 + }, + { + "epoch": 0.09893733968486625, + "grad_norm": 0.4592940212317861, + "learning_rate": 9.871794871794872e-06, + "loss": 0.4871, + "step": 540 + }, + { + "epoch": 0.09912055698057896, + "grad_norm": 0.42717127053329573, + "learning_rate": 9.890109890109892e-06, + "loss": 0.5261, + "step": 541 + }, + { + "epoch": 0.09930377427629168, + "grad_norm": 0.43970535809042904, + "learning_rate": 9.90842490842491e-06, + "loss": 0.4866, + "step": 542 + }, + { + "epoch": 0.0994869915720044, + "grad_norm": 0.42143360797263807, + "learning_rate": 9.926739926739928e-06, + "loss": 0.5113, + "step": 543 + }, + { + "epoch": 0.09967020886771712, + "grad_norm": 0.4416711016318173, + "learning_rate": 9.945054945054946e-06, + "loss": 0.5116, + "step": 544 + }, + { + "epoch": 0.09985342616342983, + "grad_norm": 0.45325267149265236, + "learning_rate": 9.963369963369965e-06, + "loss": 0.5283, + "step": 545 + }, + { + "epoch": 0.10003664345914254, + "grad_norm": 0.3948707773420248, + "learning_rate": 9.981684981684983e-06, + "loss": 0.5114, + "step": 546 + }, + { + "epoch": 0.10021986075485526, + "grad_norm": 0.42983426021717475, + "learning_rate": 1e-05, + "loss": 0.5169, + "step": 547 + }, + { + "epoch": 0.10040307805056797, + "grad_norm": 0.5119731461361727, + "learning_rate": 9.999998977359419e-06, + "loss": 0.518, + "step": 548 + }, + { + "epoch": 0.10058629534628069, + "grad_norm": 0.48479775231872485, + "learning_rate": 9.999995909438092e-06, + "loss": 0.5207, + "step": 549 + }, + { + "epoch": 0.10076951264199341, + "grad_norm": 0.45005420614602115, + "learning_rate": 9.999990796237274e-06, + "loss": 0.4878, + "step": 550 + }, + { + "epoch": 0.10095272993770613, + "grad_norm": 0.40595587220813967, + "learning_rate": 9.999983637759059e-06, + "loss": 0.4712, + "step": 551 + }, + { + "epoch": 0.10113594723341883, + "grad_norm": 0.43893870910937327, + "learning_rate": 9.999974434006372e-06, + "loss": 0.4964, + "step": 552 + }, + { + "epoch": 0.10131916452913155, + "grad_norm": 0.4638079942473788, + "learning_rate": 9.99996318498298e-06, + "loss": 0.4834, + "step": 553 + }, + { + "epoch": 0.10150238182484426, + "grad_norm": 0.45078724585599195, + "learning_rate": 9.999949890693484e-06, + "loss": 0.4828, + "step": 554 + }, + { + "epoch": 0.10168559912055698, + "grad_norm": 0.4968230440871285, + "learning_rate": 9.999934551143319e-06, + "loss": 0.5146, + "step": 555 + }, + { + "epoch": 0.1018688164162697, + "grad_norm": 0.4431824586567993, + "learning_rate": 9.999917166338767e-06, + "loss": 0.5117, + "step": 556 + }, + { + "epoch": 0.10205203371198242, + "grad_norm": 0.40809319706771474, + "learning_rate": 9.999897736286932e-06, + "loss": 0.4942, + "step": 557 + }, + { + "epoch": 0.10223525100769512, + "grad_norm": 0.46904649406632837, + "learning_rate": 9.999876260995767e-06, + "loss": 0.5223, + "step": 558 + }, + { + "epoch": 0.10241846830340784, + "grad_norm": 0.45993616230767786, + "learning_rate": 9.999852740474054e-06, + "loss": 0.498, + "step": 559 + }, + { + "epoch": 0.10260168559912056, + "grad_norm": 0.41328723379038895, + "learning_rate": 9.999827174731414e-06, + "loss": 0.5294, + "step": 560 + }, + { + "epoch": 0.10278490289483327, + "grad_norm": 0.4584774156426854, + "learning_rate": 9.999799563778307e-06, + "loss": 0.5138, + "step": 561 + }, + { + "epoch": 0.10296812019054599, + "grad_norm": 0.4626480547701558, + "learning_rate": 9.999769907626024e-06, + "loss": 0.5122, + "step": 562 + }, + { + "epoch": 0.10315133748625871, + "grad_norm": 0.4647555669122866, + "learning_rate": 9.9997382062867e-06, + "loss": 0.5276, + "step": 563 + }, + { + "epoch": 0.10333455478197141, + "grad_norm": 0.41824885855037686, + "learning_rate": 9.9997044597733e-06, + "loss": 0.4976, + "step": 564 + }, + { + "epoch": 0.10351777207768413, + "grad_norm": 0.41187305924650414, + "learning_rate": 9.999668668099628e-06, + "loss": 0.4953, + "step": 565 + }, + { + "epoch": 0.10370098937339685, + "grad_norm": 0.42660595756816655, + "learning_rate": 9.999630831280329e-06, + "loss": 0.514, + "step": 566 + }, + { + "epoch": 0.10388420666910957, + "grad_norm": 0.5099786508796531, + "learning_rate": 9.999590949330876e-06, + "loss": 0.5038, + "step": 567 + }, + { + "epoch": 0.10406742396482228, + "grad_norm": 0.4926825490754442, + "learning_rate": 9.999549022267582e-06, + "loss": 0.5535, + "step": 568 + }, + { + "epoch": 0.104250641260535, + "grad_norm": 0.4578490443823955, + "learning_rate": 9.999505050107601e-06, + "loss": 0.5075, + "step": 569 + }, + { + "epoch": 0.1044338585562477, + "grad_norm": 0.4733323605035433, + "learning_rate": 9.99945903286892e-06, + "loss": 0.4747, + "step": 570 + }, + { + "epoch": 0.10461707585196042, + "grad_norm": 0.44893796479946446, + "learning_rate": 9.999410970570358e-06, + "loss": 0.4711, + "step": 571 + }, + { + "epoch": 0.10480029314767314, + "grad_norm": 0.5254914900315981, + "learning_rate": 9.99936086323158e-06, + "loss": 0.5358, + "step": 572 + }, + { + "epoch": 0.10498351044338586, + "grad_norm": 0.537923540220477, + "learning_rate": 9.999308710873083e-06, + "loss": 0.5079, + "step": 573 + }, + { + "epoch": 0.10516672773909858, + "grad_norm": 0.45536567675754, + "learning_rate": 9.999254513516196e-06, + "loss": 0.4908, + "step": 574 + }, + { + "epoch": 0.10534994503481128, + "grad_norm": 0.38323869490085266, + "learning_rate": 9.999198271183094e-06, + "loss": 0.4983, + "step": 575 + }, + { + "epoch": 0.105533162330524, + "grad_norm": 0.43580085679521197, + "learning_rate": 9.999139983896779e-06, + "loss": 0.5031, + "step": 576 + }, + { + "epoch": 0.10571637962623671, + "grad_norm": 0.4193926406628939, + "learning_rate": 9.999079651681096e-06, + "loss": 0.5026, + "step": 577 + }, + { + "epoch": 0.10589959692194943, + "grad_norm": 0.5149327291476714, + "learning_rate": 9.999017274560722e-06, + "loss": 0.5414, + "step": 578 + }, + { + "epoch": 0.10608281421766215, + "grad_norm": 0.44098660936462153, + "learning_rate": 9.998952852561176e-06, + "loss": 0.5032, + "step": 579 + }, + { + "epoch": 0.10626603151337487, + "grad_norm": 0.4413320145497695, + "learning_rate": 9.998886385708807e-06, + "loss": 0.497, + "step": 580 + }, + { + "epoch": 0.10644924880908757, + "grad_norm": 0.45575842795199323, + "learning_rate": 9.998817874030808e-06, + "loss": 0.5207, + "step": 581 + }, + { + "epoch": 0.10663246610480029, + "grad_norm": 0.532654166955093, + "learning_rate": 9.9987473175552e-06, + "loss": 0.5212, + "step": 582 + }, + { + "epoch": 0.106815683400513, + "grad_norm": 0.4157619606008366, + "learning_rate": 9.998674716310846e-06, + "loss": 0.5085, + "step": 583 + }, + { + "epoch": 0.10699890069622572, + "grad_norm": 0.4868388397165033, + "learning_rate": 9.998600070327444e-06, + "loss": 0.5306, + "step": 584 + }, + { + "epoch": 0.10718211799193844, + "grad_norm": 0.49401937562080056, + "learning_rate": 9.998523379635527e-06, + "loss": 0.5429, + "step": 585 + }, + { + "epoch": 0.10736533528765116, + "grad_norm": 0.3986626914030523, + "learning_rate": 9.99844464426647e-06, + "loss": 0.4695, + "step": 586 + }, + { + "epoch": 0.10754855258336386, + "grad_norm": 0.49524254606268836, + "learning_rate": 9.998363864252474e-06, + "loss": 0.5226, + "step": 587 + }, + { + "epoch": 0.10773176987907658, + "grad_norm": 0.46186447850660334, + "learning_rate": 9.998281039626588e-06, + "loss": 0.5294, + "step": 588 + }, + { + "epoch": 0.1079149871747893, + "grad_norm": 0.4207082230286084, + "learning_rate": 9.99819617042269e-06, + "loss": 0.5096, + "step": 589 + }, + { + "epoch": 0.10809820447050202, + "grad_norm": 0.3955565788721964, + "learning_rate": 9.998109256675496e-06, + "loss": 0.4931, + "step": 590 + }, + { + "epoch": 0.10828142176621473, + "grad_norm": 0.36723747707336185, + "learning_rate": 9.998020298420559e-06, + "loss": 0.4964, + "step": 591 + }, + { + "epoch": 0.10846463906192745, + "grad_norm": 0.4185544526751163, + "learning_rate": 9.997929295694266e-06, + "loss": 0.4801, + "step": 592 + }, + { + "epoch": 0.10864785635764015, + "grad_norm": 0.42948593775837285, + "learning_rate": 9.997836248533845e-06, + "loss": 0.5197, + "step": 593 + }, + { + "epoch": 0.10883107365335287, + "grad_norm": 0.43966759837587943, + "learning_rate": 9.997741156977356e-06, + "loss": 0.5254, + "step": 594 + }, + { + "epoch": 0.10901429094906559, + "grad_norm": 0.4632445570564666, + "learning_rate": 9.997644021063698e-06, + "loss": 0.4802, + "step": 595 + }, + { + "epoch": 0.10919750824477831, + "grad_norm": 0.46480621716730125, + "learning_rate": 9.997544840832604e-06, + "loss": 0.5219, + "step": 596 + }, + { + "epoch": 0.10938072554049103, + "grad_norm": 0.44567173311984926, + "learning_rate": 9.997443616324645e-06, + "loss": 0.5105, + "step": 597 + }, + { + "epoch": 0.10956394283620374, + "grad_norm": 0.4534263401515789, + "learning_rate": 9.997340347581226e-06, + "loss": 0.5212, + "step": 598 + }, + { + "epoch": 0.10974716013191645, + "grad_norm": 0.40359525147680225, + "learning_rate": 9.99723503464459e-06, + "loss": 0.5261, + "step": 599 + }, + { + "epoch": 0.10993037742762916, + "grad_norm": 0.8305073137318505, + "learning_rate": 9.99712767755782e-06, + "loss": 0.5173, + "step": 600 + }, + { + "epoch": 0.11011359472334188, + "grad_norm": 0.5039765167575108, + "learning_rate": 9.997018276364825e-06, + "loss": 0.4983, + "step": 601 + }, + { + "epoch": 0.1102968120190546, + "grad_norm": 0.4000444070826362, + "learning_rate": 9.99690683111036e-06, + "loss": 0.4876, + "step": 602 + }, + { + "epoch": 0.11048002931476732, + "grad_norm": 0.45535979366815893, + "learning_rate": 9.99679334184001e-06, + "loss": 0.5116, + "step": 603 + }, + { + "epoch": 0.11066324661048003, + "grad_norm": 0.5208092576973751, + "learning_rate": 9.996677808600202e-06, + "loss": 0.5079, + "step": 604 + }, + { + "epoch": 0.11084646390619274, + "grad_norm": 0.47721138790862244, + "learning_rate": 9.996560231438193e-06, + "loss": 0.5031, + "step": 605 + }, + { + "epoch": 0.11102968120190546, + "grad_norm": 0.4106250945716929, + "learning_rate": 9.996440610402078e-06, + "loss": 0.4789, + "step": 606 + }, + { + "epoch": 0.11121289849761817, + "grad_norm": 0.4981443340009068, + "learning_rate": 9.996318945540792e-06, + "loss": 0.4749, + "step": 607 + }, + { + "epoch": 0.11139611579333089, + "grad_norm": 0.4365748987869686, + "learning_rate": 9.996195236904098e-06, + "loss": 0.503, + "step": 608 + }, + { + "epoch": 0.11157933308904361, + "grad_norm": 0.42769357659836243, + "learning_rate": 9.996069484542605e-06, + "loss": 0.4883, + "step": 609 + }, + { + "epoch": 0.11176255038475633, + "grad_norm": 0.4468425715852285, + "learning_rate": 9.995941688507749e-06, + "loss": 0.528, + "step": 610 + }, + { + "epoch": 0.11194576768046903, + "grad_norm": 0.44305501127316244, + "learning_rate": 9.995811848851807e-06, + "loss": 0.5192, + "step": 611 + }, + { + "epoch": 0.11212898497618175, + "grad_norm": 0.4430509247761743, + "learning_rate": 9.995679965627891e-06, + "loss": 0.4879, + "step": 612 + }, + { + "epoch": 0.11231220227189447, + "grad_norm": 0.4330736232909904, + "learning_rate": 9.995546038889948e-06, + "loss": 0.5315, + "step": 613 + }, + { + "epoch": 0.11249541956760718, + "grad_norm": 0.38536887827486094, + "learning_rate": 9.995410068692763e-06, + "loss": 0.4789, + "step": 614 + }, + { + "epoch": 0.1126786368633199, + "grad_norm": 0.5420414741691238, + "learning_rate": 9.995272055091954e-06, + "loss": 0.4863, + "step": 615 + }, + { + "epoch": 0.11286185415903262, + "grad_norm": 0.4166977498062619, + "learning_rate": 9.995131998143976e-06, + "loss": 0.4863, + "step": 616 + }, + { + "epoch": 0.11304507145474532, + "grad_norm": 0.40681536551456327, + "learning_rate": 9.99498989790612e-06, + "loss": 0.5158, + "step": 617 + }, + { + "epoch": 0.11322828875045804, + "grad_norm": 0.4164498167643854, + "learning_rate": 9.994845754436516e-06, + "loss": 0.5287, + "step": 618 + }, + { + "epoch": 0.11341150604617076, + "grad_norm": 0.430875770675555, + "learning_rate": 9.994699567794123e-06, + "loss": 0.4949, + "step": 619 + }, + { + "epoch": 0.11359472334188347, + "grad_norm": 0.47209089186963143, + "learning_rate": 9.994551338038742e-06, + "loss": 0.5212, + "step": 620 + }, + { + "epoch": 0.11377794063759619, + "grad_norm": 0.4693460118731517, + "learning_rate": 9.994401065231008e-06, + "loss": 0.5218, + "step": 621 + }, + { + "epoch": 0.11396115793330891, + "grad_norm": 0.43645941737097255, + "learning_rate": 9.994248749432388e-06, + "loss": 0.5355, + "step": 622 + }, + { + "epoch": 0.11414437522902161, + "grad_norm": 0.45174829183811943, + "learning_rate": 9.994094390705189e-06, + "loss": 0.4941, + "step": 623 + }, + { + "epoch": 0.11432759252473433, + "grad_norm": 0.4609885421062077, + "learning_rate": 9.993937989112554e-06, + "loss": 0.5424, + "step": 624 + }, + { + "epoch": 0.11451080982044705, + "grad_norm": 0.4711205796072519, + "learning_rate": 9.993779544718459e-06, + "loss": 0.5054, + "step": 625 + }, + { + "epoch": 0.11469402711615977, + "grad_norm": 0.4692458224302816, + "learning_rate": 9.993619057587714e-06, + "loss": 0.5013, + "step": 626 + }, + { + "epoch": 0.11487724441187248, + "grad_norm": 0.44543459536907126, + "learning_rate": 9.99345652778597e-06, + "loss": 0.4882, + "step": 627 + }, + { + "epoch": 0.1150604617075852, + "grad_norm": 0.43461884283611496, + "learning_rate": 9.993291955379713e-06, + "loss": 0.5352, + "step": 628 + }, + { + "epoch": 0.1152436790032979, + "grad_norm": 0.41500076390959145, + "learning_rate": 9.993125340436258e-06, + "loss": 0.4895, + "step": 629 + }, + { + "epoch": 0.11542689629901062, + "grad_norm": 0.4307377636082482, + "learning_rate": 9.992956683023762e-06, + "loss": 0.5149, + "step": 630 + }, + { + "epoch": 0.11561011359472334, + "grad_norm": 0.45085961488644083, + "learning_rate": 9.992785983211214e-06, + "loss": 0.5061, + "step": 631 + }, + { + "epoch": 0.11579333089043606, + "grad_norm": 0.45702457348234676, + "learning_rate": 9.992613241068444e-06, + "loss": 0.4947, + "step": 632 + }, + { + "epoch": 0.11597654818614878, + "grad_norm": 0.46046219980195896, + "learning_rate": 9.992438456666108e-06, + "loss": 0.4727, + "step": 633 + }, + { + "epoch": 0.1161597654818615, + "grad_norm": 0.5109084834664002, + "learning_rate": 9.992261630075704e-06, + "loss": 0.5015, + "step": 634 + }, + { + "epoch": 0.1163429827775742, + "grad_norm": 0.4605028589793491, + "learning_rate": 9.992082761369567e-06, + "loss": 0.518, + "step": 635 + }, + { + "epoch": 0.11652620007328691, + "grad_norm": 0.4135254020313275, + "learning_rate": 9.991901850620861e-06, + "loss": 0.4833, + "step": 636 + }, + { + "epoch": 0.11670941736899963, + "grad_norm": 0.5226822189268372, + "learning_rate": 9.99171889790359e-06, + "loss": 0.5118, + "step": 637 + }, + { + "epoch": 0.11689263466471235, + "grad_norm": 0.4061164579366039, + "learning_rate": 9.991533903292592e-06, + "loss": 0.4953, + "step": 638 + }, + { + "epoch": 0.11707585196042507, + "grad_norm": 2.3765281557608486, + "learning_rate": 9.99134686686354e-06, + "loss": 0.4708, + "step": 639 + }, + { + "epoch": 0.11725906925613779, + "grad_norm": 0.4675538148921928, + "learning_rate": 9.991157788692942e-06, + "loss": 0.4994, + "step": 640 + }, + { + "epoch": 0.11744228655185049, + "grad_norm": 0.4467597960418954, + "learning_rate": 9.990966668858144e-06, + "loss": 0.4982, + "step": 641 + }, + { + "epoch": 0.1176255038475632, + "grad_norm": 0.46222076234046117, + "learning_rate": 9.99077350743732e-06, + "loss": 0.5208, + "step": 642 + }, + { + "epoch": 0.11780872114327592, + "grad_norm": 0.45856899421216063, + "learning_rate": 9.990578304509488e-06, + "loss": 0.5109, + "step": 643 + }, + { + "epoch": 0.11799193843898864, + "grad_norm": 0.5006843127431095, + "learning_rate": 9.990381060154496e-06, + "loss": 0.4941, + "step": 644 + }, + { + "epoch": 0.11817515573470136, + "grad_norm": 0.4638055815431298, + "learning_rate": 9.990181774453028e-06, + "loss": 0.509, + "step": 645 + }, + { + "epoch": 0.11835837303041408, + "grad_norm": 0.4429725714841303, + "learning_rate": 9.989980447486601e-06, + "loss": 0.5196, + "step": 646 + }, + { + "epoch": 0.11854159032612678, + "grad_norm": 0.44051330732830757, + "learning_rate": 9.989777079337572e-06, + "loss": 0.5316, + "step": 647 + }, + { + "epoch": 0.1187248076218395, + "grad_norm": 0.41555588890083767, + "learning_rate": 9.989571670089129e-06, + "loss": 0.5126, + "step": 648 + }, + { + "epoch": 0.11890802491755222, + "grad_norm": 0.46239837428682623, + "learning_rate": 9.989364219825295e-06, + "loss": 0.506, + "step": 649 + }, + { + "epoch": 0.11909124221326493, + "grad_norm": 0.416093035128269, + "learning_rate": 9.98915472863093e-06, + "loss": 0.5194, + "step": 650 + }, + { + "epoch": 0.11927445950897765, + "grad_norm": 0.43847114193541153, + "learning_rate": 9.988943196591727e-06, + "loss": 0.4994, + "step": 651 + }, + { + "epoch": 0.11945767680469037, + "grad_norm": 0.44451525143619486, + "learning_rate": 9.988729623794215e-06, + "loss": 0.5283, + "step": 652 + }, + { + "epoch": 0.11964089410040307, + "grad_norm": 0.4041297930697904, + "learning_rate": 9.988514010325758e-06, + "loss": 0.5049, + "step": 653 + }, + { + "epoch": 0.11982411139611579, + "grad_norm": 0.4075716530480146, + "learning_rate": 9.988296356274551e-06, + "loss": 0.4921, + "step": 654 + }, + { + "epoch": 0.12000732869182851, + "grad_norm": 0.40736195443308915, + "learning_rate": 9.988076661729631e-06, + "loss": 0.4805, + "step": 655 + }, + { + "epoch": 0.12019054598754123, + "grad_norm": 0.4545077167148167, + "learning_rate": 9.987854926780863e-06, + "loss": 0.4789, + "step": 656 + }, + { + "epoch": 0.12037376328325394, + "grad_norm": 0.5314832542562551, + "learning_rate": 9.987631151518948e-06, + "loss": 0.5263, + "step": 657 + }, + { + "epoch": 0.12055698057896666, + "grad_norm": 0.4967524623940482, + "learning_rate": 9.987405336035425e-06, + "loss": 0.5106, + "step": 658 + }, + { + "epoch": 0.12074019787467936, + "grad_norm": 0.4306664623952609, + "learning_rate": 9.987177480422663e-06, + "loss": 0.5376, + "step": 659 + }, + { + "epoch": 0.12092341517039208, + "grad_norm": 0.4158801663482348, + "learning_rate": 9.98694758477387e-06, + "loss": 0.4838, + "step": 660 + }, + { + "epoch": 0.1211066324661048, + "grad_norm": 0.40605820771792417, + "learning_rate": 9.986715649183084e-06, + "loss": 0.505, + "step": 661 + }, + { + "epoch": 0.12128984976181752, + "grad_norm": 0.4085721106839356, + "learning_rate": 9.986481673745183e-06, + "loss": 0.4993, + "step": 662 + }, + { + "epoch": 0.12147306705753023, + "grad_norm": 0.4785768918832099, + "learning_rate": 9.986245658555873e-06, + "loss": 0.5387, + "step": 663 + }, + { + "epoch": 0.12165628435324295, + "grad_norm": 0.4175528791505364, + "learning_rate": 9.986007603711698e-06, + "loss": 0.4976, + "step": 664 + }, + { + "epoch": 0.12183950164895566, + "grad_norm": 0.4323331836919794, + "learning_rate": 9.985767509310035e-06, + "loss": 0.5346, + "step": 665 + }, + { + "epoch": 0.12202271894466837, + "grad_norm": 0.4868011056584651, + "learning_rate": 9.9855253754491e-06, + "loss": 0.5309, + "step": 666 + }, + { + "epoch": 0.12220593624038109, + "grad_norm": 0.46372133085485157, + "learning_rate": 9.985281202227936e-06, + "loss": 0.5053, + "step": 667 + }, + { + "epoch": 0.12238915353609381, + "grad_norm": 0.48326269099644514, + "learning_rate": 9.985034989746423e-06, + "loss": 0.4941, + "step": 668 + }, + { + "epoch": 0.12257237083180653, + "grad_norm": 0.4557552947362274, + "learning_rate": 9.984786738105279e-06, + "loss": 0.5121, + "step": 669 + }, + { + "epoch": 0.12275558812751924, + "grad_norm": 0.39704238565295197, + "learning_rate": 9.98453644740605e-06, + "loss": 0.4962, + "step": 670 + }, + { + "epoch": 0.12293880542323195, + "grad_norm": 0.40645243334504044, + "learning_rate": 9.98428411775112e-06, + "loss": 0.5046, + "step": 671 + }, + { + "epoch": 0.12312202271894467, + "grad_norm": 0.42629021258457467, + "learning_rate": 9.984029749243707e-06, + "loss": 0.5084, + "step": 672 + }, + { + "epoch": 0.12330524001465738, + "grad_norm": 0.45965819318406503, + "learning_rate": 9.98377334198786e-06, + "loss": 0.4759, + "step": 673 + }, + { + "epoch": 0.1234884573103701, + "grad_norm": 0.41967629717267835, + "learning_rate": 9.983514896088466e-06, + "loss": 0.498, + "step": 674 + }, + { + "epoch": 0.12367167460608282, + "grad_norm": 0.4295939596150219, + "learning_rate": 9.983254411651242e-06, + "loss": 0.5243, + "step": 675 + }, + { + "epoch": 0.12385489190179554, + "grad_norm": 0.4808089074794197, + "learning_rate": 9.982991888782742e-06, + "loss": 0.5, + "step": 676 + }, + { + "epoch": 0.12403810919750824, + "grad_norm": 0.48728725704379316, + "learning_rate": 9.982727327590352e-06, + "loss": 0.5459, + "step": 677 + }, + { + "epoch": 0.12422132649322096, + "grad_norm": 0.42799437772375803, + "learning_rate": 9.982460728182292e-06, + "loss": 0.5488, + "step": 678 + }, + { + "epoch": 0.12440454378893367, + "grad_norm": 0.4334349975707624, + "learning_rate": 9.982192090667618e-06, + "loss": 0.501, + "step": 679 + }, + { + "epoch": 0.12458776108464639, + "grad_norm": 0.45671378695636383, + "learning_rate": 9.981921415156217e-06, + "loss": 0.5236, + "step": 680 + }, + { + "epoch": 0.12477097838035911, + "grad_norm": 0.49531406109254905, + "learning_rate": 9.98164870175881e-06, + "loss": 0.5224, + "step": 681 + }, + { + "epoch": 0.12495419567607183, + "grad_norm": 0.3902067301024602, + "learning_rate": 9.981373950586952e-06, + "loss": 0.4835, + "step": 682 + }, + { + "epoch": 0.12513741297178455, + "grad_norm": 0.46995820446689096, + "learning_rate": 9.981097161753032e-06, + "loss": 0.527, + "step": 683 + }, + { + "epoch": 0.12532063026749726, + "grad_norm": 0.43642001976560174, + "learning_rate": 9.980818335370273e-06, + "loss": 0.514, + "step": 684 + }, + { + "epoch": 0.12550384756320998, + "grad_norm": 0.5074521053811396, + "learning_rate": 9.980537471552728e-06, + "loss": 0.4897, + "step": 685 + }, + { + "epoch": 0.12568706485892267, + "grad_norm": 0.43625566798430576, + "learning_rate": 9.98025457041529e-06, + "loss": 0.5174, + "step": 686 + }, + { + "epoch": 0.1258702821546354, + "grad_norm": 0.4129578447344159, + "learning_rate": 9.979969632073678e-06, + "loss": 0.5135, + "step": 687 + }, + { + "epoch": 0.1260534994503481, + "grad_norm": 0.4553566655837448, + "learning_rate": 9.97968265664445e-06, + "loss": 0.5228, + "step": 688 + }, + { + "epoch": 0.12623671674606082, + "grad_norm": 0.4682536997157449, + "learning_rate": 9.979393644244992e-06, + "loss": 0.5039, + "step": 689 + }, + { + "epoch": 0.12641993404177354, + "grad_norm": 0.42839595687214543, + "learning_rate": 9.979102594993533e-06, + "loss": 0.5128, + "step": 690 + }, + { + "epoch": 0.12660315133748626, + "grad_norm": 0.46092632925079513, + "learning_rate": 9.978809509009121e-06, + "loss": 0.5011, + "step": 691 + }, + { + "epoch": 0.12678636863319898, + "grad_norm": 0.4486176186332823, + "learning_rate": 9.97851438641165e-06, + "loss": 0.5169, + "step": 692 + }, + { + "epoch": 0.1269695859289117, + "grad_norm": 0.47419826995019465, + "learning_rate": 9.978217227321837e-06, + "loss": 0.5131, + "step": 693 + }, + { + "epoch": 0.1271528032246244, + "grad_norm": 0.4004091408746546, + "learning_rate": 9.97791803186124e-06, + "loss": 0.4743, + "step": 694 + }, + { + "epoch": 0.12733602052033713, + "grad_norm": 0.38849935615066844, + "learning_rate": 9.977616800152248e-06, + "loss": 0.4944, + "step": 695 + }, + { + "epoch": 0.12751923781604985, + "grad_norm": 0.44216099670793213, + "learning_rate": 9.977313532318078e-06, + "loss": 0.5072, + "step": 696 + }, + { + "epoch": 0.12770245511176256, + "grad_norm": 0.4462646919434802, + "learning_rate": 9.977008228482785e-06, + "loss": 0.4901, + "step": 697 + }, + { + "epoch": 0.12788567240747525, + "grad_norm": 0.44769942596462214, + "learning_rate": 9.976700888771259e-06, + "loss": 0.5263, + "step": 698 + }, + { + "epoch": 0.12806888970318797, + "grad_norm": 0.5113829977752261, + "learning_rate": 9.976391513309212e-06, + "loss": 0.5202, + "step": 699 + }, + { + "epoch": 0.1282521069989007, + "grad_norm": 0.4245344457429005, + "learning_rate": 9.976080102223202e-06, + "loss": 0.5182, + "step": 700 + }, + { + "epoch": 0.1284353242946134, + "grad_norm": 0.4305381223047699, + "learning_rate": 9.97576665564061e-06, + "loss": 0.507, + "step": 701 + }, + { + "epoch": 0.12861854159032612, + "grad_norm": 0.47072785404283757, + "learning_rate": 9.975451173689658e-06, + "loss": 0.5166, + "step": 702 + }, + { + "epoch": 0.12880175888603884, + "grad_norm": 0.6675845713373886, + "learning_rate": 9.975133656499392e-06, + "loss": 0.47, + "step": 703 + }, + { + "epoch": 0.12898497618175156, + "grad_norm": 0.4533261672443661, + "learning_rate": 9.974814104199694e-06, + "loss": 0.5362, + "step": 704 + }, + { + "epoch": 0.12916819347746428, + "grad_norm": 0.4503943455070683, + "learning_rate": 9.97449251692128e-06, + "loss": 0.5134, + "step": 705 + }, + { + "epoch": 0.129351410773177, + "grad_norm": 0.44545859946020006, + "learning_rate": 9.974168894795698e-06, + "loss": 0.4981, + "step": 706 + }, + { + "epoch": 0.1295346280688897, + "grad_norm": 0.430864283858029, + "learning_rate": 9.973843237955328e-06, + "loss": 0.5329, + "step": 707 + }, + { + "epoch": 0.12971784536460243, + "grad_norm": 0.7865986917626206, + "learning_rate": 9.973515546533379e-06, + "loss": 0.504, + "step": 708 + }, + { + "epoch": 0.12990106266031515, + "grad_norm": 0.44976211133712124, + "learning_rate": 9.973185820663897e-06, + "loss": 0.4737, + "step": 709 + }, + { + "epoch": 0.13008427995602784, + "grad_norm": 0.4549304394815992, + "learning_rate": 9.97285406048176e-06, + "loss": 0.5425, + "step": 710 + }, + { + "epoch": 0.13026749725174055, + "grad_norm": 0.5049990473394447, + "learning_rate": 9.972520266122676e-06, + "loss": 0.5117, + "step": 711 + }, + { + "epoch": 0.13045071454745327, + "grad_norm": 0.42657669286902156, + "learning_rate": 9.972184437723182e-06, + "loss": 0.5057, + "step": 712 + }, + { + "epoch": 0.130633931843166, + "grad_norm": 0.5299863700854671, + "learning_rate": 9.971846575420656e-06, + "loss": 0.5262, + "step": 713 + }, + { + "epoch": 0.1308171491388787, + "grad_norm": 0.45345740582998295, + "learning_rate": 9.9715066793533e-06, + "loss": 0.4997, + "step": 714 + }, + { + "epoch": 0.13100036643459143, + "grad_norm": 0.39393086616308026, + "learning_rate": 9.971164749660149e-06, + "loss": 0.4996, + "step": 715 + }, + { + "epoch": 0.13118358373030414, + "grad_norm": 0.4577611248522203, + "learning_rate": 9.970820786481075e-06, + "loss": 0.5018, + "step": 716 + }, + { + "epoch": 0.13136680102601686, + "grad_norm": 0.43279352533859367, + "learning_rate": 9.970474789956775e-06, + "loss": 0.4914, + "step": 717 + }, + { + "epoch": 0.13155001832172958, + "grad_norm": 0.4366448132250816, + "learning_rate": 9.970126760228785e-06, + "loss": 0.4864, + "step": 718 + }, + { + "epoch": 0.1317332356174423, + "grad_norm": 0.43998158412052296, + "learning_rate": 9.969776697439463e-06, + "loss": 0.4834, + "step": 719 + }, + { + "epoch": 0.131916452913155, + "grad_norm": 0.46940772392002544, + "learning_rate": 9.969424601732011e-06, + "loss": 0.501, + "step": 720 + }, + { + "epoch": 0.1320996702088677, + "grad_norm": 0.43117110167391715, + "learning_rate": 9.96907047325045e-06, + "loss": 0.5562, + "step": 721 + }, + { + "epoch": 0.13228288750458042, + "grad_norm": 0.4119352967972661, + "learning_rate": 9.968714312139642e-06, + "loss": 0.4748, + "step": 722 + }, + { + "epoch": 0.13246610480029314, + "grad_norm": 0.5028443530663289, + "learning_rate": 9.968356118545277e-06, + "loss": 0.5344, + "step": 723 + }, + { + "epoch": 0.13264932209600586, + "grad_norm": 0.43089419048593447, + "learning_rate": 9.967995892613875e-06, + "loss": 0.4704, + "step": 724 + }, + { + "epoch": 0.13283253939171857, + "grad_norm": 0.5044933769783081, + "learning_rate": 9.967633634492788e-06, + "loss": 0.4996, + "step": 725 + }, + { + "epoch": 0.1330157566874313, + "grad_norm": 0.442554097189932, + "learning_rate": 9.967269344330201e-06, + "loss": 0.5278, + "step": 726 + }, + { + "epoch": 0.133198973983144, + "grad_norm": 0.431094495901828, + "learning_rate": 9.966903022275131e-06, + "loss": 0.4943, + "step": 727 + }, + { + "epoch": 0.13338219127885673, + "grad_norm": 0.6222105275898999, + "learning_rate": 9.966534668477421e-06, + "loss": 0.5215, + "step": 728 + }, + { + "epoch": 0.13356540857456944, + "grad_norm": 0.4562449049230116, + "learning_rate": 9.96616428308775e-06, + "loss": 0.5112, + "step": 729 + }, + { + "epoch": 0.13374862587028216, + "grad_norm": 0.4249667668518143, + "learning_rate": 9.965791866257626e-06, + "loss": 0.5083, + "step": 730 + }, + { + "epoch": 0.13393184316599488, + "grad_norm": 0.42736565388331654, + "learning_rate": 9.96541741813939e-06, + "loss": 0.5078, + "step": 731 + }, + { + "epoch": 0.1341150604617076, + "grad_norm": 0.42789645341508464, + "learning_rate": 9.96504093888621e-06, + "loss": 0.5148, + "step": 732 + }, + { + "epoch": 0.1342982777574203, + "grad_norm": 0.5018533252539279, + "learning_rate": 9.964662428652088e-06, + "loss": 0.4994, + "step": 733 + }, + { + "epoch": 0.134481495053133, + "grad_norm": 0.4561559749816332, + "learning_rate": 9.964281887591856e-06, + "loss": 0.5036, + "step": 734 + }, + { + "epoch": 0.13466471234884572, + "grad_norm": 0.43874828747054045, + "learning_rate": 9.963899315861176e-06, + "loss": 0.4976, + "step": 735 + }, + { + "epoch": 0.13484792964455844, + "grad_norm": 0.44683273370213666, + "learning_rate": 9.963514713616544e-06, + "loss": 0.4837, + "step": 736 + }, + { + "epoch": 0.13503114694027116, + "grad_norm": 0.4575940082218496, + "learning_rate": 9.963128081015282e-06, + "loss": 0.531, + "step": 737 + }, + { + "epoch": 0.13521436423598387, + "grad_norm": 0.4190671986873532, + "learning_rate": 9.962739418215545e-06, + "loss": 0.525, + "step": 738 + }, + { + "epoch": 0.1353975815316966, + "grad_norm": 0.41284775197321, + "learning_rate": 9.962348725376318e-06, + "loss": 0.519, + "step": 739 + }, + { + "epoch": 0.1355807988274093, + "grad_norm": 0.41220938724190687, + "learning_rate": 9.961956002657414e-06, + "loss": 0.535, + "step": 740 + }, + { + "epoch": 0.13576401612312203, + "grad_norm": 0.4428272636340964, + "learning_rate": 9.961561250219482e-06, + "loss": 0.5287, + "step": 741 + }, + { + "epoch": 0.13594723341883475, + "grad_norm": 0.42565025293852765, + "learning_rate": 9.961164468223996e-06, + "loss": 0.5329, + "step": 742 + }, + { + "epoch": 0.13613045071454746, + "grad_norm": 0.5963216630193675, + "learning_rate": 9.960765656833263e-06, + "loss": 0.5036, + "step": 743 + }, + { + "epoch": 0.13631366801026018, + "grad_norm": 0.42494573499796934, + "learning_rate": 9.96036481621042e-06, + "loss": 0.4936, + "step": 744 + }, + { + "epoch": 0.13649688530597287, + "grad_norm": 0.39224716002216886, + "learning_rate": 9.959961946519431e-06, + "loss": 0.4704, + "step": 745 + }, + { + "epoch": 0.1366801026016856, + "grad_norm": 0.42574265198042266, + "learning_rate": 9.959557047925095e-06, + "loss": 0.4964, + "step": 746 + }, + { + "epoch": 0.1368633198973983, + "grad_norm": 0.457270415221301, + "learning_rate": 9.959150120593035e-06, + "loss": 0.5322, + "step": 747 + }, + { + "epoch": 0.13704653719311102, + "grad_norm": 0.4245718410150897, + "learning_rate": 9.95874116468971e-06, + "loss": 0.4952, + "step": 748 + }, + { + "epoch": 0.13722975448882374, + "grad_norm": 0.47589556246220993, + "learning_rate": 9.958330180382405e-06, + "loss": 0.5124, + "step": 749 + }, + { + "epoch": 0.13741297178453646, + "grad_norm": 0.45440270321018955, + "learning_rate": 9.957917167839238e-06, + "loss": 0.5109, + "step": 750 + }, + { + "epoch": 0.13759618908024918, + "grad_norm": 0.5232685566183007, + "learning_rate": 9.95750212722915e-06, + "loss": 0.5174, + "step": 751 + }, + { + "epoch": 0.1377794063759619, + "grad_norm": 0.538012343254673, + "learning_rate": 9.957085058721918e-06, + "loss": 0.5105, + "step": 752 + }, + { + "epoch": 0.1379626236716746, + "grad_norm": 0.38959097348139504, + "learning_rate": 9.956665962488148e-06, + "loss": 0.4533, + "step": 753 + }, + { + "epoch": 0.13814584096738733, + "grad_norm": 0.4404610898092818, + "learning_rate": 9.956244838699271e-06, + "loss": 0.4978, + "step": 754 + }, + { + "epoch": 0.13832905826310005, + "grad_norm": 0.47384922621236214, + "learning_rate": 9.955821687527554e-06, + "loss": 0.5309, + "step": 755 + }, + { + "epoch": 0.13851227555881276, + "grad_norm": 0.45266354792605684, + "learning_rate": 9.955396509146084e-06, + "loss": 0.5151, + "step": 756 + }, + { + "epoch": 0.13869549285452545, + "grad_norm": 0.4098036061670406, + "learning_rate": 9.95496930372879e-06, + "loss": 0.5, + "step": 757 + }, + { + "epoch": 0.13887871015023817, + "grad_norm": 0.4520393902857722, + "learning_rate": 9.954540071450418e-06, + "loss": 0.4766, + "step": 758 + }, + { + "epoch": 0.1390619274459509, + "grad_norm": 0.41592431105518124, + "learning_rate": 9.95410881248655e-06, + "loss": 0.5279, + "step": 759 + }, + { + "epoch": 0.1392451447416636, + "grad_norm": 0.9055926932881602, + "learning_rate": 9.953675527013594e-06, + "loss": 0.4544, + "step": 760 + }, + { + "epoch": 0.13942836203737632, + "grad_norm": 0.4230083797544526, + "learning_rate": 9.953240215208787e-06, + "loss": 0.5174, + "step": 761 + }, + { + "epoch": 0.13961157933308904, + "grad_norm": 0.46167105853029766, + "learning_rate": 9.9528028772502e-06, + "loss": 0.5087, + "step": 762 + }, + { + "epoch": 0.13979479662880176, + "grad_norm": 0.4614988799168283, + "learning_rate": 9.952363513316727e-06, + "loss": 0.5035, + "step": 763 + }, + { + "epoch": 0.13997801392451448, + "grad_norm": 0.4599378559524754, + "learning_rate": 9.951922123588091e-06, + "loss": 0.4982, + "step": 764 + }, + { + "epoch": 0.1401612312202272, + "grad_norm": 0.4169187563233499, + "learning_rate": 9.951478708244847e-06, + "loss": 0.5205, + "step": 765 + }, + { + "epoch": 0.1403444485159399, + "grad_norm": 1.4346541301623688, + "learning_rate": 9.951033267468375e-06, + "loss": 0.4704, + "step": 766 + }, + { + "epoch": 0.14052766581165263, + "grad_norm": 0.4477016515011828, + "learning_rate": 9.950585801440889e-06, + "loss": 0.489, + "step": 767 + }, + { + "epoch": 0.14071088310736535, + "grad_norm": 0.4178701034776549, + "learning_rate": 9.950136310345425e-06, + "loss": 0.5147, + "step": 768 + }, + { + "epoch": 0.14089410040307804, + "grad_norm": 0.4893104037393506, + "learning_rate": 9.949684794365848e-06, + "loss": 0.4913, + "step": 769 + }, + { + "epoch": 0.14107731769879076, + "grad_norm": 0.4018838415277579, + "learning_rate": 9.949231253686857e-06, + "loss": 0.464, + "step": 770 + }, + { + "epoch": 0.14126053499450347, + "grad_norm": 0.38608379924790914, + "learning_rate": 9.948775688493974e-06, + "loss": 0.5173, + "step": 771 + }, + { + "epoch": 0.1414437522902162, + "grad_norm": 0.4162388206396457, + "learning_rate": 9.948318098973552e-06, + "loss": 0.4917, + "step": 772 + }, + { + "epoch": 0.1416269695859289, + "grad_norm": 0.3781950328417496, + "learning_rate": 9.947858485312772e-06, + "loss": 0.4446, + "step": 773 + }, + { + "epoch": 0.14181018688164163, + "grad_norm": 0.45959051057998834, + "learning_rate": 9.947396847699638e-06, + "loss": 0.4918, + "step": 774 + }, + { + "epoch": 0.14199340417735434, + "grad_norm": 0.480317102541819, + "learning_rate": 9.946933186322988e-06, + "loss": 0.5369, + "step": 775 + }, + { + "epoch": 0.14217662147306706, + "grad_norm": 0.5008985493720574, + "learning_rate": 9.946467501372485e-06, + "loss": 0.4992, + "step": 776 + }, + { + "epoch": 0.14235983876877978, + "grad_norm": 0.4117591563183844, + "learning_rate": 9.94599979303862e-06, + "loss": 0.4818, + "step": 777 + }, + { + "epoch": 0.1425430560644925, + "grad_norm": 0.531108956987235, + "learning_rate": 9.945530061512714e-06, + "loss": 0.517, + "step": 778 + }, + { + "epoch": 0.1427262733602052, + "grad_norm": 0.43817523582155865, + "learning_rate": 9.945058306986911e-06, + "loss": 0.5473, + "step": 779 + }, + { + "epoch": 0.14290949065591793, + "grad_norm": 0.4747080227792923, + "learning_rate": 9.944584529654187e-06, + "loss": 0.4819, + "step": 780 + }, + { + "epoch": 0.14309270795163062, + "grad_norm": 0.4249215515551791, + "learning_rate": 9.944108729708342e-06, + "loss": 0.4979, + "step": 781 + }, + { + "epoch": 0.14327592524734334, + "grad_norm": 0.43558257752578505, + "learning_rate": 9.943630907344008e-06, + "loss": 0.484, + "step": 782 + }, + { + "epoch": 0.14345914254305606, + "grad_norm": 0.4750219883418816, + "learning_rate": 9.943151062756638e-06, + "loss": 0.4825, + "step": 783 + }, + { + "epoch": 0.14364235983876877, + "grad_norm": 0.4332476417990451, + "learning_rate": 9.942669196142516e-06, + "loss": 0.4929, + "step": 784 + }, + { + "epoch": 0.1438255771344815, + "grad_norm": 0.48307720479460303, + "learning_rate": 9.942185307698754e-06, + "loss": 0.5099, + "step": 785 + }, + { + "epoch": 0.1440087944301942, + "grad_norm": 0.43004630110605213, + "learning_rate": 9.941699397623289e-06, + "loss": 0.4435, + "step": 786 + }, + { + "epoch": 0.14419201172590693, + "grad_norm": 0.5526090978976514, + "learning_rate": 9.941211466114883e-06, + "loss": 0.4831, + "step": 787 + }, + { + "epoch": 0.14437522902161964, + "grad_norm": 0.4299548315161719, + "learning_rate": 9.94072151337313e-06, + "loss": 0.5055, + "step": 788 + }, + { + "epoch": 0.14455844631733236, + "grad_norm": 0.46488795405005734, + "learning_rate": 9.940229539598449e-06, + "loss": 0.5247, + "step": 789 + }, + { + "epoch": 0.14474166361304508, + "grad_norm": 0.45711196453956754, + "learning_rate": 9.939735544992084e-06, + "loss": 0.5155, + "step": 790 + }, + { + "epoch": 0.1449248809087578, + "grad_norm": 0.39377908902960373, + "learning_rate": 9.939239529756106e-06, + "loss": 0.4861, + "step": 791 + }, + { + "epoch": 0.14510809820447051, + "grad_norm": 0.3780973854580167, + "learning_rate": 9.938741494093413e-06, + "loss": 0.4743, + "step": 792 + }, + { + "epoch": 0.1452913155001832, + "grad_norm": 0.4442667746426281, + "learning_rate": 9.93824143820773e-06, + "loss": 0.5124, + "step": 793 + }, + { + "epoch": 0.14547453279589592, + "grad_norm": 0.47139079982459514, + "learning_rate": 9.93773936230361e-06, + "loss": 0.5095, + "step": 794 + }, + { + "epoch": 0.14565775009160864, + "grad_norm": 0.47478113201389704, + "learning_rate": 9.937235266586425e-06, + "loss": 0.5242, + "step": 795 + }, + { + "epoch": 0.14584096738732136, + "grad_norm": 0.4253566981856297, + "learning_rate": 9.936729151262383e-06, + "loss": 0.4816, + "step": 796 + }, + { + "epoch": 0.14602418468303408, + "grad_norm": 0.3995444614750483, + "learning_rate": 9.936221016538514e-06, + "loss": 0.5114, + "step": 797 + }, + { + "epoch": 0.1462074019787468, + "grad_norm": 0.3958644449360531, + "learning_rate": 9.935710862622671e-06, + "loss": 0.5044, + "step": 798 + }, + { + "epoch": 0.1463906192744595, + "grad_norm": 0.39672728856696343, + "learning_rate": 9.935198689723537e-06, + "loss": 0.5079, + "step": 799 + }, + { + "epoch": 0.14657383657017223, + "grad_norm": 0.4401216989939036, + "learning_rate": 9.934684498050619e-06, + "loss": 0.5169, + "step": 800 + }, + { + "epoch": 0.14675705386588495, + "grad_norm": 0.42531985045283716, + "learning_rate": 9.93416828781425e-06, + "loss": 0.5039, + "step": 801 + }, + { + "epoch": 0.14694027116159766, + "grad_norm": 0.43641771981673, + "learning_rate": 9.93365005922559e-06, + "loss": 0.4705, + "step": 802 + }, + { + "epoch": 0.14712348845731038, + "grad_norm": 0.4154214727248315, + "learning_rate": 9.933129812496623e-06, + "loss": 0.4801, + "step": 803 + }, + { + "epoch": 0.1473067057530231, + "grad_norm": 0.4668765888893056, + "learning_rate": 9.932607547840161e-06, + "loss": 0.5262, + "step": 804 + }, + { + "epoch": 0.1474899230487358, + "grad_norm": 0.45163409733644266, + "learning_rate": 9.932083265469836e-06, + "loss": 0.5009, + "step": 805 + }, + { + "epoch": 0.1476731403444485, + "grad_norm": 0.44602336917824525, + "learning_rate": 9.93155696560011e-06, + "loss": 0.5071, + "step": 806 + }, + { + "epoch": 0.14785635764016122, + "grad_norm": 0.40571563214546336, + "learning_rate": 9.931028648446273e-06, + "loss": 0.4867, + "step": 807 + }, + { + "epoch": 0.14803957493587394, + "grad_norm": 0.412866633306789, + "learning_rate": 9.930498314224433e-06, + "loss": 0.5204, + "step": 808 + }, + { + "epoch": 0.14822279223158666, + "grad_norm": 0.4491417286505596, + "learning_rate": 9.929965963151526e-06, + "loss": 0.4959, + "step": 809 + }, + { + "epoch": 0.14840600952729938, + "grad_norm": 0.483698524685917, + "learning_rate": 9.929431595445315e-06, + "loss": 0.4945, + "step": 810 + }, + { + "epoch": 0.1485892268230121, + "grad_norm": 0.45809625895317874, + "learning_rate": 9.928895211324387e-06, + "loss": 0.4795, + "step": 811 + }, + { + "epoch": 0.1487724441187248, + "grad_norm": 0.4671855759216309, + "learning_rate": 9.928356811008153e-06, + "loss": 0.502, + "step": 812 + }, + { + "epoch": 0.14895566141443753, + "grad_norm": 0.49991252806247843, + "learning_rate": 9.927816394716847e-06, + "loss": 0.4989, + "step": 813 + }, + { + "epoch": 0.14913887871015025, + "grad_norm": 0.4930888660574426, + "learning_rate": 9.92727396267153e-06, + "loss": 0.5256, + "step": 814 + }, + { + "epoch": 0.14932209600586296, + "grad_norm": 0.4301245288190715, + "learning_rate": 9.926729515094092e-06, + "loss": 0.5062, + "step": 815 + }, + { + "epoch": 0.14950531330157568, + "grad_norm": 0.4254345495226345, + "learning_rate": 9.926183052207235e-06, + "loss": 0.4885, + "step": 816 + }, + { + "epoch": 0.14968853059728837, + "grad_norm": 0.40854658960332646, + "learning_rate": 9.925634574234499e-06, + "loss": 0.4909, + "step": 817 + }, + { + "epoch": 0.1498717478930011, + "grad_norm": 0.4307966660913799, + "learning_rate": 9.925084081400241e-06, + "loss": 0.4974, + "step": 818 + }, + { + "epoch": 0.1500549651887138, + "grad_norm": 0.44710983697734263, + "learning_rate": 9.924531573929641e-06, + "loss": 0.5107, + "step": 819 + }, + { + "epoch": 0.15023818248442652, + "grad_norm": 0.47247983608563787, + "learning_rate": 9.923977052048708e-06, + "loss": 0.5185, + "step": 820 + }, + { + "epoch": 0.15042139978013924, + "grad_norm": 0.43408666873611645, + "learning_rate": 9.923420515984272e-06, + "loss": 0.5256, + "step": 821 + }, + { + "epoch": 0.15060461707585196, + "grad_norm": 0.3618361050633503, + "learning_rate": 9.922861965963987e-06, + "loss": 0.4835, + "step": 822 + }, + { + "epoch": 0.15078783437156468, + "grad_norm": 0.478572088704784, + "learning_rate": 9.922301402216334e-06, + "loss": 0.4971, + "step": 823 + }, + { + "epoch": 0.1509710516672774, + "grad_norm": 0.4956723559981853, + "learning_rate": 9.921738824970611e-06, + "loss": 0.4664, + "step": 824 + }, + { + "epoch": 0.1511542689629901, + "grad_norm": 0.43104833444065105, + "learning_rate": 9.921174234456947e-06, + "loss": 0.5056, + "step": 825 + }, + { + "epoch": 0.15133748625870283, + "grad_norm": 0.38636058698792763, + "learning_rate": 9.920607630906289e-06, + "loss": 0.509, + "step": 826 + }, + { + "epoch": 0.15152070355441555, + "grad_norm": 0.4638108326048624, + "learning_rate": 9.920039014550413e-06, + "loss": 0.4928, + "step": 827 + }, + { + "epoch": 0.15170392085012827, + "grad_norm": 0.38962500919500936, + "learning_rate": 9.919468385621912e-06, + "loss": 0.5025, + "step": 828 + }, + { + "epoch": 0.15188713814584096, + "grad_norm": 0.423466480397551, + "learning_rate": 9.918895744354204e-06, + "loss": 0.4903, + "step": 829 + }, + { + "epoch": 0.15207035544155367, + "grad_norm": 0.5945183810824295, + "learning_rate": 9.918321090981537e-06, + "loss": 0.4961, + "step": 830 + }, + { + "epoch": 0.1522535727372664, + "grad_norm": 0.3857346095853992, + "learning_rate": 9.917744425738971e-06, + "loss": 0.4903, + "step": 831 + }, + { + "epoch": 0.1524367900329791, + "grad_norm": 0.4368019305898112, + "learning_rate": 9.917165748862398e-06, + "loss": 0.4985, + "step": 832 + }, + { + "epoch": 0.15262000732869183, + "grad_norm": 0.4622792781214721, + "learning_rate": 9.916585060588526e-06, + "loss": 0.5298, + "step": 833 + }, + { + "epoch": 0.15280322462440454, + "grad_norm": 0.43464073300377387, + "learning_rate": 9.916002361154894e-06, + "loss": 0.4956, + "step": 834 + }, + { + "epoch": 0.15298644192011726, + "grad_norm": 0.37914154472744427, + "learning_rate": 9.915417650799855e-06, + "loss": 0.4914, + "step": 835 + }, + { + "epoch": 0.15316965921582998, + "grad_norm": 0.41110768627195465, + "learning_rate": 9.914830929762588e-06, + "loss": 0.4933, + "step": 836 + }, + { + "epoch": 0.1533528765115427, + "grad_norm": 0.3922989216974703, + "learning_rate": 9.914242198283099e-06, + "loss": 0.5, + "step": 837 + }, + { + "epoch": 0.1535360938072554, + "grad_norm": 0.4310098182316404, + "learning_rate": 9.91365145660221e-06, + "loss": 0.4817, + "step": 838 + }, + { + "epoch": 0.15371931110296813, + "grad_norm": 0.40690680117071937, + "learning_rate": 9.913058704961565e-06, + "loss": 0.4778, + "step": 839 + }, + { + "epoch": 0.15390252839868085, + "grad_norm": 0.4584415585967629, + "learning_rate": 9.912463943603635e-06, + "loss": 0.4585, + "step": 840 + }, + { + "epoch": 0.15408574569439354, + "grad_norm": 0.48734349658323495, + "learning_rate": 9.911867172771711e-06, + "loss": 0.5345, + "step": 841 + }, + { + "epoch": 0.15426896299010626, + "grad_norm": 0.442686833828428, + "learning_rate": 9.911268392709908e-06, + "loss": 0.5134, + "step": 842 + }, + { + "epoch": 0.15445218028581897, + "grad_norm": 0.44506293028939, + "learning_rate": 9.910667603663156e-06, + "loss": 0.5158, + "step": 843 + }, + { + "epoch": 0.1546353975815317, + "grad_norm": 0.4311992637968541, + "learning_rate": 9.910064805877214e-06, + "loss": 0.4837, + "step": 844 + }, + { + "epoch": 0.1548186148772444, + "grad_norm": 0.4493099076015729, + "learning_rate": 9.90945999959866e-06, + "loss": 0.5108, + "step": 845 + }, + { + "epoch": 0.15500183217295713, + "grad_norm": 0.45478365363321205, + "learning_rate": 9.908853185074896e-06, + "loss": 0.4998, + "step": 846 + }, + { + "epoch": 0.15518504946866984, + "grad_norm": 0.46301414888311854, + "learning_rate": 9.90824436255414e-06, + "loss": 0.5225, + "step": 847 + }, + { + "epoch": 0.15536826676438256, + "grad_norm": 0.40118617155548936, + "learning_rate": 9.907633532285435e-06, + "loss": 0.4927, + "step": 848 + }, + { + "epoch": 0.15555148406009528, + "grad_norm": 0.49481295874709574, + "learning_rate": 9.907020694518646e-06, + "loss": 0.4933, + "step": 849 + }, + { + "epoch": 0.155734701355808, + "grad_norm": 0.4102626498510104, + "learning_rate": 9.90640584950446e-06, + "loss": 0.5111, + "step": 850 + }, + { + "epoch": 0.15591791865152071, + "grad_norm": 0.46514690031713113, + "learning_rate": 9.905788997494377e-06, + "loss": 0.4728, + "step": 851 + }, + { + "epoch": 0.15610113594723343, + "grad_norm": 0.4054441718781369, + "learning_rate": 9.905170138740732e-06, + "loss": 0.4828, + "step": 852 + }, + { + "epoch": 0.15628435324294612, + "grad_norm": 0.4400252577977126, + "learning_rate": 9.904549273496666e-06, + "loss": 0.501, + "step": 853 + }, + { + "epoch": 0.15646757053865884, + "grad_norm": 0.42167618939817325, + "learning_rate": 9.903926402016153e-06, + "loss": 0.4922, + "step": 854 + }, + { + "epoch": 0.15665078783437156, + "grad_norm": 0.4093707700968775, + "learning_rate": 9.90330152455398e-06, + "loss": 0.5014, + "step": 855 + }, + { + "epoch": 0.15683400513008428, + "grad_norm": 0.4523522809487888, + "learning_rate": 9.902674641365756e-06, + "loss": 0.4829, + "step": 856 + }, + { + "epoch": 0.157017222425797, + "grad_norm": 0.49143897815029397, + "learning_rate": 9.902045752707916e-06, + "loss": 0.5074, + "step": 857 + }, + { + "epoch": 0.1572004397215097, + "grad_norm": 0.43621740114283203, + "learning_rate": 9.901414858837706e-06, + "loss": 0.5182, + "step": 858 + }, + { + "epoch": 0.15738365701722243, + "grad_norm": 0.40732498942379025, + "learning_rate": 9.9007819600132e-06, + "loss": 0.4674, + "step": 859 + }, + { + "epoch": 0.15756687431293515, + "grad_norm": 0.43675476681733244, + "learning_rate": 9.900147056493285e-06, + "loss": 0.5374, + "step": 860 + }, + { + "epoch": 0.15775009160864786, + "grad_norm": 0.43060074322101793, + "learning_rate": 9.89951014853768e-06, + "loss": 0.4635, + "step": 861 + }, + { + "epoch": 0.15793330890436058, + "grad_norm": 0.41179202097945616, + "learning_rate": 9.898871236406907e-06, + "loss": 0.4951, + "step": 862 + }, + { + "epoch": 0.1581165262000733, + "grad_norm": 0.371370018002044, + "learning_rate": 9.898230320362323e-06, + "loss": 0.479, + "step": 863 + }, + { + "epoch": 0.158299743495786, + "grad_norm": 0.4300134296591519, + "learning_rate": 9.897587400666097e-06, + "loss": 0.4868, + "step": 864 + }, + { + "epoch": 0.1584829607914987, + "grad_norm": 0.43589209027697184, + "learning_rate": 9.896942477581221e-06, + "loss": 0.5124, + "step": 865 + }, + { + "epoch": 0.15866617808721142, + "grad_norm": 0.575247632832188, + "learning_rate": 9.896295551371503e-06, + "loss": 0.5053, + "step": 866 + }, + { + "epoch": 0.15884939538292414, + "grad_norm": 0.4273872012355264, + "learning_rate": 9.89564662230157e-06, + "loss": 0.5138, + "step": 867 + }, + { + "epoch": 0.15903261267863686, + "grad_norm": 0.4543163505662005, + "learning_rate": 9.894995690636874e-06, + "loss": 0.5229, + "step": 868 + }, + { + "epoch": 0.15921582997434958, + "grad_norm": 0.4491410597219586, + "learning_rate": 9.894342756643684e-06, + "loss": 0.5045, + "step": 869 + }, + { + "epoch": 0.1593990472700623, + "grad_norm": 0.47087708750396473, + "learning_rate": 9.893687820589081e-06, + "loss": 0.5087, + "step": 870 + }, + { + "epoch": 0.159582264565775, + "grad_norm": 0.4240236693967998, + "learning_rate": 9.893030882740977e-06, + "loss": 0.5052, + "step": 871 + }, + { + "epoch": 0.15976548186148773, + "grad_norm": 0.4287525963131784, + "learning_rate": 9.892371943368092e-06, + "loss": 0.527, + "step": 872 + }, + { + "epoch": 0.15994869915720045, + "grad_norm": 0.4131563755872441, + "learning_rate": 9.891711002739971e-06, + "loss": 0.5074, + "step": 873 + }, + { + "epoch": 0.16013191645291316, + "grad_norm": 0.39231143806918656, + "learning_rate": 9.891048061126975e-06, + "loss": 0.5018, + "step": 874 + }, + { + "epoch": 0.16031513374862588, + "grad_norm": 0.4391493046066672, + "learning_rate": 9.890383118800287e-06, + "loss": 0.518, + "step": 875 + }, + { + "epoch": 0.16049835104433857, + "grad_norm": 0.468016398258625, + "learning_rate": 9.889716176031903e-06, + "loss": 0.5191, + "step": 876 + }, + { + "epoch": 0.1606815683400513, + "grad_norm": 0.4619425954928644, + "learning_rate": 9.88904723309464e-06, + "loss": 0.5114, + "step": 877 + }, + { + "epoch": 0.160864785635764, + "grad_norm": 0.4581492770476195, + "learning_rate": 9.888376290262134e-06, + "loss": 0.5349, + "step": 878 + }, + { + "epoch": 0.16104800293147672, + "grad_norm": 0.390022001202663, + "learning_rate": 9.887703347808838e-06, + "loss": 0.507, + "step": 879 + }, + { + "epoch": 0.16123122022718944, + "grad_norm": 0.38271876665722593, + "learning_rate": 9.887028406010026e-06, + "loss": 0.4987, + "step": 880 + }, + { + "epoch": 0.16141443752290216, + "grad_norm": 0.39991523006604696, + "learning_rate": 9.886351465141785e-06, + "loss": 0.4807, + "step": 881 + }, + { + "epoch": 0.16159765481861488, + "grad_norm": 0.3867195675970219, + "learning_rate": 9.88567252548102e-06, + "loss": 0.515, + "step": 882 + }, + { + "epoch": 0.1617808721143276, + "grad_norm": 0.47566001886735476, + "learning_rate": 9.884991587305459e-06, + "loss": 0.4461, + "step": 883 + }, + { + "epoch": 0.1619640894100403, + "grad_norm": 0.4453653597243864, + "learning_rate": 9.884308650893642e-06, + "loss": 0.5493, + "step": 884 + }, + { + "epoch": 0.16214730670575303, + "grad_norm": 0.38787040868110323, + "learning_rate": 9.883623716524929e-06, + "loss": 0.4998, + "step": 885 + }, + { + "epoch": 0.16233052400146575, + "grad_norm": 0.4570258592705414, + "learning_rate": 9.882936784479498e-06, + "loss": 0.5226, + "step": 886 + }, + { + "epoch": 0.16251374129717847, + "grad_norm": 0.4438200453897428, + "learning_rate": 9.882247855038339e-06, + "loss": 0.5073, + "step": 887 + }, + { + "epoch": 0.16269695859289116, + "grad_norm": 0.3887912773501902, + "learning_rate": 9.881556928483266e-06, + "loss": 0.5213, + "step": 888 + }, + { + "epoch": 0.16288017588860387, + "grad_norm": 0.44048355093202696, + "learning_rate": 9.880864005096906e-06, + "loss": 0.511, + "step": 889 + }, + { + "epoch": 0.1630633931843166, + "grad_norm": 0.44235079469227967, + "learning_rate": 9.880169085162703e-06, + "loss": 0.5223, + "step": 890 + }, + { + "epoch": 0.1632466104800293, + "grad_norm": 0.3877832289424, + "learning_rate": 9.87947216896492e-06, + "loss": 0.4537, + "step": 891 + }, + { + "epoch": 0.16342982777574203, + "grad_norm": 0.44869938145307137, + "learning_rate": 9.878773256788635e-06, + "loss": 0.5013, + "step": 892 + }, + { + "epoch": 0.16361304507145474, + "grad_norm": 0.46375171983760244, + "learning_rate": 9.878072348919738e-06, + "loss": 0.489, + "step": 893 + }, + { + "epoch": 0.16379626236716746, + "grad_norm": 0.49608852880970994, + "learning_rate": 9.877369445644945e-06, + "loss": 0.5177, + "step": 894 + }, + { + "epoch": 0.16397947966288018, + "grad_norm": 0.4082183549684993, + "learning_rate": 9.876664547251781e-06, + "loss": 0.5091, + "step": 895 + }, + { + "epoch": 0.1641626969585929, + "grad_norm": 0.4347968628628821, + "learning_rate": 9.875957654028588e-06, + "loss": 0.5242, + "step": 896 + }, + { + "epoch": 0.16434591425430561, + "grad_norm": 0.45545255399303036, + "learning_rate": 9.875248766264527e-06, + "loss": 0.5406, + "step": 897 + }, + { + "epoch": 0.16452913155001833, + "grad_norm": 0.41358644063630506, + "learning_rate": 9.874537884249574e-06, + "loss": 0.511, + "step": 898 + }, + { + "epoch": 0.16471234884573105, + "grad_norm": 0.4576667619206321, + "learning_rate": 9.873825008274514e-06, + "loss": 0.5073, + "step": 899 + }, + { + "epoch": 0.16489556614144374, + "grad_norm": 0.47721781865779744, + "learning_rate": 9.87311013863096e-06, + "loss": 0.5044, + "step": 900 + }, + { + "epoch": 0.16507878343715646, + "grad_norm": 0.4622288619794578, + "learning_rate": 9.872393275611329e-06, + "loss": 0.5096, + "step": 901 + }, + { + "epoch": 0.16526200073286917, + "grad_norm": 0.41177688817899083, + "learning_rate": 9.871674419508864e-06, + "loss": 0.523, + "step": 902 + }, + { + "epoch": 0.1654452180285819, + "grad_norm": 0.371054810469729, + "learning_rate": 9.87095357061761e-06, + "loss": 0.4897, + "step": 903 + }, + { + "epoch": 0.1656284353242946, + "grad_norm": 0.41574608839188226, + "learning_rate": 9.87023072923244e-06, + "loss": 0.4823, + "step": 904 + }, + { + "epoch": 0.16581165262000733, + "grad_norm": 0.4218915251343208, + "learning_rate": 9.869505895649036e-06, + "loss": 0.5079, + "step": 905 + }, + { + "epoch": 0.16599486991572004, + "grad_norm": 0.5549831381708766, + "learning_rate": 9.868779070163895e-06, + "loss": 0.484, + "step": 906 + }, + { + "epoch": 0.16617808721143276, + "grad_norm": 0.3891473564827362, + "learning_rate": 9.868050253074328e-06, + "loss": 0.4903, + "step": 907 + }, + { + "epoch": 0.16636130450714548, + "grad_norm": 0.4268945705632106, + "learning_rate": 9.867319444678465e-06, + "loss": 0.5024, + "step": 908 + }, + { + "epoch": 0.1665445218028582, + "grad_norm": 0.40332206359603545, + "learning_rate": 9.866586645275247e-06, + "loss": 0.4477, + "step": 909 + }, + { + "epoch": 0.16672773909857092, + "grad_norm": 0.3720747589583317, + "learning_rate": 9.86585185516443e-06, + "loss": 0.469, + "step": 910 + }, + { + "epoch": 0.16691095639428363, + "grad_norm": 0.451113552639029, + "learning_rate": 9.865115074646583e-06, + "loss": 0.5221, + "step": 911 + }, + { + "epoch": 0.16709417368999632, + "grad_norm": 0.4454998332020719, + "learning_rate": 9.864376304023092e-06, + "loss": 0.495, + "step": 912 + }, + { + "epoch": 0.16727739098570904, + "grad_norm": 0.5363910151131063, + "learning_rate": 9.863635543596156e-06, + "loss": 0.5233, + "step": 913 + }, + { + "epoch": 0.16746060828142176, + "grad_norm": 0.4392403856880424, + "learning_rate": 9.862892793668787e-06, + "loss": 0.4998, + "step": 914 + }, + { + "epoch": 0.16764382557713448, + "grad_norm": 0.42093143672832734, + "learning_rate": 9.862148054544812e-06, + "loss": 0.4977, + "step": 915 + }, + { + "epoch": 0.1678270428728472, + "grad_norm": 0.39275787265485074, + "learning_rate": 9.86140132652887e-06, + "loss": 0.4694, + "step": 916 + }, + { + "epoch": 0.1680102601685599, + "grad_norm": 0.4159908861548446, + "learning_rate": 9.860652609926417e-06, + "loss": 0.4886, + "step": 917 + }, + { + "epoch": 0.16819347746427263, + "grad_norm": 0.38579180273261776, + "learning_rate": 9.859901905043718e-06, + "loss": 0.48, + "step": 918 + }, + { + "epoch": 0.16837669475998535, + "grad_norm": 0.4273621763370794, + "learning_rate": 9.859149212187855e-06, + "loss": 0.512, + "step": 919 + }, + { + "epoch": 0.16855991205569806, + "grad_norm": 0.45913559298162226, + "learning_rate": 9.85839453166672e-06, + "loss": 0.4877, + "step": 920 + }, + { + "epoch": 0.16874312935141078, + "grad_norm": 0.3847363797846655, + "learning_rate": 9.85763786378902e-06, + "loss": 0.4586, + "step": 921 + }, + { + "epoch": 0.1689263466471235, + "grad_norm": 0.4872256895625686, + "learning_rate": 9.856879208864277e-06, + "loss": 0.5035, + "step": 922 + }, + { + "epoch": 0.16910956394283622, + "grad_norm": 0.47756329412725707, + "learning_rate": 9.85611856720282e-06, + "loss": 0.5031, + "step": 923 + }, + { + "epoch": 0.1692927812385489, + "grad_norm": 0.474431940864528, + "learning_rate": 9.8553559391158e-06, + "loss": 0.512, + "step": 924 + }, + { + "epoch": 0.16947599853426162, + "grad_norm": 0.41972238109665655, + "learning_rate": 9.85459132491517e-06, + "loss": 0.5386, + "step": 925 + }, + { + "epoch": 0.16965921582997434, + "grad_norm": 0.4369771091332424, + "learning_rate": 9.8538247249137e-06, + "loss": 0.4909, + "step": 926 + }, + { + "epoch": 0.16984243312568706, + "grad_norm": 0.4523772832392917, + "learning_rate": 9.853056139424974e-06, + "loss": 0.4951, + "step": 927 + }, + { + "epoch": 0.17002565042139978, + "grad_norm": 0.40235086982371876, + "learning_rate": 9.852285568763387e-06, + "loss": 0.4828, + "step": 928 + }, + { + "epoch": 0.1702088677171125, + "grad_norm": 0.4150844402567988, + "learning_rate": 9.851513013244144e-06, + "loss": 0.4933, + "step": 929 + }, + { + "epoch": 0.1703920850128252, + "grad_norm": 0.3941602147072389, + "learning_rate": 9.850738473183266e-06, + "loss": 0.4855, + "step": 930 + }, + { + "epoch": 0.17057530230853793, + "grad_norm": 0.47118394280764797, + "learning_rate": 9.849961948897582e-06, + "loss": 0.5024, + "step": 931 + }, + { + "epoch": 0.17075851960425065, + "grad_norm": 0.3912878663413955, + "learning_rate": 9.849183440704735e-06, + "loss": 0.5041, + "step": 932 + }, + { + "epoch": 0.17094173689996336, + "grad_norm": 0.4001961484144372, + "learning_rate": 9.848402948923177e-06, + "loss": 0.4753, + "step": 933 + }, + { + "epoch": 0.17112495419567608, + "grad_norm": 0.4441137892348275, + "learning_rate": 9.847620473872172e-06, + "loss": 0.5102, + "step": 934 + }, + { + "epoch": 0.1713081714913888, + "grad_norm": 0.40101758764349804, + "learning_rate": 9.846836015871802e-06, + "loss": 0.4824, + "step": 935 + }, + { + "epoch": 0.1714913887871015, + "grad_norm": 0.41075944565503664, + "learning_rate": 9.846049575242949e-06, + "loss": 0.5235, + "step": 936 + }, + { + "epoch": 0.1716746060828142, + "grad_norm": 0.3780837940291423, + "learning_rate": 9.845261152307312e-06, + "loss": 0.5211, + "step": 937 + }, + { + "epoch": 0.17185782337852692, + "grad_norm": 0.45070991069476113, + "learning_rate": 9.844470747387403e-06, + "loss": 0.4926, + "step": 938 + }, + { + "epoch": 0.17204104067423964, + "grad_norm": 0.40836878453605646, + "learning_rate": 9.843678360806542e-06, + "loss": 0.5173, + "step": 939 + }, + { + "epoch": 0.17222425796995236, + "grad_norm": 0.36676941914948374, + "learning_rate": 9.842883992888855e-06, + "loss": 0.5101, + "step": 940 + }, + { + "epoch": 0.17240747526566508, + "grad_norm": 0.42894488109200807, + "learning_rate": 9.842087643959288e-06, + "loss": 0.5398, + "step": 941 + }, + { + "epoch": 0.1725906925613778, + "grad_norm": 0.4004768449690441, + "learning_rate": 9.841289314343591e-06, + "loss": 0.4967, + "step": 942 + }, + { + "epoch": 0.1727739098570905, + "grad_norm": 0.3781076276341992, + "learning_rate": 9.840489004368325e-06, + "loss": 0.4731, + "step": 943 + }, + { + "epoch": 0.17295712715280323, + "grad_norm": 0.45307945315267256, + "learning_rate": 9.839686714360864e-06, + "loss": 0.4737, + "step": 944 + }, + { + "epoch": 0.17314034444851595, + "grad_norm": 0.4079247794136399, + "learning_rate": 9.838882444649387e-06, + "loss": 0.4927, + "step": 945 + }, + { + "epoch": 0.17332356174422867, + "grad_norm": 0.47171018664243136, + "learning_rate": 9.838076195562886e-06, + "loss": 0.5009, + "step": 946 + }, + { + "epoch": 0.17350677903994138, + "grad_norm": 0.44572108382734166, + "learning_rate": 9.837267967431164e-06, + "loss": 0.4858, + "step": 947 + }, + { + "epoch": 0.17368999633565407, + "grad_norm": 0.40649730883742624, + "learning_rate": 9.83645776058483e-06, + "loss": 0.4943, + "step": 948 + }, + { + "epoch": 0.1738732136313668, + "grad_norm": 0.43002759908477306, + "learning_rate": 9.835645575355304e-06, + "loss": 0.4872, + "step": 949 + }, + { + "epoch": 0.1740564309270795, + "grad_norm": 0.39064612647120167, + "learning_rate": 9.834831412074816e-06, + "loss": 0.4409, + "step": 950 + }, + { + "epoch": 0.17423964822279223, + "grad_norm": 0.3761573603114394, + "learning_rate": 9.834015271076405e-06, + "loss": 0.459, + "step": 951 + }, + { + "epoch": 0.17442286551850494, + "grad_norm": 0.42796519720566795, + "learning_rate": 9.83319715269392e-06, + "loss": 0.4906, + "step": 952 + }, + { + "epoch": 0.17460608281421766, + "grad_norm": 0.5091807084893948, + "learning_rate": 9.832377057262015e-06, + "loss": 0.5516, + "step": 953 + }, + { + "epoch": 0.17478930010993038, + "grad_norm": 0.48065067696295527, + "learning_rate": 9.831554985116155e-06, + "loss": 0.5097, + "step": 954 + }, + { + "epoch": 0.1749725174056431, + "grad_norm": 0.4310364985277951, + "learning_rate": 9.830730936592615e-06, + "loss": 0.5025, + "step": 955 + }, + { + "epoch": 0.17515573470135581, + "grad_norm": 0.39964086093662415, + "learning_rate": 9.829904912028477e-06, + "loss": 0.494, + "step": 956 + }, + { + "epoch": 0.17533895199706853, + "grad_norm": 0.40935380850458175, + "learning_rate": 9.829076911761631e-06, + "loss": 0.5182, + "step": 957 + }, + { + "epoch": 0.17552216929278125, + "grad_norm": 0.454766087870644, + "learning_rate": 9.828246936130777e-06, + "loss": 0.5113, + "step": 958 + }, + { + "epoch": 0.17570538658849397, + "grad_norm": 0.5087972352515657, + "learning_rate": 9.827414985475419e-06, + "loss": 0.4797, + "step": 959 + }, + { + "epoch": 0.17588860388420666, + "grad_norm": 0.4255562501722622, + "learning_rate": 9.826581060135873e-06, + "loss": 0.5045, + "step": 960 + }, + { + "epoch": 0.17607182117991937, + "grad_norm": 0.45583787528367603, + "learning_rate": 9.825745160453264e-06, + "loss": 0.4755, + "step": 961 + }, + { + "epoch": 0.1762550384756321, + "grad_norm": 0.45648883029913284, + "learning_rate": 9.824907286769519e-06, + "loss": 0.4736, + "step": 962 + }, + { + "epoch": 0.1764382557713448, + "grad_norm": 0.436071698669034, + "learning_rate": 9.824067439427374e-06, + "loss": 0.5036, + "step": 963 + }, + { + "epoch": 0.17662147306705753, + "grad_norm": 0.41502000197263017, + "learning_rate": 9.823225618770378e-06, + "loss": 0.5121, + "step": 964 + }, + { + "epoch": 0.17680469036277024, + "grad_norm": 0.450252173902017, + "learning_rate": 9.822381825142879e-06, + "loss": 0.506, + "step": 965 + }, + { + "epoch": 0.17698790765848296, + "grad_norm": 0.43561778776659554, + "learning_rate": 9.82153605889004e-06, + "loss": 0.5304, + "step": 966 + }, + { + "epoch": 0.17717112495419568, + "grad_norm": 0.4306520452468297, + "learning_rate": 9.820688320357823e-06, + "loss": 0.5112, + "step": 967 + }, + { + "epoch": 0.1773543422499084, + "grad_norm": 0.39680977809710444, + "learning_rate": 9.819838609893005e-06, + "loss": 0.4904, + "step": 968 + }, + { + "epoch": 0.17753755954562112, + "grad_norm": 0.4225898008101114, + "learning_rate": 9.81898692784316e-06, + "loss": 0.5137, + "step": 969 + }, + { + "epoch": 0.17772077684133383, + "grad_norm": 0.4267062187327317, + "learning_rate": 9.818133274556679e-06, + "loss": 0.517, + "step": 970 + }, + { + "epoch": 0.17790399413704655, + "grad_norm": 0.4450452397955529, + "learning_rate": 9.81727765038275e-06, + "loss": 0.4949, + "step": 971 + }, + { + "epoch": 0.17808721143275924, + "grad_norm": 0.42730228271929965, + "learning_rate": 9.816420055671374e-06, + "loss": 0.5013, + "step": 972 + }, + { + "epoch": 0.17827042872847196, + "grad_norm": 0.4421383626296405, + "learning_rate": 9.815560490773356e-06, + "loss": 0.4679, + "step": 973 + }, + { + "epoch": 0.17845364602418468, + "grad_norm": 0.43307265096730885, + "learning_rate": 9.814698956040305e-06, + "loss": 0.4976, + "step": 974 + }, + { + "epoch": 0.1786368633198974, + "grad_norm": 0.40376464412932705, + "learning_rate": 9.813835451824636e-06, + "loss": 0.5021, + "step": 975 + }, + { + "epoch": 0.1788200806156101, + "grad_norm": 0.44775776339538953, + "learning_rate": 9.812969978479573e-06, + "loss": 0.4824, + "step": 976 + }, + { + "epoch": 0.17900329791132283, + "grad_norm": 0.468036097947629, + "learning_rate": 9.812102536359142e-06, + "loss": 0.4966, + "step": 977 + }, + { + "epoch": 0.17918651520703555, + "grad_norm": 0.49832149697786493, + "learning_rate": 9.811233125818176e-06, + "loss": 0.4853, + "step": 978 + }, + { + "epoch": 0.17936973250274826, + "grad_norm": 0.44754723833693727, + "learning_rate": 9.810361747212313e-06, + "loss": 0.4999, + "step": 979 + }, + { + "epoch": 0.17955294979846098, + "grad_norm": 0.4747011696315986, + "learning_rate": 9.809488400897996e-06, + "loss": 0.5108, + "step": 980 + }, + { + "epoch": 0.1797361670941737, + "grad_norm": 0.4055842519238836, + "learning_rate": 9.808613087232473e-06, + "loss": 0.4786, + "step": 981 + }, + { + "epoch": 0.17991938438988642, + "grad_norm": 0.3971701443501286, + "learning_rate": 9.807735806573795e-06, + "loss": 0.4925, + "step": 982 + }, + { + "epoch": 0.18010260168559913, + "grad_norm": 0.3784480421793096, + "learning_rate": 9.806856559280819e-06, + "loss": 0.4762, + "step": 983 + }, + { + "epoch": 0.18028581898131182, + "grad_norm": 0.41465183173286063, + "learning_rate": 9.80597534571321e-06, + "loss": 0.5037, + "step": 984 + }, + { + "epoch": 0.18046903627702454, + "grad_norm": 0.4713550462839489, + "learning_rate": 9.80509216623143e-06, + "loss": 0.4816, + "step": 985 + }, + { + "epoch": 0.18065225357273726, + "grad_norm": 0.34923786309805893, + "learning_rate": 9.804207021196751e-06, + "loss": 0.4903, + "step": 986 + }, + { + "epoch": 0.18083547086844998, + "grad_norm": 0.4359203385963395, + "learning_rate": 9.803319910971248e-06, + "loss": 0.5252, + "step": 987 + }, + { + "epoch": 0.1810186881641627, + "grad_norm": 0.40286824933567683, + "learning_rate": 9.802430835917796e-06, + "loss": 0.5248, + "step": 988 + }, + { + "epoch": 0.1812019054598754, + "grad_norm": 0.44314680454994215, + "learning_rate": 9.801539796400078e-06, + "loss": 0.5071, + "step": 989 + }, + { + "epoch": 0.18138512275558813, + "grad_norm": 0.44074612784921385, + "learning_rate": 9.80064679278258e-06, + "loss": 0.5057, + "step": 990 + }, + { + "epoch": 0.18156834005130085, + "grad_norm": 0.40968360376086815, + "learning_rate": 9.799751825430592e-06, + "loss": 0.4919, + "step": 991 + }, + { + "epoch": 0.18175155734701356, + "grad_norm": 0.4204735967781832, + "learning_rate": 9.798854894710202e-06, + "loss": 0.546, + "step": 992 + }, + { + "epoch": 0.18193477464272628, + "grad_norm": 0.36210624498761007, + "learning_rate": 9.79795600098831e-06, + "loss": 0.4904, + "step": 993 + }, + { + "epoch": 0.182117991938439, + "grad_norm": 0.45193717949202633, + "learning_rate": 9.797055144632609e-06, + "loss": 0.5292, + "step": 994 + }, + { + "epoch": 0.18230120923415172, + "grad_norm": 0.415630496229026, + "learning_rate": 9.796152326011604e-06, + "loss": 0.533, + "step": 995 + }, + { + "epoch": 0.1824844265298644, + "grad_norm": 0.39372729416385926, + "learning_rate": 9.795247545494594e-06, + "loss": 0.5147, + "step": 996 + }, + { + "epoch": 0.18266764382557713, + "grad_norm": 0.43540226077312505, + "learning_rate": 9.794340803451692e-06, + "loss": 0.49, + "step": 997 + }, + { + "epoch": 0.18285086112128984, + "grad_norm": 0.39372791981341043, + "learning_rate": 9.7934321002538e-06, + "loss": 0.4776, + "step": 998 + }, + { + "epoch": 0.18303407841700256, + "grad_norm": 0.40381508343023165, + "learning_rate": 9.792521436272633e-06, + "loss": 0.4566, + "step": 999 + }, + { + "epoch": 0.18321729571271528, + "grad_norm": 0.47729065322343966, + "learning_rate": 9.791608811880702e-06, + "loss": 0.5013, + "step": 1000 + }, + { + "epoch": 0.183400513008428, + "grad_norm": 0.4569898838129068, + "learning_rate": 9.79069422745132e-06, + "loss": 0.4834, + "step": 1001 + }, + { + "epoch": 0.1835837303041407, + "grad_norm": 0.41137738030360954, + "learning_rate": 9.789777683358607e-06, + "loss": 0.4982, + "step": 1002 + }, + { + "epoch": 0.18376694759985343, + "grad_norm": 0.443462511432857, + "learning_rate": 9.788859179977478e-06, + "loss": 0.4822, + "step": 1003 + }, + { + "epoch": 0.18395016489556615, + "grad_norm": 0.3803483898936361, + "learning_rate": 9.787938717683654e-06, + "loss": 0.4781, + "step": 1004 + }, + { + "epoch": 0.18413338219127887, + "grad_norm": 0.4711405888417879, + "learning_rate": 9.787016296853657e-06, + "loss": 0.5104, + "step": 1005 + }, + { + "epoch": 0.18431659948699158, + "grad_norm": 0.42018309024133216, + "learning_rate": 9.786091917864807e-06, + "loss": 0.4922, + "step": 1006 + }, + { + "epoch": 0.18449981678270427, + "grad_norm": 0.46053661876076, + "learning_rate": 9.785165581095225e-06, + "loss": 0.4777, + "step": 1007 + }, + { + "epoch": 0.184683034078417, + "grad_norm": 0.38543837395940844, + "learning_rate": 9.78423728692384e-06, + "loss": 0.5053, + "step": 1008 + }, + { + "epoch": 0.1848662513741297, + "grad_norm": 0.3681107156627339, + "learning_rate": 9.783307035730375e-06, + "loss": 0.463, + "step": 1009 + }, + { + "epoch": 0.18504946866984243, + "grad_norm": 0.42988658038887384, + "learning_rate": 9.782374827895353e-06, + "loss": 0.5179, + "step": 1010 + }, + { + "epoch": 0.18523268596555514, + "grad_norm": 0.3985975808728344, + "learning_rate": 9.781440663800099e-06, + "loss": 0.4979, + "step": 1011 + }, + { + "epoch": 0.18541590326126786, + "grad_norm": 0.419869548881785, + "learning_rate": 9.780504543826741e-06, + "loss": 0.4952, + "step": 1012 + }, + { + "epoch": 0.18559912055698058, + "grad_norm": 0.4348607167141751, + "learning_rate": 9.779566468358205e-06, + "loss": 0.4999, + "step": 1013 + }, + { + "epoch": 0.1857823378526933, + "grad_norm": 0.43518629406794973, + "learning_rate": 9.778626437778214e-06, + "loss": 0.5028, + "step": 1014 + }, + { + "epoch": 0.18596555514840601, + "grad_norm": 0.4475968637904264, + "learning_rate": 9.777684452471296e-06, + "loss": 0.4997, + "step": 1015 + }, + { + "epoch": 0.18614877244411873, + "grad_norm": 0.4834533700606862, + "learning_rate": 9.776740512822773e-06, + "loss": 0.5139, + "step": 1016 + }, + { + "epoch": 0.18633198973983145, + "grad_norm": 0.49443428074486423, + "learning_rate": 9.775794619218773e-06, + "loss": 0.4818, + "step": 1017 + }, + { + "epoch": 0.18651520703554417, + "grad_norm": 0.42771715112146097, + "learning_rate": 9.774846772046216e-06, + "loss": 0.507, + "step": 1018 + }, + { + "epoch": 0.18669842433125686, + "grad_norm": 0.43776726147874706, + "learning_rate": 9.77389697169283e-06, + "loss": 0.5002, + "step": 1019 + }, + { + "epoch": 0.18688164162696957, + "grad_norm": 0.44345646316302056, + "learning_rate": 9.77294521854713e-06, + "loss": 0.4953, + "step": 1020 + }, + { + "epoch": 0.1870648589226823, + "grad_norm": 0.4199805356687226, + "learning_rate": 9.77199151299844e-06, + "loss": 0.491, + "step": 1021 + }, + { + "epoch": 0.187248076218395, + "grad_norm": 0.4157991523056247, + "learning_rate": 9.77103585543688e-06, + "loss": 0.4809, + "step": 1022 + }, + { + "epoch": 0.18743129351410773, + "grad_norm": 0.3883786524287937, + "learning_rate": 9.770078246253367e-06, + "loss": 0.4811, + "step": 1023 + }, + { + "epoch": 0.18761451080982045, + "grad_norm": 0.4460460362572771, + "learning_rate": 9.769118685839616e-06, + "loss": 0.4869, + "step": 1024 + }, + { + "epoch": 0.18779772810553316, + "grad_norm": 0.4153757045736682, + "learning_rate": 9.768157174588144e-06, + "loss": 0.4464, + "step": 1025 + }, + { + "epoch": 0.18798094540124588, + "grad_norm": 0.4146331433621528, + "learning_rate": 9.767193712892259e-06, + "loss": 0.4588, + "step": 1026 + }, + { + "epoch": 0.1881641626969586, + "grad_norm": 0.42852199415357106, + "learning_rate": 9.766228301146074e-06, + "loss": 0.5132, + "step": 1027 + }, + { + "epoch": 0.18834737999267132, + "grad_norm": 0.40617799593038223, + "learning_rate": 9.765260939744496e-06, + "loss": 0.4715, + "step": 1028 + }, + { + "epoch": 0.18853059728838403, + "grad_norm": 0.43057345885222986, + "learning_rate": 9.76429162908323e-06, + "loss": 0.4613, + "step": 1029 + }, + { + "epoch": 0.18871381458409675, + "grad_norm": 0.4287852170015339, + "learning_rate": 9.763320369558777e-06, + "loss": 0.5029, + "step": 1030 + }, + { + "epoch": 0.18889703187980944, + "grad_norm": 0.4338612990572842, + "learning_rate": 9.762347161568441e-06, + "loss": 0.51, + "step": 1031 + }, + { + "epoch": 0.18908024917552216, + "grad_norm": 0.6099286344029112, + "learning_rate": 9.761372005510315e-06, + "loss": 0.5176, + "step": 1032 + }, + { + "epoch": 0.18926346647123488, + "grad_norm": 0.4139699553573524, + "learning_rate": 9.760394901783294e-06, + "loss": 0.4803, + "step": 1033 + }, + { + "epoch": 0.1894466837669476, + "grad_norm": 0.4613287614032801, + "learning_rate": 9.759415850787068e-06, + "loss": 0.5114, + "step": 1034 + }, + { + "epoch": 0.1896299010626603, + "grad_norm": 0.43042214487008984, + "learning_rate": 9.758434852922124e-06, + "loss": 0.5049, + "step": 1035 + }, + { + "epoch": 0.18981311835837303, + "grad_norm": 0.4708254422812291, + "learning_rate": 9.757451908589746e-06, + "loss": 0.5466, + "step": 1036 + }, + { + "epoch": 0.18999633565408575, + "grad_norm": 0.4823877518080949, + "learning_rate": 9.756467018192013e-06, + "loss": 0.4851, + "step": 1037 + }, + { + "epoch": 0.19017955294979846, + "grad_norm": 0.49290249361142535, + "learning_rate": 9.7554801821318e-06, + "loss": 0.4915, + "step": 1038 + }, + { + "epoch": 0.19036277024551118, + "grad_norm": 0.5097570694075713, + "learning_rate": 9.754491400812779e-06, + "loss": 0.5195, + "step": 1039 + }, + { + "epoch": 0.1905459875412239, + "grad_norm": 0.41886567376911943, + "learning_rate": 9.753500674639417e-06, + "loss": 0.5133, + "step": 1040 + }, + { + "epoch": 0.19072920483693662, + "grad_norm": 0.48924775085841404, + "learning_rate": 9.752508004016976e-06, + "loss": 0.5174, + "step": 1041 + }, + { + "epoch": 0.19091242213264933, + "grad_norm": 0.44418461742112236, + "learning_rate": 9.751513389351517e-06, + "loss": 0.4711, + "step": 1042 + }, + { + "epoch": 0.19109563942836202, + "grad_norm": 0.4103673664900802, + "learning_rate": 9.75051683104989e-06, + "loss": 0.478, + "step": 1043 + }, + { + "epoch": 0.19127885672407474, + "grad_norm": 0.45394849049437136, + "learning_rate": 9.749518329519745e-06, + "loss": 0.5069, + "step": 1044 + }, + { + "epoch": 0.19146207401978746, + "grad_norm": 0.4841087058964896, + "learning_rate": 9.748517885169525e-06, + "loss": 0.5226, + "step": 1045 + }, + { + "epoch": 0.19164529131550018, + "grad_norm": 0.4138362206735341, + "learning_rate": 9.747515498408466e-06, + "loss": 0.5085, + "step": 1046 + }, + { + "epoch": 0.1918285086112129, + "grad_norm": 0.4267417097417006, + "learning_rate": 9.746511169646604e-06, + "loss": 0.5074, + "step": 1047 + }, + { + "epoch": 0.1920117259069256, + "grad_norm": 0.4397686656970245, + "learning_rate": 9.745504899294764e-06, + "loss": 0.5242, + "step": 1048 + }, + { + "epoch": 0.19219494320263833, + "grad_norm": 0.6027364595972364, + "learning_rate": 9.744496687764568e-06, + "loss": 0.5046, + "step": 1049 + }, + { + "epoch": 0.19237816049835105, + "grad_norm": 0.45130752478163744, + "learning_rate": 9.743486535468431e-06, + "loss": 0.4699, + "step": 1050 + }, + { + "epoch": 0.19256137779406376, + "grad_norm": 0.4971542747077622, + "learning_rate": 9.742474442819561e-06, + "loss": 0.4446, + "step": 1051 + }, + { + "epoch": 0.19274459508977648, + "grad_norm": 0.566649600088142, + "learning_rate": 9.741460410231962e-06, + "loss": 0.5531, + "step": 1052 + }, + { + "epoch": 0.1929278123854892, + "grad_norm": 0.44162327591540923, + "learning_rate": 9.74044443812043e-06, + "loss": 0.5194, + "step": 1053 + }, + { + "epoch": 0.19311102968120192, + "grad_norm": 0.40323605263226747, + "learning_rate": 9.739426526900555e-06, + "loss": 0.5013, + "step": 1054 + }, + { + "epoch": 0.1932942469769146, + "grad_norm": 0.45567902759630347, + "learning_rate": 9.738406676988721e-06, + "loss": 0.4397, + "step": 1055 + }, + { + "epoch": 0.19347746427262733, + "grad_norm": 0.4604601552443739, + "learning_rate": 9.737384888802099e-06, + "loss": 0.4888, + "step": 1056 + }, + { + "epoch": 0.19366068156834004, + "grad_norm": 0.3927639599518843, + "learning_rate": 9.736361162758665e-06, + "loss": 0.4632, + "step": 1057 + }, + { + "epoch": 0.19384389886405276, + "grad_norm": 0.40880550818933825, + "learning_rate": 9.735335499277176e-06, + "loss": 0.484, + "step": 1058 + }, + { + "epoch": 0.19402711615976548, + "grad_norm": 0.4813436727815402, + "learning_rate": 9.734307898777187e-06, + "loss": 0.5351, + "step": 1059 + }, + { + "epoch": 0.1942103334554782, + "grad_norm": 0.44626763914485296, + "learning_rate": 9.733278361679042e-06, + "loss": 0.4847, + "step": 1060 + }, + { + "epoch": 0.1943935507511909, + "grad_norm": 0.3999429153352251, + "learning_rate": 9.732246888403885e-06, + "loss": 0.4888, + "step": 1061 + }, + { + "epoch": 0.19457676804690363, + "grad_norm": 0.39498667028752193, + "learning_rate": 9.731213479373643e-06, + "loss": 0.4947, + "step": 1062 + }, + { + "epoch": 0.19475998534261635, + "grad_norm": 0.4096597841615662, + "learning_rate": 9.730178135011036e-06, + "loss": 0.4747, + "step": 1063 + }, + { + "epoch": 0.19494320263832907, + "grad_norm": 0.4279641584926225, + "learning_rate": 9.729140855739585e-06, + "loss": 0.5013, + "step": 1064 + }, + { + "epoch": 0.19512641993404178, + "grad_norm": 0.5191493361942445, + "learning_rate": 9.72810164198359e-06, + "loss": 0.4947, + "step": 1065 + }, + { + "epoch": 0.1953096372297545, + "grad_norm": 0.4196384047011208, + "learning_rate": 9.727060494168148e-06, + "loss": 0.4825, + "step": 1066 + }, + { + "epoch": 0.1954928545254672, + "grad_norm": 0.4345875246850262, + "learning_rate": 9.726017412719151e-06, + "loss": 0.4734, + "step": 1067 + }, + { + "epoch": 0.1956760718211799, + "grad_norm": 0.45563595275942326, + "learning_rate": 9.724972398063273e-06, + "loss": 0.5153, + "step": 1068 + }, + { + "epoch": 0.19585928911689263, + "grad_norm": 0.44886022596575603, + "learning_rate": 9.723925450627988e-06, + "loss": 0.478, + "step": 1069 + }, + { + "epoch": 0.19604250641260534, + "grad_norm": 0.413358196754176, + "learning_rate": 9.722876570841554e-06, + "loss": 0.5132, + "step": 1070 + }, + { + "epoch": 0.19622572370831806, + "grad_norm": 0.41859249060654946, + "learning_rate": 9.721825759133022e-06, + "loss": 0.4911, + "step": 1071 + }, + { + "epoch": 0.19640894100403078, + "grad_norm": 0.4261104024613703, + "learning_rate": 9.720773015932234e-06, + "loss": 0.512, + "step": 1072 + }, + { + "epoch": 0.1965921582997435, + "grad_norm": 0.41207466582794494, + "learning_rate": 9.71971834166982e-06, + "loss": 0.4739, + "step": 1073 + }, + { + "epoch": 0.19677537559545621, + "grad_norm": 0.4191753315427551, + "learning_rate": 9.718661736777202e-06, + "loss": 0.4906, + "step": 1074 + }, + { + "epoch": 0.19695859289116893, + "grad_norm": 0.42085342925013736, + "learning_rate": 9.71760320168659e-06, + "loss": 0.5164, + "step": 1075 + }, + { + "epoch": 0.19714181018688165, + "grad_norm": 0.4268987095250944, + "learning_rate": 9.716542736830985e-06, + "loss": 0.4861, + "step": 1076 + }, + { + "epoch": 0.19732502748259437, + "grad_norm": 0.42610930294181376, + "learning_rate": 9.715480342644177e-06, + "loss": 0.475, + "step": 1077 + }, + { + "epoch": 0.19750824477830708, + "grad_norm": 0.46458365169769283, + "learning_rate": 9.714416019560744e-06, + "loss": 0.4949, + "step": 1078 + }, + { + "epoch": 0.19769146207401977, + "grad_norm": 0.40388410318318285, + "learning_rate": 9.713349768016055e-06, + "loss": 0.4886, + "step": 1079 + }, + { + "epoch": 0.1978746793697325, + "grad_norm": 0.47359324461508207, + "learning_rate": 9.712281588446267e-06, + "loss": 0.5166, + "step": 1080 + }, + { + "epoch": 0.1980578966654452, + "grad_norm": 0.3996969974243518, + "learning_rate": 9.711211481288323e-06, + "loss": 0.5208, + "step": 1081 + }, + { + "epoch": 0.19824111396115793, + "grad_norm": 0.37701213979445153, + "learning_rate": 9.710139446979961e-06, + "loss": 0.5, + "step": 1082 + }, + { + "epoch": 0.19842433125687065, + "grad_norm": 0.43491339118601396, + "learning_rate": 9.7090654859597e-06, + "loss": 0.4899, + "step": 1083 + }, + { + "epoch": 0.19860754855258336, + "grad_norm": 0.4651922376126236, + "learning_rate": 9.707989598666852e-06, + "loss": 0.4684, + "step": 1084 + }, + { + "epoch": 0.19879076584829608, + "grad_norm": 0.41200327863289965, + "learning_rate": 9.706911785541515e-06, + "loss": 0.4408, + "step": 1085 + }, + { + "epoch": 0.1989739831440088, + "grad_norm": 0.48300341584043166, + "learning_rate": 9.705832047024574e-06, + "loss": 0.53, + "step": 1086 + }, + { + "epoch": 0.19915720043972152, + "grad_norm": 0.4403619842708917, + "learning_rate": 9.704750383557707e-06, + "loss": 0.5361, + "step": 1087 + }, + { + "epoch": 0.19934041773543423, + "grad_norm": 0.3719557224033517, + "learning_rate": 9.70366679558337e-06, + "loss": 0.4948, + "step": 1088 + }, + { + "epoch": 0.19952363503114695, + "grad_norm": 0.4176787500801207, + "learning_rate": 9.702581283544813e-06, + "loss": 0.5145, + "step": 1089 + }, + { + "epoch": 0.19970685232685967, + "grad_norm": 0.4221402861265115, + "learning_rate": 9.701493847886075e-06, + "loss": 0.5128, + "step": 1090 + }, + { + "epoch": 0.19989006962257236, + "grad_norm": 0.40690517337963633, + "learning_rate": 9.700404489051974e-06, + "loss": 0.4911, + "step": 1091 + }, + { + "epoch": 0.20007328691828508, + "grad_norm": 0.46227138813892915, + "learning_rate": 9.69931320748812e-06, + "loss": 0.5037, + "step": 1092 + }, + { + "epoch": 0.2002565042139978, + "grad_norm": 0.3860882899117114, + "learning_rate": 9.69822000364091e-06, + "loss": 0.5063, + "step": 1093 + }, + { + "epoch": 0.2004397215097105, + "grad_norm": 0.512213232549586, + "learning_rate": 9.697124877957524e-06, + "loss": 0.5019, + "step": 1094 + }, + { + "epoch": 0.20062293880542323, + "grad_norm": 0.455116276042089, + "learning_rate": 9.69602783088593e-06, + "loss": 0.5145, + "step": 1095 + }, + { + "epoch": 0.20080615610113595, + "grad_norm": 0.4251117291337367, + "learning_rate": 9.694928862874883e-06, + "loss": 0.5096, + "step": 1096 + }, + { + "epoch": 0.20098937339684866, + "grad_norm": 0.46240748219044453, + "learning_rate": 9.693827974373925e-06, + "loss": 0.4506, + "step": 1097 + }, + { + "epoch": 0.20117259069256138, + "grad_norm": 0.35673376148210156, + "learning_rate": 9.692725165833377e-06, + "loss": 0.4688, + "step": 1098 + }, + { + "epoch": 0.2013558079882741, + "grad_norm": 0.469083523866663, + "learning_rate": 9.69162043770435e-06, + "loss": 0.5049, + "step": 1099 + }, + { + "epoch": 0.20153902528398682, + "grad_norm": 0.42797732688089807, + "learning_rate": 9.690513790438743e-06, + "loss": 0.4696, + "step": 1100 + }, + { + "epoch": 0.20172224257969953, + "grad_norm": 0.4281881317554375, + "learning_rate": 9.689405224489235e-06, + "loss": 0.4973, + "step": 1101 + }, + { + "epoch": 0.20190545987541225, + "grad_norm": 0.4513286311031276, + "learning_rate": 9.688294740309292e-06, + "loss": 0.4694, + "step": 1102 + }, + { + "epoch": 0.20208867717112494, + "grad_norm": 0.3940963878288531, + "learning_rate": 9.687182338353166e-06, + "loss": 0.4765, + "step": 1103 + }, + { + "epoch": 0.20227189446683766, + "grad_norm": 0.3902749441842877, + "learning_rate": 9.68606801907589e-06, + "loss": 0.4759, + "step": 1104 + }, + { + "epoch": 0.20245511176255038, + "grad_norm": 0.4789649740688253, + "learning_rate": 9.684951782933282e-06, + "loss": 0.4948, + "step": 1105 + }, + { + "epoch": 0.2026383290582631, + "grad_norm": 0.47833293392492787, + "learning_rate": 9.683833630381949e-06, + "loss": 0.5239, + "step": 1106 + }, + { + "epoch": 0.2028215463539758, + "grad_norm": 0.38231892821814056, + "learning_rate": 9.682713561879275e-06, + "loss": 0.4958, + "step": 1107 + }, + { + "epoch": 0.20300476364968853, + "grad_norm": 0.37675479403578893, + "learning_rate": 9.681591577883433e-06, + "loss": 0.4778, + "step": 1108 + }, + { + "epoch": 0.20318798094540125, + "grad_norm": 0.3570523937587689, + "learning_rate": 9.680467678853378e-06, + "loss": 0.484, + "step": 1109 + }, + { + "epoch": 0.20337119824111397, + "grad_norm": 0.4245386369867795, + "learning_rate": 9.679341865248848e-06, + "loss": 0.4866, + "step": 1110 + }, + { + "epoch": 0.20355441553682668, + "grad_norm": 0.4405417241764804, + "learning_rate": 9.678214137530361e-06, + "loss": 0.4532, + "step": 1111 + }, + { + "epoch": 0.2037376328325394, + "grad_norm": 0.3965079787279517, + "learning_rate": 9.677084496159224e-06, + "loss": 0.4731, + "step": 1112 + }, + { + "epoch": 0.20392085012825212, + "grad_norm": 0.38208118269449826, + "learning_rate": 9.675952941597522e-06, + "loss": 0.4397, + "step": 1113 + }, + { + "epoch": 0.20410406742396484, + "grad_norm": 0.39488634039550174, + "learning_rate": 9.674819474308126e-06, + "loss": 0.4928, + "step": 1114 + }, + { + "epoch": 0.20428728471967753, + "grad_norm": 0.3739080933909842, + "learning_rate": 9.673684094754686e-06, + "loss": 0.486, + "step": 1115 + }, + { + "epoch": 0.20447050201539024, + "grad_norm": 0.3958666571427383, + "learning_rate": 9.672546803401638e-06, + "loss": 0.5102, + "step": 1116 + }, + { + "epoch": 0.20465371931110296, + "grad_norm": 1.2610206438333627, + "learning_rate": 9.671407600714197e-06, + "loss": 0.4555, + "step": 1117 + }, + { + "epoch": 0.20483693660681568, + "grad_norm": 0.4461796609741918, + "learning_rate": 9.670266487158363e-06, + "loss": 0.4718, + "step": 1118 + }, + { + "epoch": 0.2050201539025284, + "grad_norm": 0.43024748470939017, + "learning_rate": 9.669123463200914e-06, + "loss": 0.4953, + "step": 1119 + }, + { + "epoch": 0.2052033711982411, + "grad_norm": 0.44154226794099993, + "learning_rate": 9.66797852930941e-06, + "loss": 0.5181, + "step": 1120 + }, + { + "epoch": 0.20538658849395383, + "grad_norm": 0.41502499575208784, + "learning_rate": 9.666831685952194e-06, + "loss": 0.4772, + "step": 1121 + }, + { + "epoch": 0.20556980578966655, + "grad_norm": 0.4084436936730005, + "learning_rate": 9.66568293359839e-06, + "loss": 0.5047, + "step": 1122 + }, + { + "epoch": 0.20575302308537927, + "grad_norm": 0.43479037966133816, + "learning_rate": 9.664532272717902e-06, + "loss": 0.4698, + "step": 1123 + }, + { + "epoch": 0.20593624038109198, + "grad_norm": 0.7606951229520282, + "learning_rate": 9.663379703781414e-06, + "loss": 0.4841, + "step": 1124 + }, + { + "epoch": 0.2061194576768047, + "grad_norm": 0.4006994667526495, + "learning_rate": 9.662225227260395e-06, + "loss": 0.4963, + "step": 1125 + }, + { + "epoch": 0.20630267497251742, + "grad_norm": 0.39866531966563096, + "learning_rate": 9.661068843627088e-06, + "loss": 0.4777, + "step": 1126 + }, + { + "epoch": 0.2064858922682301, + "grad_norm": 0.4035785610200985, + "learning_rate": 9.65991055335452e-06, + "loss": 0.4666, + "step": 1127 + }, + { + "epoch": 0.20666910956394283, + "grad_norm": 0.438045344209569, + "learning_rate": 9.658750356916494e-06, + "loss": 0.4956, + "step": 1128 + }, + { + "epoch": 0.20685232685965554, + "grad_norm": 0.3835030416878391, + "learning_rate": 9.657588254787598e-06, + "loss": 0.4886, + "step": 1129 + }, + { + "epoch": 0.20703554415536826, + "grad_norm": 0.4044887736362754, + "learning_rate": 9.6564242474432e-06, + "loss": 0.4831, + "step": 1130 + }, + { + "epoch": 0.20721876145108098, + "grad_norm": 0.40816269155213675, + "learning_rate": 9.655258335359438e-06, + "loss": 0.497, + "step": 1131 + }, + { + "epoch": 0.2074019787467937, + "grad_norm": 0.39413831022845913, + "learning_rate": 9.65409051901324e-06, + "loss": 0.4588, + "step": 1132 + }, + { + "epoch": 0.20758519604250641, + "grad_norm": 0.5053059172212088, + "learning_rate": 9.652920798882307e-06, + "loss": 0.4985, + "step": 1133 + }, + { + "epoch": 0.20776841333821913, + "grad_norm": 0.3669036545586826, + "learning_rate": 9.651749175445123e-06, + "loss": 0.4917, + "step": 1134 + }, + { + "epoch": 0.20795163063393185, + "grad_norm": 0.4268516645401171, + "learning_rate": 9.650575649180944e-06, + "loss": 0.4849, + "step": 1135 + }, + { + "epoch": 0.20813484792964457, + "grad_norm": 0.4262534436550349, + "learning_rate": 9.649400220569812e-06, + "loss": 0.5128, + "step": 1136 + }, + { + "epoch": 0.20831806522535729, + "grad_norm": 0.3944076446197719, + "learning_rate": 9.648222890092538e-06, + "loss": 0.5014, + "step": 1137 + }, + { + "epoch": 0.20850128252107, + "grad_norm": 0.40237247723443686, + "learning_rate": 9.647043658230723e-06, + "loss": 0.5162, + "step": 1138 + }, + { + "epoch": 0.2086844998167827, + "grad_norm": 0.40871120340037176, + "learning_rate": 9.645862525466734e-06, + "loss": 0.5021, + "step": 1139 + }, + { + "epoch": 0.2088677171124954, + "grad_norm": 0.40979376980666626, + "learning_rate": 9.644679492283723e-06, + "loss": 0.5104, + "step": 1140 + }, + { + "epoch": 0.20905093440820813, + "grad_norm": 0.41263650327798096, + "learning_rate": 9.643494559165616e-06, + "loss": 0.4764, + "step": 1141 + }, + { + "epoch": 0.20923415170392085, + "grad_norm": 0.37127003400941044, + "learning_rate": 9.64230772659712e-06, + "loss": 0.4858, + "step": 1142 + }, + { + "epoch": 0.20941736899963356, + "grad_norm": 0.4556209336445142, + "learning_rate": 9.641118995063712e-06, + "loss": 0.4668, + "step": 1143 + }, + { + "epoch": 0.20960058629534628, + "grad_norm": 0.37966915346806135, + "learning_rate": 9.639928365051655e-06, + "loss": 0.4733, + "step": 1144 + }, + { + "epoch": 0.209783803591059, + "grad_norm": 0.5527182868106905, + "learning_rate": 9.638735837047976e-06, + "loss": 0.5182, + "step": 1145 + }, + { + "epoch": 0.20996702088677172, + "grad_norm": 0.37939914018444165, + "learning_rate": 9.637541411540496e-06, + "loss": 0.4866, + "step": 1146 + }, + { + "epoch": 0.21015023818248443, + "grad_norm": 0.44598863023348195, + "learning_rate": 9.636345089017795e-06, + "loss": 0.5115, + "step": 1147 + }, + { + "epoch": 0.21033345547819715, + "grad_norm": 0.4509291233105335, + "learning_rate": 9.635146869969239e-06, + "loss": 0.5099, + "step": 1148 + }, + { + "epoch": 0.21051667277390987, + "grad_norm": 0.43540452743790453, + "learning_rate": 9.633946754884963e-06, + "loss": 0.4971, + "step": 1149 + }, + { + "epoch": 0.21069989006962256, + "grad_norm": 0.43503175511293474, + "learning_rate": 9.63274474425589e-06, + "loss": 0.535, + "step": 1150 + }, + { + "epoch": 0.21088310736533528, + "grad_norm": 0.4284909564211116, + "learning_rate": 9.6315408385737e-06, + "loss": 0.5195, + "step": 1151 + }, + { + "epoch": 0.211066324661048, + "grad_norm": 0.4144953761451503, + "learning_rate": 9.630335038330867e-06, + "loss": 0.4917, + "step": 1152 + }, + { + "epoch": 0.2112495419567607, + "grad_norm": 0.4897343272100308, + "learning_rate": 9.629127344020625e-06, + "loss": 0.5315, + "step": 1153 + }, + { + "epoch": 0.21143275925247343, + "grad_norm": 0.422768220232635, + "learning_rate": 9.627917756136991e-06, + "loss": 0.4751, + "step": 1154 + }, + { + "epoch": 0.21161597654818615, + "grad_norm": 0.4834794267838247, + "learning_rate": 9.626706275174754e-06, + "loss": 0.49, + "step": 1155 + }, + { + "epoch": 0.21179919384389886, + "grad_norm": 0.43277895942455413, + "learning_rate": 9.625492901629478e-06, + "loss": 0.4942, + "step": 1156 + }, + { + "epoch": 0.21198241113961158, + "grad_norm": 0.4500403514457418, + "learning_rate": 9.624277635997503e-06, + "loss": 0.4794, + "step": 1157 + }, + { + "epoch": 0.2121656284353243, + "grad_norm": 0.4099295617842837, + "learning_rate": 9.62306047877594e-06, + "loss": 0.5079, + "step": 1158 + }, + { + "epoch": 0.21234884573103702, + "grad_norm": 0.39697880470360636, + "learning_rate": 9.62184143046267e-06, + "loss": 0.4718, + "step": 1159 + }, + { + "epoch": 0.21253206302674973, + "grad_norm": 0.45285900320929984, + "learning_rate": 9.620620491556359e-06, + "loss": 0.4815, + "step": 1160 + }, + { + "epoch": 0.21271528032246245, + "grad_norm": 0.4616778523748189, + "learning_rate": 9.619397662556434e-06, + "loss": 0.5072, + "step": 1161 + }, + { + "epoch": 0.21289849761817514, + "grad_norm": 0.42472446401669417, + "learning_rate": 9.618172943963107e-06, + "loss": 0.5205, + "step": 1162 + }, + { + "epoch": 0.21308171491388786, + "grad_norm": 1.1202188747857207, + "learning_rate": 9.61694633627735e-06, + "loss": 0.4941, + "step": 1163 + }, + { + "epoch": 0.21326493220960058, + "grad_norm": 0.4238555504122449, + "learning_rate": 9.615717840000922e-06, + "loss": 0.5259, + "step": 1164 + }, + { + "epoch": 0.2134481495053133, + "grad_norm": 0.4331940724685078, + "learning_rate": 9.614487455636341e-06, + "loss": 0.4792, + "step": 1165 + }, + { + "epoch": 0.213631366801026, + "grad_norm": 0.42864744785979175, + "learning_rate": 9.613255183686906e-06, + "loss": 0.4737, + "step": 1166 + }, + { + "epoch": 0.21381458409673873, + "grad_norm": 0.43365153166481973, + "learning_rate": 9.612021024656685e-06, + "loss": 0.4709, + "step": 1167 + }, + { + "epoch": 0.21399780139245145, + "grad_norm": 0.43747080354215073, + "learning_rate": 9.610784979050519e-06, + "loss": 0.5012, + "step": 1168 + }, + { + "epoch": 0.21418101868816417, + "grad_norm": 0.4204470389686437, + "learning_rate": 9.609547047374018e-06, + "loss": 0.4905, + "step": 1169 + }, + { + "epoch": 0.21436423598387688, + "grad_norm": 0.42512506549883267, + "learning_rate": 9.608307230133566e-06, + "loss": 0.4672, + "step": 1170 + }, + { + "epoch": 0.2145474532795896, + "grad_norm": 0.4326161922702061, + "learning_rate": 9.607065527836324e-06, + "loss": 0.5035, + "step": 1171 + }, + { + "epoch": 0.21473067057530232, + "grad_norm": 0.4012535242609492, + "learning_rate": 9.60582194099021e-06, + "loss": 0.5042, + "step": 1172 + }, + { + "epoch": 0.21491388787101504, + "grad_norm": 0.407044902400079, + "learning_rate": 9.604576470103923e-06, + "loss": 0.4751, + "step": 1173 + }, + { + "epoch": 0.21509710516672773, + "grad_norm": 0.420735767335274, + "learning_rate": 9.603329115686934e-06, + "loss": 0.4872, + "step": 1174 + }, + { + "epoch": 0.21528032246244044, + "grad_norm": 0.41386495522739447, + "learning_rate": 9.60207987824948e-06, + "loss": 0.4554, + "step": 1175 + }, + { + "epoch": 0.21546353975815316, + "grad_norm": 0.3993847396510676, + "learning_rate": 9.600828758302568e-06, + "loss": 0.4769, + "step": 1176 + }, + { + "epoch": 0.21564675705386588, + "grad_norm": 0.43412675596114314, + "learning_rate": 9.599575756357974e-06, + "loss": 0.4604, + "step": 1177 + }, + { + "epoch": 0.2158299743495786, + "grad_norm": 0.42400411471398847, + "learning_rate": 9.598320872928251e-06, + "loss": 0.4644, + "step": 1178 + }, + { + "epoch": 0.2160131916452913, + "grad_norm": 0.43220524123609294, + "learning_rate": 9.597064108526715e-06, + "loss": 0.4931, + "step": 1179 + }, + { + "epoch": 0.21619640894100403, + "grad_norm": 0.4139513466321143, + "learning_rate": 9.595805463667452e-06, + "loss": 0.502, + "step": 1180 + }, + { + "epoch": 0.21637962623671675, + "grad_norm": 0.5080108758300018, + "learning_rate": 9.59454493886532e-06, + "loss": 0.4825, + "step": 1181 + }, + { + "epoch": 0.21656284353242947, + "grad_norm": 0.47000114005402754, + "learning_rate": 9.593282534635945e-06, + "loss": 0.4678, + "step": 1182 + }, + { + "epoch": 0.21674606082814218, + "grad_norm": 0.42032814714737154, + "learning_rate": 9.59201825149572e-06, + "loss": 0.5086, + "step": 1183 + }, + { + "epoch": 0.2169292781238549, + "grad_norm": 0.4011425053622946, + "learning_rate": 9.590752089961811e-06, + "loss": 0.5067, + "step": 1184 + }, + { + "epoch": 0.21711249541956762, + "grad_norm": 0.38415361503458767, + "learning_rate": 9.589484050552142e-06, + "loss": 0.5192, + "step": 1185 + }, + { + "epoch": 0.2172957127152803, + "grad_norm": 0.4099189657711175, + "learning_rate": 9.588214133785421e-06, + "loss": 0.5087, + "step": 1186 + }, + { + "epoch": 0.21747893001099303, + "grad_norm": 0.4259010610337274, + "learning_rate": 9.58694234018111e-06, + "loss": 0.4931, + "step": 1187 + }, + { + "epoch": 0.21766214730670574, + "grad_norm": 0.4709131450449139, + "learning_rate": 9.585668670259446e-06, + "loss": 0.4949, + "step": 1188 + }, + { + "epoch": 0.21784536460241846, + "grad_norm": 0.44547955151410057, + "learning_rate": 9.584393124541431e-06, + "loss": 0.4771, + "step": 1189 + }, + { + "epoch": 0.21802858189813118, + "grad_norm": 0.43579531600042903, + "learning_rate": 9.583115703548835e-06, + "loss": 0.5121, + "step": 1190 + }, + { + "epoch": 0.2182117991938439, + "grad_norm": 0.3959951122458681, + "learning_rate": 9.581836407804196e-06, + "loss": 0.4739, + "step": 1191 + }, + { + "epoch": 0.21839501648955661, + "grad_norm": 0.39913718386402464, + "learning_rate": 9.580555237830817e-06, + "loss": 0.476, + "step": 1192 + }, + { + "epoch": 0.21857823378526933, + "grad_norm": 0.3981984945355583, + "learning_rate": 9.579272194152767e-06, + "loss": 0.5042, + "step": 1193 + }, + { + "epoch": 0.21876145108098205, + "grad_norm": 0.4369820172983135, + "learning_rate": 9.577987277294887e-06, + "loss": 0.5072, + "step": 1194 + }, + { + "epoch": 0.21894466837669477, + "grad_norm": 0.43452616563791835, + "learning_rate": 9.576700487782775e-06, + "loss": 0.5111, + "step": 1195 + }, + { + "epoch": 0.21912788567240749, + "grad_norm": 0.45908350014787985, + "learning_rate": 9.575411826142806e-06, + "loss": 0.5284, + "step": 1196 + }, + { + "epoch": 0.2193111029681202, + "grad_norm": 0.40541576693311915, + "learning_rate": 9.57412129290211e-06, + "loss": 0.465, + "step": 1197 + }, + { + "epoch": 0.2194943202638329, + "grad_norm": 0.3894120792636639, + "learning_rate": 9.572828888588593e-06, + "loss": 0.502, + "step": 1198 + }, + { + "epoch": 0.2196775375595456, + "grad_norm": 0.41214065106611036, + "learning_rate": 9.571534613730915e-06, + "loss": 0.4999, + "step": 1199 + }, + { + "epoch": 0.21986075485525833, + "grad_norm": 0.38064251771626856, + "learning_rate": 9.57023846885851e-06, + "loss": 0.485, + "step": 1200 + }, + { + "epoch": 0.22004397215097105, + "grad_norm": 0.3546653255179806, + "learning_rate": 9.568940454501578e-06, + "loss": 0.4575, + "step": 1201 + }, + { + "epoch": 0.22022718944668376, + "grad_norm": 0.40806335023375157, + "learning_rate": 9.567640571191073e-06, + "loss": 0.4516, + "step": 1202 + }, + { + "epoch": 0.22041040674239648, + "grad_norm": 0.4688612954629896, + "learning_rate": 9.566338819458726e-06, + "loss": 0.5134, + "step": 1203 + }, + { + "epoch": 0.2205936240381092, + "grad_norm": 0.39969548266649896, + "learning_rate": 9.56503519983702e-06, + "loss": 0.4847, + "step": 1204 + }, + { + "epoch": 0.22077684133382192, + "grad_norm": 0.4383715060312841, + "learning_rate": 9.563729712859216e-06, + "loss": 0.5254, + "step": 1205 + }, + { + "epoch": 0.22096005862953463, + "grad_norm": 0.4352285842560572, + "learning_rate": 9.562422359059328e-06, + "loss": 0.5016, + "step": 1206 + }, + { + "epoch": 0.22114327592524735, + "grad_norm": 0.39691391609356236, + "learning_rate": 9.561113138972138e-06, + "loss": 0.5089, + "step": 1207 + }, + { + "epoch": 0.22132649322096007, + "grad_norm": 0.3960087971203996, + "learning_rate": 9.55980205313319e-06, + "loss": 0.4977, + "step": 1208 + }, + { + "epoch": 0.2215097105166728, + "grad_norm": 0.4287362367803256, + "learning_rate": 9.558489102078792e-06, + "loss": 0.4641, + "step": 1209 + }, + { + "epoch": 0.22169292781238548, + "grad_norm": 0.3811089768117096, + "learning_rate": 9.557174286346014e-06, + "loss": 0.5213, + "step": 1210 + }, + { + "epoch": 0.2218761451080982, + "grad_norm": 0.4593900206484617, + "learning_rate": 9.555857606472692e-06, + "loss": 0.514, + "step": 1211 + }, + { + "epoch": 0.2220593624038109, + "grad_norm": 0.3711599555884036, + "learning_rate": 9.554539062997421e-06, + "loss": 0.5003, + "step": 1212 + }, + { + "epoch": 0.22224257969952363, + "grad_norm": 0.42622144015650515, + "learning_rate": 9.553218656459558e-06, + "loss": 0.5486, + "step": 1213 + }, + { + "epoch": 0.22242579699523635, + "grad_norm": 0.4167867779019379, + "learning_rate": 9.551896387399226e-06, + "loss": 0.5184, + "step": 1214 + }, + { + "epoch": 0.22260901429094906, + "grad_norm": 0.3809013711065984, + "learning_rate": 9.550572256357305e-06, + "loss": 0.5022, + "step": 1215 + }, + { + "epoch": 0.22279223158666178, + "grad_norm": 0.44157858433876945, + "learning_rate": 9.54924626387544e-06, + "loss": 0.4605, + "step": 1216 + }, + { + "epoch": 0.2229754488823745, + "grad_norm": 0.38145210565349935, + "learning_rate": 9.547918410496037e-06, + "loss": 0.497, + "step": 1217 + }, + { + "epoch": 0.22315866617808722, + "grad_norm": 0.4245361578656831, + "learning_rate": 9.546588696762262e-06, + "loss": 0.5249, + "step": 1218 + }, + { + "epoch": 0.22334188347379993, + "grad_norm": 0.3806441240590457, + "learning_rate": 9.545257123218043e-06, + "loss": 0.489, + "step": 1219 + }, + { + "epoch": 0.22352510076951265, + "grad_norm": 0.3771756284454525, + "learning_rate": 9.54392369040807e-06, + "loss": 0.4835, + "step": 1220 + }, + { + "epoch": 0.22370831806522537, + "grad_norm": 0.42014861145100263, + "learning_rate": 9.542588398877787e-06, + "loss": 0.4908, + "step": 1221 + }, + { + "epoch": 0.22389153536093806, + "grad_norm": 0.40603250481690667, + "learning_rate": 9.54125124917341e-06, + "loss": 0.4936, + "step": 1222 + }, + { + "epoch": 0.22407475265665078, + "grad_norm": 0.4061177784088164, + "learning_rate": 9.539912241841904e-06, + "loss": 0.4827, + "step": 1223 + }, + { + "epoch": 0.2242579699523635, + "grad_norm": 0.3861564878611609, + "learning_rate": 9.538571377431e-06, + "loss": 0.4676, + "step": 1224 + }, + { + "epoch": 0.2244411872480762, + "grad_norm": 0.42382201486067383, + "learning_rate": 9.537228656489187e-06, + "loss": 0.4602, + "step": 1225 + }, + { + "epoch": 0.22462440454378893, + "grad_norm": 0.3822471479641513, + "learning_rate": 9.53588407956571e-06, + "loss": 0.4508, + "step": 1226 + }, + { + "epoch": 0.22480762183950165, + "grad_norm": 0.3882035970294104, + "learning_rate": 9.534537647210582e-06, + "loss": 0.4723, + "step": 1227 + }, + { + "epoch": 0.22499083913521437, + "grad_norm": 0.46953161111998076, + "learning_rate": 9.533189359974564e-06, + "loss": 0.5037, + "step": 1228 + }, + { + "epoch": 0.22517405643092708, + "grad_norm": 0.3742186217573921, + "learning_rate": 9.531839218409186e-06, + "loss": 0.4821, + "step": 1229 + }, + { + "epoch": 0.2253572737266398, + "grad_norm": 0.42732489656890116, + "learning_rate": 9.53048722306673e-06, + "loss": 0.4959, + "step": 1230 + }, + { + "epoch": 0.22554049102235252, + "grad_norm": 0.36370123103326807, + "learning_rate": 9.52913337450024e-06, + "loss": 0.509, + "step": 1231 + }, + { + "epoch": 0.22572370831806524, + "grad_norm": 0.3842157959300981, + "learning_rate": 9.527777673263512e-06, + "loss": 0.4776, + "step": 1232 + }, + { + "epoch": 0.22590692561377795, + "grad_norm": 0.39958584358419963, + "learning_rate": 9.526420119911109e-06, + "loss": 0.4937, + "step": 1233 + }, + { + "epoch": 0.22609014290949064, + "grad_norm": 0.3795874511893427, + "learning_rate": 9.525060714998341e-06, + "loss": 0.4673, + "step": 1234 + }, + { + "epoch": 0.22627336020520336, + "grad_norm": 0.443615180266169, + "learning_rate": 9.523699459081285e-06, + "loss": 0.497, + "step": 1235 + }, + { + "epoch": 0.22645657750091608, + "grad_norm": 0.4126187255162815, + "learning_rate": 9.52233635271677e-06, + "loss": 0.5026, + "step": 1236 + }, + { + "epoch": 0.2266397947966288, + "grad_norm": 0.3898731986202734, + "learning_rate": 9.520971396462383e-06, + "loss": 0.5142, + "step": 1237 + }, + { + "epoch": 0.2268230120923415, + "grad_norm": 0.4303787774334329, + "learning_rate": 9.519604590876471e-06, + "loss": 0.5027, + "step": 1238 + }, + { + "epoch": 0.22700622938805423, + "grad_norm": 0.4881712350764513, + "learning_rate": 9.51823593651813e-06, + "loss": 0.5194, + "step": 1239 + }, + { + "epoch": 0.22718944668376695, + "grad_norm": 0.3958516733915845, + "learning_rate": 9.516865433947218e-06, + "loss": 0.4873, + "step": 1240 + }, + { + "epoch": 0.22737266397947967, + "grad_norm": 0.37588183496475464, + "learning_rate": 9.515493083724348e-06, + "loss": 0.483, + "step": 1241 + }, + { + "epoch": 0.22755588127519238, + "grad_norm": 0.42594083794952703, + "learning_rate": 9.514118886410889e-06, + "loss": 0.515, + "step": 1242 + }, + { + "epoch": 0.2277390985709051, + "grad_norm": 0.42418243334298944, + "learning_rate": 9.512742842568964e-06, + "loss": 0.5057, + "step": 1243 + }, + { + "epoch": 0.22792231586661782, + "grad_norm": 0.4402484021876601, + "learning_rate": 9.511364952761453e-06, + "loss": 0.4961, + "step": 1244 + }, + { + "epoch": 0.22810553316233054, + "grad_norm": 0.39304351029537393, + "learning_rate": 9.509985217551989e-06, + "loss": 0.4803, + "step": 1245 + }, + { + "epoch": 0.22828875045804323, + "grad_norm": 0.40248532246553786, + "learning_rate": 9.508603637504962e-06, + "loss": 0.4801, + "step": 1246 + }, + { + "epoch": 0.22847196775375594, + "grad_norm": 0.4302259083201261, + "learning_rate": 9.507220213185517e-06, + "loss": 0.5027, + "step": 1247 + }, + { + "epoch": 0.22865518504946866, + "grad_norm": 0.40571061563755584, + "learning_rate": 9.505834945159552e-06, + "loss": 0.5162, + "step": 1248 + }, + { + "epoch": 0.22883840234518138, + "grad_norm": 0.4032017209651049, + "learning_rate": 9.504447833993717e-06, + "loss": 0.4948, + "step": 1249 + }, + { + "epoch": 0.2290216196408941, + "grad_norm": 0.4195307039259119, + "learning_rate": 9.503058880255423e-06, + "loss": 0.5007, + "step": 1250 + }, + { + "epoch": 0.22920483693660682, + "grad_norm": 0.4677433036389937, + "learning_rate": 9.501668084512827e-06, + "loss": 0.4676, + "step": 1251 + }, + { + "epoch": 0.22938805423231953, + "grad_norm": 0.44203412318706997, + "learning_rate": 9.500275447334843e-06, + "loss": 0.5178, + "step": 1252 + }, + { + "epoch": 0.22957127152803225, + "grad_norm": 0.3745369415002761, + "learning_rate": 9.498880969291138e-06, + "loss": 0.4705, + "step": 1253 + }, + { + "epoch": 0.22975448882374497, + "grad_norm": 0.3760740011440762, + "learning_rate": 9.497484650952133e-06, + "loss": 0.4982, + "step": 1254 + }, + { + "epoch": 0.22993770611945769, + "grad_norm": 0.38594463821493263, + "learning_rate": 9.496086492889001e-06, + "loss": 0.4887, + "step": 1255 + }, + { + "epoch": 0.2301209234151704, + "grad_norm": 0.4212966089256434, + "learning_rate": 9.494686495673665e-06, + "loss": 0.4837, + "step": 1256 + }, + { + "epoch": 0.23030414071088312, + "grad_norm": 0.4210112942058116, + "learning_rate": 9.493284659878802e-06, + "loss": 0.517, + "step": 1257 + }, + { + "epoch": 0.2304873580065958, + "grad_norm": 0.41531125869001356, + "learning_rate": 9.491880986077846e-06, + "loss": 0.5263, + "step": 1258 + }, + { + "epoch": 0.23067057530230853, + "grad_norm": 0.46209059357171517, + "learning_rate": 9.490475474844976e-06, + "loss": 0.5232, + "step": 1259 + }, + { + "epoch": 0.23085379259802125, + "grad_norm": 0.4297726055183601, + "learning_rate": 9.489068126755124e-06, + "loss": 0.4613, + "step": 1260 + }, + { + "epoch": 0.23103700989373396, + "grad_norm": 0.46903664754890034, + "learning_rate": 9.487658942383975e-06, + "loss": 0.5076, + "step": 1261 + }, + { + "epoch": 0.23122022718944668, + "grad_norm": 0.48896965014560756, + "learning_rate": 9.486247922307967e-06, + "loss": 0.487, + "step": 1262 + }, + { + "epoch": 0.2314034444851594, + "grad_norm": 0.4218926363173936, + "learning_rate": 9.484835067104285e-06, + "loss": 0.4914, + "step": 1263 + }, + { + "epoch": 0.23158666178087212, + "grad_norm": 0.40940946307427944, + "learning_rate": 9.483420377350865e-06, + "loss": 0.4984, + "step": 1264 + }, + { + "epoch": 0.23176987907658483, + "grad_norm": 0.45293257509435014, + "learning_rate": 9.482003853626396e-06, + "loss": 0.5082, + "step": 1265 + }, + { + "epoch": 0.23195309637229755, + "grad_norm": 0.3935023699694669, + "learning_rate": 9.480585496510315e-06, + "loss": 0.5099, + "step": 1266 + }, + { + "epoch": 0.23213631366801027, + "grad_norm": 0.39198925832901294, + "learning_rate": 9.479165306582811e-06, + "loss": 0.4682, + "step": 1267 + }, + { + "epoch": 0.232319530963723, + "grad_norm": 0.34440630702814135, + "learning_rate": 9.47774328442482e-06, + "loss": 0.4835, + "step": 1268 + }, + { + "epoch": 0.2325027482594357, + "grad_norm": 0.4696525716200109, + "learning_rate": 9.476319430618033e-06, + "loss": 0.5098, + "step": 1269 + }, + { + "epoch": 0.2326859655551484, + "grad_norm": 0.4172646920959579, + "learning_rate": 9.47489374574488e-06, + "loss": 0.5023, + "step": 1270 + }, + { + "epoch": 0.2328691828508611, + "grad_norm": 0.3756156161424675, + "learning_rate": 9.473466230388552e-06, + "loss": 0.4754, + "step": 1271 + }, + { + "epoch": 0.23305240014657383, + "grad_norm": 1.9398173346161394, + "learning_rate": 9.472036885132979e-06, + "loss": 0.5082, + "step": 1272 + }, + { + "epoch": 0.23323561744228655, + "grad_norm": 0.47073412743230747, + "learning_rate": 9.470605710562845e-06, + "loss": 0.5285, + "step": 1273 + }, + { + "epoch": 0.23341883473799926, + "grad_norm": 0.38789836660988597, + "learning_rate": 9.469172707263582e-06, + "loss": 0.4925, + "step": 1274 + }, + { + "epoch": 0.23360205203371198, + "grad_norm": 0.3825601936653152, + "learning_rate": 9.467737875821368e-06, + "loss": 0.4659, + "step": 1275 + }, + { + "epoch": 0.2337852693294247, + "grad_norm": 0.3973414450558941, + "learning_rate": 9.46630121682313e-06, + "loss": 0.4772, + "step": 1276 + }, + { + "epoch": 0.23396848662513742, + "grad_norm": 0.40322202930522744, + "learning_rate": 9.464862730856542e-06, + "loss": 0.4832, + "step": 1277 + }, + { + "epoch": 0.23415170392085013, + "grad_norm": 0.42699100304138604, + "learning_rate": 9.463422418510024e-06, + "loss": 0.5212, + "step": 1278 + }, + { + "epoch": 0.23433492121656285, + "grad_norm": 0.42176298384733996, + "learning_rate": 9.461980280372748e-06, + "loss": 0.4669, + "step": 1279 + }, + { + "epoch": 0.23451813851227557, + "grad_norm": 0.4081443681151287, + "learning_rate": 9.460536317034627e-06, + "loss": 0.5113, + "step": 1280 + }, + { + "epoch": 0.2347013558079883, + "grad_norm": 0.38795147083032466, + "learning_rate": 9.459090529086325e-06, + "loss": 0.4392, + "step": 1281 + }, + { + "epoch": 0.23488457310370098, + "grad_norm": 0.5217298258213892, + "learning_rate": 9.457642917119249e-06, + "loss": 0.5181, + "step": 1282 + }, + { + "epoch": 0.2350677903994137, + "grad_norm": 0.42457976380897766, + "learning_rate": 9.456193481725555e-06, + "loss": 0.5034, + "step": 1283 + }, + { + "epoch": 0.2352510076951264, + "grad_norm": 0.37057304524699214, + "learning_rate": 9.454742223498145e-06, + "loss": 0.4502, + "step": 1284 + }, + { + "epoch": 0.23543422499083913, + "grad_norm": 0.3946439030149264, + "learning_rate": 9.453289143030662e-06, + "loss": 0.4777, + "step": 1285 + }, + { + "epoch": 0.23561744228655185, + "grad_norm": 0.44885227772255887, + "learning_rate": 9.451834240917498e-06, + "loss": 0.4847, + "step": 1286 + }, + { + "epoch": 0.23580065958226457, + "grad_norm": 0.43519145234243767, + "learning_rate": 9.45037751775379e-06, + "loss": 0.5088, + "step": 1287 + }, + { + "epoch": 0.23598387687797728, + "grad_norm": 0.4309093065470007, + "learning_rate": 9.448918974135424e-06, + "loss": 0.4677, + "step": 1288 + }, + { + "epoch": 0.23616709417369, + "grad_norm": 0.47978496271668625, + "learning_rate": 9.447458610659019e-06, + "loss": 0.5116, + "step": 1289 + }, + { + "epoch": 0.23635031146940272, + "grad_norm": 0.41021471445709895, + "learning_rate": 9.445996427921951e-06, + "loss": 0.523, + "step": 1290 + }, + { + "epoch": 0.23653352876511544, + "grad_norm": 0.4296171869446803, + "learning_rate": 9.444532426522334e-06, + "loss": 0.4691, + "step": 1291 + }, + { + "epoch": 0.23671674606082815, + "grad_norm": 0.4077954317933853, + "learning_rate": 9.443066607059026e-06, + "loss": 0.4772, + "step": 1292 + }, + { + "epoch": 0.23689996335654084, + "grad_norm": 0.35793769775976464, + "learning_rate": 9.44159897013163e-06, + "loss": 0.4924, + "step": 1293 + }, + { + "epoch": 0.23708318065225356, + "grad_norm": 0.39909841974854016, + "learning_rate": 9.440129516340492e-06, + "loss": 0.4823, + "step": 1294 + }, + { + "epoch": 0.23726639794796628, + "grad_norm": 0.3675254686648848, + "learning_rate": 9.4386582462867e-06, + "loss": 0.506, + "step": 1295 + }, + { + "epoch": 0.237449615243679, + "grad_norm": 0.37951723029275103, + "learning_rate": 9.43718516057209e-06, + "loss": 0.4996, + "step": 1296 + }, + { + "epoch": 0.23763283253939171, + "grad_norm": 0.4395625646258237, + "learning_rate": 9.435710259799234e-06, + "loss": 0.4841, + "step": 1297 + }, + { + "epoch": 0.23781604983510443, + "grad_norm": 0.4409131115143356, + "learning_rate": 9.434233544571446e-06, + "loss": 0.524, + "step": 1298 + }, + { + "epoch": 0.23799926713081715, + "grad_norm": 0.38035953245650445, + "learning_rate": 9.432755015492794e-06, + "loss": 0.4833, + "step": 1299 + }, + { + "epoch": 0.23818248442652987, + "grad_norm": 0.37603352556682307, + "learning_rate": 9.431274673168072e-06, + "loss": 0.4771, + "step": 1300 + }, + { + "epoch": 0.23836570172224258, + "grad_norm": 0.3973590263816894, + "learning_rate": 9.429792518202826e-06, + "loss": 0.4519, + "step": 1301 + }, + { + "epoch": 0.2385489190179553, + "grad_norm": 0.46279982818002663, + "learning_rate": 9.428308551203342e-06, + "loss": 0.5067, + "step": 1302 + }, + { + "epoch": 0.23873213631366802, + "grad_norm": 0.44423083147602355, + "learning_rate": 9.426822772776645e-06, + "loss": 0.4955, + "step": 1303 + }, + { + "epoch": 0.23891535360938074, + "grad_norm": 0.3765473227250332, + "learning_rate": 9.425335183530501e-06, + "loss": 0.4688, + "step": 1304 + }, + { + "epoch": 0.23909857090509343, + "grad_norm": 0.38352159262743835, + "learning_rate": 9.42384578407342e-06, + "loss": 0.5013, + "step": 1305 + }, + { + "epoch": 0.23928178820080614, + "grad_norm": 0.38474924269669325, + "learning_rate": 9.422354575014644e-06, + "loss": 0.5259, + "step": 1306 + }, + { + "epoch": 0.23946500549651886, + "grad_norm": 0.3764815533565891, + "learning_rate": 9.42086155696417e-06, + "loss": 0.4907, + "step": 1307 + }, + { + "epoch": 0.23964822279223158, + "grad_norm": 0.40120663442717996, + "learning_rate": 9.41936673053272e-06, + "loss": 0.5082, + "step": 1308 + }, + { + "epoch": 0.2398314400879443, + "grad_norm": 0.38410734753609116, + "learning_rate": 9.417870096331764e-06, + "loss": 0.5235, + "step": 1309 + }, + { + "epoch": 0.24001465738365702, + "grad_norm": 0.4824543008653062, + "learning_rate": 9.416371654973513e-06, + "loss": 0.5306, + "step": 1310 + }, + { + "epoch": 0.24019787467936973, + "grad_norm": 0.4448880086308854, + "learning_rate": 9.414871407070906e-06, + "loss": 0.5045, + "step": 1311 + }, + { + "epoch": 0.24038109197508245, + "grad_norm": 0.3818507137941972, + "learning_rate": 9.413369353237637e-06, + "loss": 0.4903, + "step": 1312 + }, + { + "epoch": 0.24056430927079517, + "grad_norm": 0.42284475981294684, + "learning_rate": 9.411865494088124e-06, + "loss": 0.4964, + "step": 1313 + }, + { + "epoch": 0.24074752656650789, + "grad_norm": 0.39875491330187324, + "learning_rate": 9.410359830237534e-06, + "loss": 0.5081, + "step": 1314 + }, + { + "epoch": 0.2409307438622206, + "grad_norm": 0.3905340354403746, + "learning_rate": 9.408852362301768e-06, + "loss": 0.5006, + "step": 1315 + }, + { + "epoch": 0.24111396115793332, + "grad_norm": 0.5340945724213095, + "learning_rate": 9.407343090897464e-06, + "loss": 0.5028, + "step": 1316 + }, + { + "epoch": 0.241297178453646, + "grad_norm": 0.43175601981581013, + "learning_rate": 9.405832016641997e-06, + "loss": 0.5053, + "step": 1317 + }, + { + "epoch": 0.24148039574935873, + "grad_norm": 0.4128194078123402, + "learning_rate": 9.404319140153484e-06, + "loss": 0.446, + "step": 1318 + }, + { + "epoch": 0.24166361304507145, + "grad_norm": 0.3673682428697322, + "learning_rate": 9.402804462050776e-06, + "loss": 0.4664, + "step": 1319 + }, + { + "epoch": 0.24184683034078416, + "grad_norm": 0.35927179937888154, + "learning_rate": 9.401287982953462e-06, + "loss": 0.4772, + "step": 1320 + }, + { + "epoch": 0.24203004763649688, + "grad_norm": 0.436824106029447, + "learning_rate": 9.399769703481865e-06, + "loss": 0.4973, + "step": 1321 + }, + { + "epoch": 0.2422132649322096, + "grad_norm": 0.44788915497504966, + "learning_rate": 9.39824962425705e-06, + "loss": 0.5297, + "step": 1322 + }, + { + "epoch": 0.24239648222792232, + "grad_norm": 0.43045640486065784, + "learning_rate": 9.396727745900811e-06, + "loss": 0.4906, + "step": 1323 + }, + { + "epoch": 0.24257969952363503, + "grad_norm": 0.5421905192174398, + "learning_rate": 9.395204069035686e-06, + "loss": 0.4659, + "step": 1324 + }, + { + "epoch": 0.24276291681934775, + "grad_norm": 0.4219323385690796, + "learning_rate": 9.39367859428494e-06, + "loss": 0.5003, + "step": 1325 + }, + { + "epoch": 0.24294613411506047, + "grad_norm": 0.4119758712615104, + "learning_rate": 9.392151322272583e-06, + "loss": 0.499, + "step": 1326 + }, + { + "epoch": 0.2431293514107732, + "grad_norm": 0.4287819235082752, + "learning_rate": 9.390622253623353e-06, + "loss": 0.4939, + "step": 1327 + }, + { + "epoch": 0.2433125687064859, + "grad_norm": 0.38055263240566445, + "learning_rate": 9.389091388962723e-06, + "loss": 0.4712, + "step": 1328 + }, + { + "epoch": 0.2434957860021986, + "grad_norm": 0.36009711673291733, + "learning_rate": 9.387558728916905e-06, + "loss": 0.4728, + "step": 1329 + }, + { + "epoch": 0.2436790032979113, + "grad_norm": 0.43571235014716525, + "learning_rate": 9.386024274112842e-06, + "loss": 0.4821, + "step": 1330 + }, + { + "epoch": 0.24386222059362403, + "grad_norm": 0.3659773293799275, + "learning_rate": 9.384488025178214e-06, + "loss": 0.4572, + "step": 1331 + }, + { + "epoch": 0.24404543788933675, + "grad_norm": 0.37704886297840073, + "learning_rate": 9.382949982741429e-06, + "loss": 0.4336, + "step": 1332 + }, + { + "epoch": 0.24422865518504946, + "grad_norm": 0.4586883760808847, + "learning_rate": 9.38141014743164e-06, + "loss": 0.4783, + "step": 1333 + }, + { + "epoch": 0.24441187248076218, + "grad_norm": 0.5054332600007021, + "learning_rate": 9.379868519878718e-06, + "loss": 0.4883, + "step": 1334 + }, + { + "epoch": 0.2445950897764749, + "grad_norm": 0.3980887593736299, + "learning_rate": 9.378325100713283e-06, + "loss": 0.4865, + "step": 1335 + }, + { + "epoch": 0.24477830707218762, + "grad_norm": 0.4215392454714871, + "learning_rate": 9.376779890566675e-06, + "loss": 0.5154, + "step": 1336 + }, + { + "epoch": 0.24496152436790034, + "grad_norm": 0.42292016972151214, + "learning_rate": 9.375232890070973e-06, + "loss": 0.5184, + "step": 1337 + }, + { + "epoch": 0.24514474166361305, + "grad_norm": 0.38508367690050965, + "learning_rate": 9.373684099858989e-06, + "loss": 0.5014, + "step": 1338 + }, + { + "epoch": 0.24532795895932577, + "grad_norm": 0.4481685963998327, + "learning_rate": 9.372133520564264e-06, + "loss": 0.4831, + "step": 1339 + }, + { + "epoch": 0.2455111762550385, + "grad_norm": 0.5247732605935221, + "learning_rate": 9.37058115282107e-06, + "loss": 0.5238, + "step": 1340 + }, + { + "epoch": 0.24569439355075118, + "grad_norm": 0.3817180851569393, + "learning_rate": 9.369026997264417e-06, + "loss": 0.4513, + "step": 1341 + }, + { + "epoch": 0.2458776108464639, + "grad_norm": 0.41437756480910404, + "learning_rate": 9.36747105453004e-06, + "loss": 0.4692, + "step": 1342 + }, + { + "epoch": 0.2460608281421766, + "grad_norm": 0.45961904721533503, + "learning_rate": 9.365913325254406e-06, + "loss": 0.4824, + "step": 1343 + }, + { + "epoch": 0.24624404543788933, + "grad_norm": 0.411323007740071, + "learning_rate": 9.364353810074716e-06, + "loss": 0.4767, + "step": 1344 + }, + { + "epoch": 0.24642726273360205, + "grad_norm": 0.40884023238234113, + "learning_rate": 9.362792509628897e-06, + "loss": 0.5129, + "step": 1345 + }, + { + "epoch": 0.24661048002931477, + "grad_norm": 0.4143220748874932, + "learning_rate": 9.361229424555609e-06, + "loss": 0.4969, + "step": 1346 + }, + { + "epoch": 0.24679369732502748, + "grad_norm": 0.42312617400069397, + "learning_rate": 9.359664555494244e-06, + "loss": 0.4586, + "step": 1347 + }, + { + "epoch": 0.2469769146207402, + "grad_norm": 0.476793297997145, + "learning_rate": 9.35809790308492e-06, + "loss": 0.4933, + "step": 1348 + }, + { + "epoch": 0.24716013191645292, + "grad_norm": 0.42441332538018967, + "learning_rate": 9.356529467968485e-06, + "loss": 0.4588, + "step": 1349 + }, + { + "epoch": 0.24734334921216564, + "grad_norm": 0.39138693081487125, + "learning_rate": 9.354959250786519e-06, + "loss": 0.4602, + "step": 1350 + }, + { + "epoch": 0.24752656650787835, + "grad_norm": 0.44700626066197013, + "learning_rate": 9.353387252181328e-06, + "loss": 0.4954, + "step": 1351 + }, + { + "epoch": 0.24770978380359107, + "grad_norm": 0.5055157069210112, + "learning_rate": 9.351813472795947e-06, + "loss": 0.4796, + "step": 1352 + }, + { + "epoch": 0.24789300109930376, + "grad_norm": 0.45338574174735713, + "learning_rate": 9.350237913274143e-06, + "loss": 0.5119, + "step": 1353 + }, + { + "epoch": 0.24807621839501648, + "grad_norm": 0.392746885183511, + "learning_rate": 9.348660574260406e-06, + "loss": 0.514, + "step": 1354 + }, + { + "epoch": 0.2482594356907292, + "grad_norm": 0.4164788422614105, + "learning_rate": 9.347081456399958e-06, + "loss": 0.5089, + "step": 1355 + }, + { + "epoch": 0.24844265298644191, + "grad_norm": 0.7070530239591771, + "learning_rate": 9.345500560338745e-06, + "loss": 0.4702, + "step": 1356 + }, + { + "epoch": 0.24862587028215463, + "grad_norm": 0.4394194345754478, + "learning_rate": 9.343917886723444e-06, + "loss": 0.477, + "step": 1357 + }, + { + "epoch": 0.24880908757786735, + "grad_norm": 0.42911702918417705, + "learning_rate": 9.342333436201457e-06, + "loss": 0.4923, + "step": 1358 + }, + { + "epoch": 0.24899230487358007, + "grad_norm": 0.4470370063218885, + "learning_rate": 9.340747209420913e-06, + "loss": 0.5228, + "step": 1359 + }, + { + "epoch": 0.24917552216929278, + "grad_norm": 0.4599577217037793, + "learning_rate": 9.339159207030668e-06, + "loss": 0.5182, + "step": 1360 + }, + { + "epoch": 0.2493587394650055, + "grad_norm": 0.4405760361683581, + "learning_rate": 9.337569429680306e-06, + "loss": 0.5319, + "step": 1361 + }, + { + "epoch": 0.24954195676071822, + "grad_norm": 0.4164050414490823, + "learning_rate": 9.335977878020132e-06, + "loss": 0.5231, + "step": 1362 + }, + { + "epoch": 0.24972517405643094, + "grad_norm": 0.4968041900926753, + "learning_rate": 9.334384552701183e-06, + "loss": 0.5057, + "step": 1363 + }, + { + "epoch": 0.24990839135214366, + "grad_norm": 0.40916476625199594, + "learning_rate": 9.332789454375219e-06, + "loss": 0.4936, + "step": 1364 + }, + { + "epoch": 0.25009160864785634, + "grad_norm": 0.40707879703628913, + "learning_rate": 9.331192583694722e-06, + "loss": 0.4911, + "step": 1365 + }, + { + "epoch": 0.2502748259435691, + "grad_norm": 0.36742835933618884, + "learning_rate": 9.329593941312904e-06, + "loss": 0.4697, + "step": 1366 + }, + { + "epoch": 0.2504580432392818, + "grad_norm": 0.417024876859677, + "learning_rate": 9.327993527883698e-06, + "loss": 0.4726, + "step": 1367 + }, + { + "epoch": 0.2506412605349945, + "grad_norm": 0.4039071282595343, + "learning_rate": 9.326391344061767e-06, + "loss": 0.4545, + "step": 1368 + }, + { + "epoch": 0.2508244778307072, + "grad_norm": 0.5042927537180502, + "learning_rate": 9.32478739050249e-06, + "loss": 0.4994, + "step": 1369 + }, + { + "epoch": 0.25100769512641996, + "grad_norm": 0.43157899355551005, + "learning_rate": 9.323181667861976e-06, + "loss": 0.5033, + "step": 1370 + }, + { + "epoch": 0.25119091242213265, + "grad_norm": 0.4065168380531392, + "learning_rate": 9.321574176797055e-06, + "loss": 0.4767, + "step": 1371 + }, + { + "epoch": 0.25137412971784534, + "grad_norm": 0.3698342286632864, + "learning_rate": 9.31996491796528e-06, + "loss": 0.4863, + "step": 1372 + }, + { + "epoch": 0.2515573470135581, + "grad_norm": 0.43408179894940496, + "learning_rate": 9.318353892024934e-06, + "loss": 0.505, + "step": 1373 + }, + { + "epoch": 0.2517405643092708, + "grad_norm": 0.39467414262175576, + "learning_rate": 9.316741099635012e-06, + "loss": 0.4922, + "step": 1374 + }, + { + "epoch": 0.2519237816049835, + "grad_norm": 0.37099615952376014, + "learning_rate": 9.315126541455237e-06, + "loss": 0.4723, + "step": 1375 + }, + { + "epoch": 0.2521069989006962, + "grad_norm": 0.40654704513015383, + "learning_rate": 9.313510218146055e-06, + "loss": 0.4858, + "step": 1376 + }, + { + "epoch": 0.25229021619640896, + "grad_norm": 0.4320196580766476, + "learning_rate": 9.311892130368636e-06, + "loss": 0.4773, + "step": 1377 + }, + { + "epoch": 0.25247343349212165, + "grad_norm": 0.4649358511506784, + "learning_rate": 9.310272278784865e-06, + "loss": 0.506, + "step": 1378 + }, + { + "epoch": 0.2526566507878344, + "grad_norm": 0.37417265682881873, + "learning_rate": 9.308650664057352e-06, + "loss": 0.474, + "step": 1379 + }, + { + "epoch": 0.2528398680835471, + "grad_norm": 0.3877983907225597, + "learning_rate": 9.307027286849434e-06, + "loss": 0.5005, + "step": 1380 + }, + { + "epoch": 0.2530230853792598, + "grad_norm": 0.3604568123656264, + "learning_rate": 9.305402147825155e-06, + "loss": 0.4931, + "step": 1381 + }, + { + "epoch": 0.2532063026749725, + "grad_norm": 0.4076431370889414, + "learning_rate": 9.303775247649297e-06, + "loss": 0.4781, + "step": 1382 + }, + { + "epoch": 0.2533895199706852, + "grad_norm": 0.4125540102911716, + "learning_rate": 9.302146586987347e-06, + "loss": 0.489, + "step": 1383 + }, + { + "epoch": 0.25357273726639795, + "grad_norm": 0.38688511296298134, + "learning_rate": 9.300516166505523e-06, + "loss": 0.4997, + "step": 1384 + }, + { + "epoch": 0.25375595456211064, + "grad_norm": 0.3901423062453873, + "learning_rate": 9.298883986870757e-06, + "loss": 0.4599, + "step": 1385 + }, + { + "epoch": 0.2539391718578234, + "grad_norm": 0.3963777690852405, + "learning_rate": 9.297250048750702e-06, + "loss": 0.5013, + "step": 1386 + }, + { + "epoch": 0.2541223891535361, + "grad_norm": 0.42814597983549907, + "learning_rate": 9.295614352813732e-06, + "loss": 0.4841, + "step": 1387 + }, + { + "epoch": 0.2543056064492488, + "grad_norm": 0.4342615131111368, + "learning_rate": 9.293976899728936e-06, + "loss": 0.4893, + "step": 1388 + }, + { + "epoch": 0.2544888237449615, + "grad_norm": 0.3801898878490571, + "learning_rate": 9.292337690166127e-06, + "loss": 0.4921, + "step": 1389 + }, + { + "epoch": 0.25467204104067426, + "grad_norm": 0.41610528591828644, + "learning_rate": 9.290696724795833e-06, + "loss": 0.5179, + "step": 1390 + }, + { + "epoch": 0.25485525833638695, + "grad_norm": 0.4071057844053913, + "learning_rate": 9.289054004289302e-06, + "loss": 0.463, + "step": 1391 + }, + { + "epoch": 0.2550384756320997, + "grad_norm": 0.44828293293160687, + "learning_rate": 9.287409529318497e-06, + "loss": 0.4899, + "step": 1392 + }, + { + "epoch": 0.2552216929278124, + "grad_norm": 0.4044272763220796, + "learning_rate": 9.285763300556101e-06, + "loss": 0.5114, + "step": 1393 + }, + { + "epoch": 0.25540491022352513, + "grad_norm": 0.4027954029425126, + "learning_rate": 9.284115318675515e-06, + "loss": 0.4706, + "step": 1394 + }, + { + "epoch": 0.2555881275192378, + "grad_norm": 0.4318050133048906, + "learning_rate": 9.282465584350856e-06, + "loss": 0.4931, + "step": 1395 + }, + { + "epoch": 0.2557713448149505, + "grad_norm": 0.39621664861383626, + "learning_rate": 9.280814098256961e-06, + "loss": 0.4561, + "step": 1396 + }, + { + "epoch": 0.25595456211066325, + "grad_norm": 0.4720751749880995, + "learning_rate": 9.279160861069376e-06, + "loss": 0.4889, + "step": 1397 + }, + { + "epoch": 0.25613777940637594, + "grad_norm": 0.47784125077057304, + "learning_rate": 9.277505873464369e-06, + "loss": 0.5292, + "step": 1398 + }, + { + "epoch": 0.2563209967020887, + "grad_norm": 0.48714587413048666, + "learning_rate": 9.275849136118926e-06, + "loss": 0.4785, + "step": 1399 + }, + { + "epoch": 0.2565042139978014, + "grad_norm": 0.403984192319981, + "learning_rate": 9.274190649710743e-06, + "loss": 0.5076, + "step": 1400 + }, + { + "epoch": 0.2566874312935141, + "grad_norm": 0.39832914408918374, + "learning_rate": 9.272530414918236e-06, + "loss": 0.508, + "step": 1401 + }, + { + "epoch": 0.2568706485892268, + "grad_norm": 0.4345591492942482, + "learning_rate": 9.270868432420532e-06, + "loss": 0.434, + "step": 1402 + }, + { + "epoch": 0.25705386588493956, + "grad_norm": 0.38575416668818974, + "learning_rate": 9.269204702897476e-06, + "loss": 0.5123, + "step": 1403 + }, + { + "epoch": 0.25723708318065225, + "grad_norm": 0.46248145640766813, + "learning_rate": 9.267539227029628e-06, + "loss": 0.462, + "step": 1404 + }, + { + "epoch": 0.257420300476365, + "grad_norm": 0.41624325170317916, + "learning_rate": 9.265872005498263e-06, + "loss": 0.5224, + "step": 1405 + }, + { + "epoch": 0.2576035177720777, + "grad_norm": 0.45537999223861836, + "learning_rate": 9.264203038985362e-06, + "loss": 0.4863, + "step": 1406 + }, + { + "epoch": 0.2577867350677904, + "grad_norm": 0.41647722499814044, + "learning_rate": 9.262532328173633e-06, + "loss": 0.498, + "step": 1407 + }, + { + "epoch": 0.2579699523635031, + "grad_norm": 0.39714530799054115, + "learning_rate": 9.260859873746487e-06, + "loss": 0.4874, + "step": 1408 + }, + { + "epoch": 0.2581531696592158, + "grad_norm": 0.3703455263000783, + "learning_rate": 9.259185676388053e-06, + "loss": 0.4676, + "step": 1409 + }, + { + "epoch": 0.25833638695492855, + "grad_norm": 0.40896486520452896, + "learning_rate": 9.257509736783173e-06, + "loss": 0.4972, + "step": 1410 + }, + { + "epoch": 0.25851960425064124, + "grad_norm": 0.38349770802111716, + "learning_rate": 9.2558320556174e-06, + "loss": 0.4986, + "step": 1411 + }, + { + "epoch": 0.258702821546354, + "grad_norm": 0.38845233640399973, + "learning_rate": 9.254152633576996e-06, + "loss": 0.4781, + "step": 1412 + }, + { + "epoch": 0.2588860388420667, + "grad_norm": 0.4414806402475634, + "learning_rate": 9.252471471348945e-06, + "loss": 0.4998, + "step": 1413 + }, + { + "epoch": 0.2590692561377794, + "grad_norm": 0.6508729529767923, + "learning_rate": 9.250788569620933e-06, + "loss": 0.4731, + "step": 1414 + }, + { + "epoch": 0.2592524734334921, + "grad_norm": 0.3943363905618656, + "learning_rate": 9.249103929081362e-06, + "loss": 0.5024, + "step": 1415 + }, + { + "epoch": 0.25943569072920486, + "grad_norm": 0.40856928097689554, + "learning_rate": 9.247417550419347e-06, + "loss": 0.5212, + "step": 1416 + }, + { + "epoch": 0.25961890802491755, + "grad_norm": 0.441062293662255, + "learning_rate": 9.245729434324708e-06, + "loss": 0.4818, + "step": 1417 + }, + { + "epoch": 0.2598021253206303, + "grad_norm": 0.3948984673415181, + "learning_rate": 9.244039581487983e-06, + "loss": 0.4666, + "step": 1418 + }, + { + "epoch": 0.259985342616343, + "grad_norm": 0.3925260255665206, + "learning_rate": 9.242347992600416e-06, + "loss": 0.5116, + "step": 1419 + }, + { + "epoch": 0.2601685599120557, + "grad_norm": 0.4304859621407188, + "learning_rate": 9.24065466835396e-06, + "loss": 0.4982, + "step": 1420 + }, + { + "epoch": 0.2603517772077684, + "grad_norm": 0.40221018211406145, + "learning_rate": 9.23895960944128e-06, + "loss": 0.5005, + "step": 1421 + }, + { + "epoch": 0.2605349945034811, + "grad_norm": 0.41700277900738475, + "learning_rate": 9.237262816555755e-06, + "loss": 0.5002, + "step": 1422 + }, + { + "epoch": 0.26071821179919386, + "grad_norm": 0.4685774232060306, + "learning_rate": 9.235564290391461e-06, + "loss": 0.511, + "step": 1423 + }, + { + "epoch": 0.26090142909490655, + "grad_norm": 0.43210491917117466, + "learning_rate": 9.233864031643199e-06, + "loss": 0.4969, + "step": 1424 + }, + { + "epoch": 0.2610846463906193, + "grad_norm": 0.39899357606382013, + "learning_rate": 9.232162041006463e-06, + "loss": 0.5067, + "step": 1425 + }, + { + "epoch": 0.261267863686332, + "grad_norm": 0.3914730405010264, + "learning_rate": 9.230458319177467e-06, + "loss": 0.4567, + "step": 1426 + }, + { + "epoch": 0.2614510809820447, + "grad_norm": 0.4224757115631004, + "learning_rate": 9.22875286685313e-06, + "loss": 0.4878, + "step": 1427 + }, + { + "epoch": 0.2616342982777574, + "grad_norm": 0.44070937438452384, + "learning_rate": 9.227045684731075e-06, + "loss": 0.4751, + "step": 1428 + }, + { + "epoch": 0.26181751557347016, + "grad_norm": 0.41874124802141355, + "learning_rate": 9.225336773509637e-06, + "loss": 0.495, + "step": 1429 + }, + { + "epoch": 0.26200073286918285, + "grad_norm": 0.39663369299149254, + "learning_rate": 9.223626133887856e-06, + "loss": 0.502, + "step": 1430 + }, + { + "epoch": 0.26218395016489554, + "grad_norm": 0.3978403037913836, + "learning_rate": 9.22191376656548e-06, + "loss": 0.5057, + "step": 1431 + }, + { + "epoch": 0.2623671674606083, + "grad_norm": 0.39621285816390917, + "learning_rate": 9.220199672242965e-06, + "loss": 0.4787, + "step": 1432 + }, + { + "epoch": 0.262550384756321, + "grad_norm": 0.36823591327830396, + "learning_rate": 9.218483851621469e-06, + "loss": 0.4759, + "step": 1433 + }, + { + "epoch": 0.2627336020520337, + "grad_norm": 0.40956001204846254, + "learning_rate": 9.216766305402862e-06, + "loss": 0.4992, + "step": 1434 + }, + { + "epoch": 0.2629168193477464, + "grad_norm": 0.4909546794694869, + "learning_rate": 9.215047034289716e-06, + "loss": 0.4893, + "step": 1435 + }, + { + "epoch": 0.26310003664345916, + "grad_norm": 0.38305868280097505, + "learning_rate": 9.213326038985308e-06, + "loss": 0.4998, + "step": 1436 + }, + { + "epoch": 0.26328325393917185, + "grad_norm": 0.4506364554374671, + "learning_rate": 9.211603320193624e-06, + "loss": 0.5376, + "step": 1437 + }, + { + "epoch": 0.2634664712348846, + "grad_norm": 0.4532783129632112, + "learning_rate": 9.209878878619354e-06, + "loss": 0.4792, + "step": 1438 + }, + { + "epoch": 0.2636496885305973, + "grad_norm": 0.3927838537868758, + "learning_rate": 9.208152714967888e-06, + "loss": 0.4671, + "step": 1439 + }, + { + "epoch": 0.26383290582631, + "grad_norm": 0.43211843791598387, + "learning_rate": 9.206424829945326e-06, + "loss": 0.5083, + "step": 1440 + }, + { + "epoch": 0.2640161231220227, + "grad_norm": 0.4598413576241595, + "learning_rate": 9.20469522425847e-06, + "loss": 0.4946, + "step": 1441 + }, + { + "epoch": 0.2641993404177354, + "grad_norm": 0.4087512250232026, + "learning_rate": 9.202963898614825e-06, + "loss": 0.5002, + "step": 1442 + }, + { + "epoch": 0.26438255771344815, + "grad_norm": 0.4584659510926636, + "learning_rate": 9.201230853722603e-06, + "loss": 0.4979, + "step": 1443 + }, + { + "epoch": 0.26456577500916084, + "grad_norm": 0.4088655212775404, + "learning_rate": 9.199496090290713e-06, + "loss": 0.5083, + "step": 1444 + }, + { + "epoch": 0.2647489923048736, + "grad_norm": 0.39750766713674895, + "learning_rate": 9.197759609028774e-06, + "loss": 0.4447, + "step": 1445 + }, + { + "epoch": 0.2649322096005863, + "grad_norm": 2.5841998747640154, + "learning_rate": 9.196021410647104e-06, + "loss": 0.4938, + "step": 1446 + }, + { + "epoch": 0.265115426896299, + "grad_norm": 0.4352826266759048, + "learning_rate": 9.194281495856724e-06, + "loss": 0.4791, + "step": 1447 + }, + { + "epoch": 0.2652986441920117, + "grad_norm": 0.4347966702352818, + "learning_rate": 9.192539865369354e-06, + "loss": 0.4713, + "step": 1448 + }, + { + "epoch": 0.26548186148772446, + "grad_norm": 0.37418945105267554, + "learning_rate": 9.190796519897423e-06, + "loss": 0.4953, + "step": 1449 + }, + { + "epoch": 0.26566507878343715, + "grad_norm": 0.45051341568089304, + "learning_rate": 9.189051460154054e-06, + "loss": 0.4844, + "step": 1450 + }, + { + "epoch": 0.2658482960791499, + "grad_norm": 0.4213633340607822, + "learning_rate": 9.187304686853078e-06, + "loss": 0.4927, + "step": 1451 + }, + { + "epoch": 0.2660315133748626, + "grad_norm": 0.4083049015051242, + "learning_rate": 9.185556200709021e-06, + "loss": 0.4819, + "step": 1452 + }, + { + "epoch": 0.26621473067057533, + "grad_norm": 0.7665941590007657, + "learning_rate": 9.183806002437112e-06, + "loss": 0.5237, + "step": 1453 + }, + { + "epoch": 0.266397947966288, + "grad_norm": 0.4231366399299115, + "learning_rate": 9.182054092753281e-06, + "loss": 0.5165, + "step": 1454 + }, + { + "epoch": 0.2665811652620007, + "grad_norm": 0.3819742000154287, + "learning_rate": 9.180300472374158e-06, + "loss": 0.5068, + "step": 1455 + }, + { + "epoch": 0.26676438255771345, + "grad_norm": 0.430881793871794, + "learning_rate": 9.178545142017073e-06, + "loss": 0.4855, + "step": 1456 + }, + { + "epoch": 0.26694759985342614, + "grad_norm": 0.4120790750347018, + "learning_rate": 9.176788102400053e-06, + "loss": 0.4885, + "step": 1457 + }, + { + "epoch": 0.2671308171491389, + "grad_norm": 0.4538815285108574, + "learning_rate": 9.175029354241827e-06, + "loss": 0.4863, + "step": 1458 + }, + { + "epoch": 0.2673140344448516, + "grad_norm": 0.3830901020270506, + "learning_rate": 9.173268898261822e-06, + "loss": 0.4863, + "step": 1459 + }, + { + "epoch": 0.2674972517405643, + "grad_norm": 0.4245356876119995, + "learning_rate": 9.171506735180164e-06, + "loss": 0.4711, + "step": 1460 + }, + { + "epoch": 0.267680469036277, + "grad_norm": 0.47590664332011035, + "learning_rate": 9.169742865717675e-06, + "loss": 0.4902, + "step": 1461 + }, + { + "epoch": 0.26786368633198976, + "grad_norm": 0.4472826376520853, + "learning_rate": 9.167977290595879e-06, + "loss": 0.4752, + "step": 1462 + }, + { + "epoch": 0.26804690362770245, + "grad_norm": 0.45383853982527605, + "learning_rate": 9.166210010536996e-06, + "loss": 0.5041, + "step": 1463 + }, + { + "epoch": 0.2682301209234152, + "grad_norm": 0.4349556389365925, + "learning_rate": 9.164441026263939e-06, + "loss": 0.4841, + "step": 1464 + }, + { + "epoch": 0.2684133382191279, + "grad_norm": 0.3896144484943624, + "learning_rate": 9.162670338500323e-06, + "loss": 0.4691, + "step": 1465 + }, + { + "epoch": 0.2685965555148406, + "grad_norm": 0.40828579323895614, + "learning_rate": 9.160897947970463e-06, + "loss": 0.5087, + "step": 1466 + }, + { + "epoch": 0.2687797728105533, + "grad_norm": 0.44921268339324344, + "learning_rate": 9.159123855399364e-06, + "loss": 0.5205, + "step": 1467 + }, + { + "epoch": 0.268962990106266, + "grad_norm": 0.3749973142689185, + "learning_rate": 9.157348061512728e-06, + "loss": 0.5054, + "step": 1468 + }, + { + "epoch": 0.26914620740197875, + "grad_norm": 0.8460064155416173, + "learning_rate": 9.155570567036956e-06, + "loss": 0.504, + "step": 1469 + }, + { + "epoch": 0.26932942469769144, + "grad_norm": 0.40099978049747603, + "learning_rate": 9.153791372699141e-06, + "loss": 0.5169, + "step": 1470 + }, + { + "epoch": 0.2695126419934042, + "grad_norm": 0.399695867626504, + "learning_rate": 9.152010479227078e-06, + "loss": 0.4435, + "step": 1471 + }, + { + "epoch": 0.2696958592891169, + "grad_norm": 0.4522941668627639, + "learning_rate": 9.15022788734925e-06, + "loss": 0.4892, + "step": 1472 + }, + { + "epoch": 0.2698790765848296, + "grad_norm": 0.3907998733925797, + "learning_rate": 9.148443597794839e-06, + "loss": 0.4763, + "step": 1473 + }, + { + "epoch": 0.2700622938805423, + "grad_norm": 0.3989688222157153, + "learning_rate": 9.146657611293714e-06, + "loss": 0.4907, + "step": 1474 + }, + { + "epoch": 0.27024551117625506, + "grad_norm": 0.4158272717294187, + "learning_rate": 9.144869928576451e-06, + "loss": 0.4585, + "step": 1475 + }, + { + "epoch": 0.27042872847196775, + "grad_norm": 0.4122595114642358, + "learning_rate": 9.14308055037431e-06, + "loss": 0.5156, + "step": 1476 + }, + { + "epoch": 0.2706119457676805, + "grad_norm": 0.5428248637214301, + "learning_rate": 9.141289477419246e-06, + "loss": 0.5146, + "step": 1477 + }, + { + "epoch": 0.2707951630633932, + "grad_norm": 0.4081668756027033, + "learning_rate": 9.139496710443911e-06, + "loss": 0.4962, + "step": 1478 + }, + { + "epoch": 0.2709783803591059, + "grad_norm": 0.43919686618422127, + "learning_rate": 9.137702250181646e-06, + "loss": 0.5238, + "step": 1479 + }, + { + "epoch": 0.2711615976548186, + "grad_norm": 0.4199352917649959, + "learning_rate": 9.135906097366486e-06, + "loss": 0.5171, + "step": 1480 + }, + { + "epoch": 0.2713448149505313, + "grad_norm": 0.3812628825329553, + "learning_rate": 9.134108252733159e-06, + "loss": 0.4919, + "step": 1481 + }, + { + "epoch": 0.27152803224624406, + "grad_norm": 0.4062350904486568, + "learning_rate": 9.132308717017084e-06, + "loss": 0.4827, + "step": 1482 + }, + { + "epoch": 0.27171124954195675, + "grad_norm": 0.37649670407672037, + "learning_rate": 9.130507490954375e-06, + "loss": 0.4862, + "step": 1483 + }, + { + "epoch": 0.2718944668376695, + "grad_norm": 0.39510364308925733, + "learning_rate": 9.12870457528183e-06, + "loss": 0.4695, + "step": 1484 + }, + { + "epoch": 0.2720776841333822, + "grad_norm": 0.4412716554236446, + "learning_rate": 9.126899970736947e-06, + "loss": 0.4929, + "step": 1485 + }, + { + "epoch": 0.2722609014290949, + "grad_norm": 0.37777764528277363, + "learning_rate": 9.125093678057909e-06, + "loss": 0.4965, + "step": 1486 + }, + { + "epoch": 0.2724441187248076, + "grad_norm": 0.3701200088318955, + "learning_rate": 9.12328569798359e-06, + "loss": 0.4857, + "step": 1487 + }, + { + "epoch": 0.27262733602052036, + "grad_norm": 0.4323306961731567, + "learning_rate": 9.121476031253557e-06, + "loss": 0.47, + "step": 1488 + }, + { + "epoch": 0.27281055331623305, + "grad_norm": 0.40253559983281856, + "learning_rate": 9.119664678608067e-06, + "loss": 0.4474, + "step": 1489 + }, + { + "epoch": 0.27299377061194574, + "grad_norm": 0.36316272808709116, + "learning_rate": 9.117851640788064e-06, + "loss": 0.4743, + "step": 1490 + }, + { + "epoch": 0.2731769879076585, + "grad_norm": 0.39987807717863316, + "learning_rate": 9.11603691853518e-06, + "loss": 0.5074, + "step": 1491 + }, + { + "epoch": 0.2733602052033712, + "grad_norm": 0.36775081110408747, + "learning_rate": 9.11422051259174e-06, + "loss": 0.4805, + "step": 1492 + }, + { + "epoch": 0.2735434224990839, + "grad_norm": 0.41940289454678054, + "learning_rate": 9.112402423700759e-06, + "loss": 0.5071, + "step": 1493 + }, + { + "epoch": 0.2737266397947966, + "grad_norm": 0.3883498469145558, + "learning_rate": 9.110582652605934e-06, + "loss": 0.4839, + "step": 1494 + }, + { + "epoch": 0.27390985709050936, + "grad_norm": 0.40736107872427696, + "learning_rate": 9.108761200051655e-06, + "loss": 0.4769, + "step": 1495 + }, + { + "epoch": 0.27409307438622205, + "grad_norm": 0.42145067923082186, + "learning_rate": 9.106938066782998e-06, + "loss": 0.5176, + "step": 1496 + }, + { + "epoch": 0.2742762916819348, + "grad_norm": 0.4436513324175086, + "learning_rate": 9.105113253545727e-06, + "loss": 0.5185, + "step": 1497 + }, + { + "epoch": 0.2744595089776475, + "grad_norm": 0.3692944363116775, + "learning_rate": 9.103286761086294e-06, + "loss": 0.4369, + "step": 1498 + }, + { + "epoch": 0.2746427262733602, + "grad_norm": 0.4760665480063585, + "learning_rate": 9.101458590151837e-06, + "loss": 0.5006, + "step": 1499 + }, + { + "epoch": 0.2748259435690729, + "grad_norm": 0.4079202639977091, + "learning_rate": 9.099628741490179e-06, + "loss": 0.486, + "step": 1500 + }, + { + "epoch": 0.27500916086478566, + "grad_norm": 0.43182526673935, + "learning_rate": 9.097797215849834e-06, + "loss": 0.5081, + "step": 1501 + }, + { + "epoch": 0.27519237816049835, + "grad_norm": 0.39431567471837187, + "learning_rate": 9.095964013979998e-06, + "loss": 0.4692, + "step": 1502 + }, + { + "epoch": 0.27537559545621104, + "grad_norm": 0.4019693616622923, + "learning_rate": 9.094129136630552e-06, + "loss": 0.476, + "step": 1503 + }, + { + "epoch": 0.2755588127519238, + "grad_norm": 0.4313498077896387, + "learning_rate": 9.092292584552064e-06, + "loss": 0.4775, + "step": 1504 + }, + { + "epoch": 0.2757420300476365, + "grad_norm": 0.3604034259823722, + "learning_rate": 9.09045435849579e-06, + "loss": 0.497, + "step": 1505 + }, + { + "epoch": 0.2759252473433492, + "grad_norm": 0.4981716823895217, + "learning_rate": 9.088614459213665e-06, + "loss": 0.5151, + "step": 1506 + }, + { + "epoch": 0.2761084646390619, + "grad_norm": 0.42985286366030234, + "learning_rate": 9.086772887458314e-06, + "loss": 0.4724, + "step": 1507 + }, + { + "epoch": 0.27629168193477466, + "grad_norm": 0.40646371473453785, + "learning_rate": 9.08492964398304e-06, + "loss": 0.4875, + "step": 1508 + }, + { + "epoch": 0.27647489923048735, + "grad_norm": 0.3651071526478735, + "learning_rate": 9.083084729541836e-06, + "loss": 0.5, + "step": 1509 + }, + { + "epoch": 0.2766581165262001, + "grad_norm": 0.49895693881998293, + "learning_rate": 9.081238144889373e-06, + "loss": 0.5184, + "step": 1510 + }, + { + "epoch": 0.2768413338219128, + "grad_norm": 0.4128024780491911, + "learning_rate": 9.079389890781011e-06, + "loss": 0.4879, + "step": 1511 + }, + { + "epoch": 0.27702455111762553, + "grad_norm": 0.41553878397064187, + "learning_rate": 9.077539967972788e-06, + "loss": 0.5132, + "step": 1512 + }, + { + "epoch": 0.2772077684133382, + "grad_norm": 0.5023456269550239, + "learning_rate": 9.075688377221428e-06, + "loss": 0.5316, + "step": 1513 + }, + { + "epoch": 0.2773909857090509, + "grad_norm": 0.42658543986166453, + "learning_rate": 9.073835119284333e-06, + "loss": 0.4954, + "step": 1514 + }, + { + "epoch": 0.27757420300476365, + "grad_norm": 0.4257754248428276, + "learning_rate": 9.071980194919592e-06, + "loss": 0.4739, + "step": 1515 + }, + { + "epoch": 0.27775742030047634, + "grad_norm": 0.36513061047252116, + "learning_rate": 9.070123604885973e-06, + "loss": 0.4666, + "step": 1516 + }, + { + "epoch": 0.2779406375961891, + "grad_norm": 0.39730883326499855, + "learning_rate": 9.068265349942926e-06, + "loss": 0.4857, + "step": 1517 + }, + { + "epoch": 0.2781238548919018, + "grad_norm": 0.4089832030333659, + "learning_rate": 9.06640543085058e-06, + "loss": 0.5033, + "step": 1518 + }, + { + "epoch": 0.2783070721876145, + "grad_norm": 0.45930382348605114, + "learning_rate": 9.064543848369749e-06, + "loss": 0.5071, + "step": 1519 + }, + { + "epoch": 0.2784902894833272, + "grad_norm": 0.420241795630585, + "learning_rate": 9.062680603261923e-06, + "loss": 0.4966, + "step": 1520 + }, + { + "epoch": 0.27867350677903996, + "grad_norm": 0.45585947515287706, + "learning_rate": 9.060815696289273e-06, + "loss": 0.475, + "step": 1521 + }, + { + "epoch": 0.27885672407475265, + "grad_norm": 0.37205804655379004, + "learning_rate": 9.058949128214655e-06, + "loss": 0.5017, + "step": 1522 + }, + { + "epoch": 0.2790399413704654, + "grad_norm": 0.4529387207435349, + "learning_rate": 9.057080899801598e-06, + "loss": 0.482, + "step": 1523 + }, + { + "epoch": 0.2792231586661781, + "grad_norm": 0.3712655517985857, + "learning_rate": 9.055211011814312e-06, + "loss": 0.458, + "step": 1524 + }, + { + "epoch": 0.27940637596189083, + "grad_norm": 0.39642625126203734, + "learning_rate": 9.053339465017685e-06, + "loss": 0.4687, + "step": 1525 + }, + { + "epoch": 0.2795895932576035, + "grad_norm": 0.4136774885883469, + "learning_rate": 9.051466260177286e-06, + "loss": 0.5088, + "step": 1526 + }, + { + "epoch": 0.2797728105533162, + "grad_norm": 0.41134059761022984, + "learning_rate": 9.049591398059364e-06, + "loss": 0.4934, + "step": 1527 + }, + { + "epoch": 0.27995602784902895, + "grad_norm": 0.42078398613472917, + "learning_rate": 9.047714879430841e-06, + "loss": 0.488, + "step": 1528 + }, + { + "epoch": 0.28013924514474164, + "grad_norm": 0.3893031394642896, + "learning_rate": 9.045836705059316e-06, + "loss": 0.4787, + "step": 1529 + }, + { + "epoch": 0.2803224624404544, + "grad_norm": 0.4093693977481649, + "learning_rate": 9.043956875713071e-06, + "loss": 0.5038, + "step": 1530 + }, + { + "epoch": 0.2805056797361671, + "grad_norm": 0.39262539132038, + "learning_rate": 9.042075392161062e-06, + "loss": 0.4789, + "step": 1531 + }, + { + "epoch": 0.2806888970318798, + "grad_norm": 0.45335811960346867, + "learning_rate": 9.040192255172919e-06, + "loss": 0.4875, + "step": 1532 + }, + { + "epoch": 0.2808721143275925, + "grad_norm": 0.36503912559147556, + "learning_rate": 9.038307465518954e-06, + "loss": 0.477, + "step": 1533 + }, + { + "epoch": 0.28105533162330526, + "grad_norm": 0.4257211312986204, + "learning_rate": 9.03642102397015e-06, + "loss": 0.469, + "step": 1534 + }, + { + "epoch": 0.28123854891901795, + "grad_norm": 0.4252988723490824, + "learning_rate": 9.034532931298169e-06, + "loss": 0.4881, + "step": 1535 + }, + { + "epoch": 0.2814217662147307, + "grad_norm": 0.40947075405147687, + "learning_rate": 9.032643188275346e-06, + "loss": 0.4943, + "step": 1536 + }, + { + "epoch": 0.2816049835104434, + "grad_norm": 0.4104856496215327, + "learning_rate": 9.030751795674693e-06, + "loss": 0.5001, + "step": 1537 + }, + { + "epoch": 0.2817882008061561, + "grad_norm": 0.4108489148032505, + "learning_rate": 9.028858754269893e-06, + "loss": 0.486, + "step": 1538 + }, + { + "epoch": 0.2819714181018688, + "grad_norm": 0.400218515841904, + "learning_rate": 9.026964064835312e-06, + "loss": 0.4875, + "step": 1539 + }, + { + "epoch": 0.2821546353975815, + "grad_norm": 0.392307962036209, + "learning_rate": 9.02506772814598e-06, + "loss": 0.4696, + "step": 1540 + }, + { + "epoch": 0.28233785269329426, + "grad_norm": 0.39418249225011087, + "learning_rate": 9.023169744977607e-06, + "loss": 0.4944, + "step": 1541 + }, + { + "epoch": 0.28252106998900695, + "grad_norm": 0.36881071138331667, + "learning_rate": 9.021270116106574e-06, + "loss": 0.458, + "step": 1542 + }, + { + "epoch": 0.2827042872847197, + "grad_norm": 0.3604688853323364, + "learning_rate": 9.019368842309937e-06, + "loss": 0.4612, + "step": 1543 + }, + { + "epoch": 0.2828875045804324, + "grad_norm": 0.43502584499562924, + "learning_rate": 9.017465924365423e-06, + "loss": 0.535, + "step": 1544 + }, + { + "epoch": 0.2830707218761451, + "grad_norm": 0.3495743983793779, + "learning_rate": 9.015561363051434e-06, + "loss": 0.4884, + "step": 1545 + }, + { + "epoch": 0.2832539391718578, + "grad_norm": 0.38746814328257545, + "learning_rate": 9.01365515914704e-06, + "loss": 0.4684, + "step": 1546 + }, + { + "epoch": 0.28343715646757056, + "grad_norm": 0.42208621683162756, + "learning_rate": 9.011747313431988e-06, + "loss": 0.4894, + "step": 1547 + }, + { + "epoch": 0.28362037376328325, + "grad_norm": 0.40174728758795464, + "learning_rate": 9.009837826686693e-06, + "loss": 0.4892, + "step": 1548 + }, + { + "epoch": 0.283803591058996, + "grad_norm": 0.3804107018996848, + "learning_rate": 9.007926699692244e-06, + "loss": 0.4781, + "step": 1549 + }, + { + "epoch": 0.2839868083547087, + "grad_norm": 0.4347596518542267, + "learning_rate": 9.006013933230398e-06, + "loss": 0.4735, + "step": 1550 + }, + { + "epoch": 0.2841700256504214, + "grad_norm": 0.4121135548481947, + "learning_rate": 9.004099528083583e-06, + "loss": 0.4719, + "step": 1551 + }, + { + "epoch": 0.2843532429461341, + "grad_norm": 0.44795396529132614, + "learning_rate": 9.0021834850349e-06, + "loss": 0.5099, + "step": 1552 + }, + { + "epoch": 0.2845364602418468, + "grad_norm": 0.44372233437035424, + "learning_rate": 9.000265804868118e-06, + "loss": 0.4532, + "step": 1553 + }, + { + "epoch": 0.28471967753755956, + "grad_norm": 0.37122246988181606, + "learning_rate": 8.998346488367675e-06, + "loss": 0.5031, + "step": 1554 + }, + { + "epoch": 0.28490289483327225, + "grad_norm": 0.4132553844079889, + "learning_rate": 8.996425536318683e-06, + "loss": 0.4973, + "step": 1555 + }, + { + "epoch": 0.285086112128985, + "grad_norm": 0.44678572881841533, + "learning_rate": 8.994502949506914e-06, + "loss": 0.4937, + "step": 1556 + }, + { + "epoch": 0.2852693294246977, + "grad_norm": 0.3961645213095161, + "learning_rate": 8.99257872871882e-06, + "loss": 0.4917, + "step": 1557 + }, + { + "epoch": 0.2854525467204104, + "grad_norm": 0.4068575794251254, + "learning_rate": 8.99065287474151e-06, + "loss": 0.5059, + "step": 1558 + }, + { + "epoch": 0.2856357640161231, + "grad_norm": 0.3439421297681402, + "learning_rate": 8.98872538836277e-06, + "loss": 0.4943, + "step": 1559 + }, + { + "epoch": 0.28581898131183586, + "grad_norm": 0.4072178672567339, + "learning_rate": 8.986796270371047e-06, + "loss": 0.4884, + "step": 1560 + }, + { + "epoch": 0.28600219860754855, + "grad_norm": 0.42449273454289244, + "learning_rate": 8.984865521555464e-06, + "loss": 0.4875, + "step": 1561 + }, + { + "epoch": 0.28618541590326124, + "grad_norm": 0.37613787684490696, + "learning_rate": 8.9829331427058e-06, + "loss": 0.4857, + "step": 1562 + }, + { + "epoch": 0.286368633198974, + "grad_norm": 0.36486974332112687, + "learning_rate": 8.980999134612512e-06, + "loss": 0.5002, + "step": 1563 + }, + { + "epoch": 0.2865518504946867, + "grad_norm": 0.38386234504756434, + "learning_rate": 8.979063498066714e-06, + "loss": 0.4885, + "step": 1564 + }, + { + "epoch": 0.2867350677903994, + "grad_norm": 0.39062088012736546, + "learning_rate": 8.977126233860193e-06, + "loss": 0.4936, + "step": 1565 + }, + { + "epoch": 0.2869182850861121, + "grad_norm": 0.4126575331178653, + "learning_rate": 8.975187342785397e-06, + "loss": 0.4903, + "step": 1566 + }, + { + "epoch": 0.28710150238182486, + "grad_norm": 0.46684149874863134, + "learning_rate": 8.973246825635441e-06, + "loss": 0.527, + "step": 1567 + }, + { + "epoch": 0.28728471967753755, + "grad_norm": 0.39335552791237355, + "learning_rate": 8.97130468320411e-06, + "loss": 0.4779, + "step": 1568 + }, + { + "epoch": 0.2874679369732503, + "grad_norm": 0.4246548051800614, + "learning_rate": 8.969360916285844e-06, + "loss": 0.4815, + "step": 1569 + }, + { + "epoch": 0.287651154268963, + "grad_norm": 0.44025612619503274, + "learning_rate": 8.967415525675756e-06, + "loss": 0.4371, + "step": 1570 + }, + { + "epoch": 0.28783437156467573, + "grad_norm": 0.9703871617250511, + "learning_rate": 8.96546851216962e-06, + "loss": 0.4856, + "step": 1571 + }, + { + "epoch": 0.2880175888603884, + "grad_norm": 0.49217374632039296, + "learning_rate": 8.96351987656387e-06, + "loss": 0.5087, + "step": 1572 + }, + { + "epoch": 0.28820080615610116, + "grad_norm": 0.41378220482583816, + "learning_rate": 8.961569619655615e-06, + "loss": 0.4736, + "step": 1573 + }, + { + "epoch": 0.28838402345181385, + "grad_norm": 0.4186462441272065, + "learning_rate": 8.959617742242615e-06, + "loss": 0.493, + "step": 1574 + }, + { + "epoch": 0.28856724074752654, + "grad_norm": 0.3880931424436376, + "learning_rate": 8.957664245123297e-06, + "loss": 0.4844, + "step": 1575 + }, + { + "epoch": 0.2887504580432393, + "grad_norm": 0.3741132383694949, + "learning_rate": 8.955709129096752e-06, + "loss": 0.4055, + "step": 1576 + }, + { + "epoch": 0.288933675338952, + "grad_norm": 0.44863598418647255, + "learning_rate": 8.953752394962736e-06, + "loss": 0.5278, + "step": 1577 + }, + { + "epoch": 0.2891168926346647, + "grad_norm": 0.3876273635035489, + "learning_rate": 8.951794043521657e-06, + "loss": 0.4859, + "step": 1578 + }, + { + "epoch": 0.2893001099303774, + "grad_norm": 0.34806890820847197, + "learning_rate": 8.949834075574595e-06, + "loss": 0.4768, + "step": 1579 + }, + { + "epoch": 0.28948332722609016, + "grad_norm": 0.4072445168776299, + "learning_rate": 8.947872491923288e-06, + "loss": 0.487, + "step": 1580 + }, + { + "epoch": 0.28966654452180285, + "grad_norm": 0.4293336896066131, + "learning_rate": 8.94590929337013e-06, + "loss": 0.5008, + "step": 1581 + }, + { + "epoch": 0.2898497618175156, + "grad_norm": 0.41128872476228806, + "learning_rate": 8.943944480718184e-06, + "loss": 0.51, + "step": 1582 + }, + { + "epoch": 0.2900329791132283, + "grad_norm": 0.36929320177047475, + "learning_rate": 8.941978054771165e-06, + "loss": 0.5116, + "step": 1583 + }, + { + "epoch": 0.29021619640894103, + "grad_norm": 0.3533069262895081, + "learning_rate": 8.940010016333453e-06, + "loss": 0.462, + "step": 1584 + }, + { + "epoch": 0.2903994137046537, + "grad_norm": 0.4290477614884879, + "learning_rate": 8.938040366210088e-06, + "loss": 0.4886, + "step": 1585 + }, + { + "epoch": 0.2905826310003664, + "grad_norm": 0.4049126999073717, + "learning_rate": 8.936069105206767e-06, + "loss": 0.488, + "step": 1586 + }, + { + "epoch": 0.29076584829607915, + "grad_norm": 0.37625165658845383, + "learning_rate": 8.934096234129843e-06, + "loss": 0.4704, + "step": 1587 + }, + { + "epoch": 0.29094906559179184, + "grad_norm": 0.4368199590248626, + "learning_rate": 8.932121753786339e-06, + "loss": 0.5089, + "step": 1588 + }, + { + "epoch": 0.2911322828875046, + "grad_norm": 0.4293928339113946, + "learning_rate": 8.930145664983921e-06, + "loss": 0.4989, + "step": 1589 + }, + { + "epoch": 0.2913155001832173, + "grad_norm": 0.362276762434061, + "learning_rate": 8.928167968530921e-06, + "loss": 0.4848, + "step": 1590 + }, + { + "epoch": 0.29149871747893, + "grad_norm": 0.4545018284719774, + "learning_rate": 8.926188665236334e-06, + "loss": 0.5297, + "step": 1591 + }, + { + "epoch": 0.2916819347746427, + "grad_norm": 0.4271204305866213, + "learning_rate": 8.924207755909801e-06, + "loss": 0.4635, + "step": 1592 + }, + { + "epoch": 0.29186515207035546, + "grad_norm": 0.3895989149171432, + "learning_rate": 8.922225241361629e-06, + "loss": 0.4802, + "step": 1593 + }, + { + "epoch": 0.29204836936606815, + "grad_norm": 0.36495793219351547, + "learning_rate": 8.920241122402773e-06, + "loss": 0.4963, + "step": 1594 + }, + { + "epoch": 0.2922315866617809, + "grad_norm": 0.38869432485787914, + "learning_rate": 8.918255399844855e-06, + "loss": 0.4996, + "step": 1595 + }, + { + "epoch": 0.2924148039574936, + "grad_norm": 0.4601922887324616, + "learning_rate": 8.916268074500141e-06, + "loss": 0.5025, + "step": 1596 + }, + { + "epoch": 0.2925980212532063, + "grad_norm": 0.39205663001486896, + "learning_rate": 8.914279147181564e-06, + "loss": 0.4603, + "step": 1597 + }, + { + "epoch": 0.292781238548919, + "grad_norm": 0.4013458043993565, + "learning_rate": 8.912288618702706e-06, + "loss": 0.487, + "step": 1598 + }, + { + "epoch": 0.2929644558446317, + "grad_norm": 0.38939321801864035, + "learning_rate": 8.910296489877803e-06, + "loss": 0.5029, + "step": 1599 + }, + { + "epoch": 0.29314767314034446, + "grad_norm": 0.35395541109470974, + "learning_rate": 8.90830276152175e-06, + "loss": 0.4551, + "step": 1600 + }, + { + "epoch": 0.29333089043605715, + "grad_norm": 0.41856036352490694, + "learning_rate": 8.906307434450092e-06, + "loss": 0.4949, + "step": 1601 + }, + { + "epoch": 0.2935141077317699, + "grad_norm": 0.5275243050289077, + "learning_rate": 8.904310509479031e-06, + "loss": 0.4961, + "step": 1602 + }, + { + "epoch": 0.2936973250274826, + "grad_norm": 0.4742284206723536, + "learning_rate": 8.902311987425422e-06, + "loss": 0.5095, + "step": 1603 + }, + { + "epoch": 0.2938805423231953, + "grad_norm": 0.3390894324421048, + "learning_rate": 8.900311869106772e-06, + "loss": 0.4675, + "step": 1604 + }, + { + "epoch": 0.294063759618908, + "grad_norm": 0.39896932989796424, + "learning_rate": 8.898310155341245e-06, + "loss": 0.4819, + "step": 1605 + }, + { + "epoch": 0.29424697691462076, + "grad_norm": 0.5410338815825568, + "learning_rate": 8.896306846947649e-06, + "loss": 0.5032, + "step": 1606 + }, + { + "epoch": 0.29443019421033345, + "grad_norm": 0.4070700524177446, + "learning_rate": 8.894301944745453e-06, + "loss": 0.48, + "step": 1607 + }, + { + "epoch": 0.2946134115060462, + "grad_norm": 0.42862296543081163, + "learning_rate": 8.892295449554777e-06, + "loss": 0.4605, + "step": 1608 + }, + { + "epoch": 0.2947966288017589, + "grad_norm": 0.4287346890153464, + "learning_rate": 8.890287362196385e-06, + "loss": 0.485, + "step": 1609 + }, + { + "epoch": 0.2949798460974716, + "grad_norm": 0.43250442423486146, + "learning_rate": 8.888277683491701e-06, + "loss": 0.4854, + "step": 1610 + }, + { + "epoch": 0.2951630633931843, + "grad_norm": 0.49215086511267214, + "learning_rate": 8.886266414262797e-06, + "loss": 0.489, + "step": 1611 + }, + { + "epoch": 0.295346280688897, + "grad_norm": 0.4186341465088225, + "learning_rate": 8.884253555332393e-06, + "loss": 0.5078, + "step": 1612 + }, + { + "epoch": 0.29552949798460976, + "grad_norm": 0.4406315027714821, + "learning_rate": 8.882239107523863e-06, + "loss": 0.4937, + "step": 1613 + }, + { + "epoch": 0.29571271528032245, + "grad_norm": 0.3726428857037365, + "learning_rate": 8.88022307166123e-06, + "loss": 0.453, + "step": 1614 + }, + { + "epoch": 0.2958959325760352, + "grad_norm": 0.42821871054675215, + "learning_rate": 8.878205448569163e-06, + "loss": 0.4934, + "step": 1615 + }, + { + "epoch": 0.2960791498717479, + "grad_norm": 0.3670484899531654, + "learning_rate": 8.876186239072988e-06, + "loss": 0.486, + "step": 1616 + }, + { + "epoch": 0.2962623671674606, + "grad_norm": 0.34461787552049733, + "learning_rate": 8.874165443998672e-06, + "loss": 0.4761, + "step": 1617 + }, + { + "epoch": 0.2964455844631733, + "grad_norm": 0.37867132917069235, + "learning_rate": 8.872143064172834e-06, + "loss": 0.4821, + "step": 1618 + }, + { + "epoch": 0.29662880175888606, + "grad_norm": 0.3712153633078572, + "learning_rate": 8.870119100422743e-06, + "loss": 0.4931, + "step": 1619 + }, + { + "epoch": 0.29681201905459875, + "grad_norm": 0.36684611413431756, + "learning_rate": 8.86809355357631e-06, + "loss": 0.4616, + "step": 1620 + }, + { + "epoch": 0.29699523635031144, + "grad_norm": 0.38108459496354996, + "learning_rate": 8.866066424462103e-06, + "loss": 0.5118, + "step": 1621 + }, + { + "epoch": 0.2971784536460242, + "grad_norm": 0.382915281692164, + "learning_rate": 8.864037713909329e-06, + "loss": 0.4564, + "step": 1622 + }, + { + "epoch": 0.2973616709417369, + "grad_norm": 0.38365339546998406, + "learning_rate": 8.862007422747842e-06, + "loss": 0.516, + "step": 1623 + }, + { + "epoch": 0.2975448882374496, + "grad_norm": 0.5453723638931661, + "learning_rate": 8.85997555180815e-06, + "loss": 0.4722, + "step": 1624 + }, + { + "epoch": 0.2977281055331623, + "grad_norm": 0.38449352037816636, + "learning_rate": 8.8579421019214e-06, + "loss": 0.4583, + "step": 1625 + }, + { + "epoch": 0.29791132282887506, + "grad_norm": 0.4091259949883011, + "learning_rate": 8.855907073919388e-06, + "loss": 0.5102, + "step": 1626 + }, + { + "epoch": 0.29809454012458775, + "grad_norm": 0.46579655657632557, + "learning_rate": 8.853870468634554e-06, + "loss": 0.504, + "step": 1627 + }, + { + "epoch": 0.2982777574203005, + "grad_norm": 0.34914615282378225, + "learning_rate": 8.851832286899984e-06, + "loss": 0.4809, + "step": 1628 + }, + { + "epoch": 0.2984609747160132, + "grad_norm": 0.3465447713300964, + "learning_rate": 8.84979252954941e-06, + "loss": 0.4714, + "step": 1629 + }, + { + "epoch": 0.29864419201172593, + "grad_norm": 0.36719763001696804, + "learning_rate": 8.847751197417208e-06, + "loss": 0.4856, + "step": 1630 + }, + { + "epoch": 0.2988274093074386, + "grad_norm": 0.3954216747573634, + "learning_rate": 8.845708291338396e-06, + "loss": 0.4802, + "step": 1631 + }, + { + "epoch": 0.29901062660315136, + "grad_norm": 0.40271380964660825, + "learning_rate": 8.84366381214864e-06, + "loss": 0.4788, + "step": 1632 + }, + { + "epoch": 0.29919384389886405, + "grad_norm": 0.3751708327909397, + "learning_rate": 8.841617760684242e-06, + "loss": 0.4834, + "step": 1633 + }, + { + "epoch": 0.29937706119457674, + "grad_norm": 0.36638025333362195, + "learning_rate": 8.839570137782157e-06, + "loss": 0.4451, + "step": 1634 + }, + { + "epoch": 0.2995602784902895, + "grad_norm": 0.34260231840043137, + "learning_rate": 8.837520944279976e-06, + "loss": 0.4816, + "step": 1635 + }, + { + "epoch": 0.2997434957860022, + "grad_norm": 0.41835359853675946, + "learning_rate": 8.835470181015935e-06, + "loss": 0.4657, + "step": 1636 + }, + { + "epoch": 0.2999267130817149, + "grad_norm": 0.3853522103777551, + "learning_rate": 8.833417848828912e-06, + "loss": 0.4798, + "step": 1637 + }, + { + "epoch": 0.3001099303774276, + "grad_norm": 0.4311087142185147, + "learning_rate": 8.831363948558422e-06, + "loss": 0.5055, + "step": 1638 + }, + { + "epoch": 0.30029314767314036, + "grad_norm": 0.4121988474004896, + "learning_rate": 8.829308481044631e-06, + "loss": 0.4946, + "step": 1639 + }, + { + "epoch": 0.30047636496885305, + "grad_norm": 0.39926874503164067, + "learning_rate": 8.82725144712834e-06, + "loss": 0.5105, + "step": 1640 + }, + { + "epoch": 0.3006595822645658, + "grad_norm": 0.42930617786556713, + "learning_rate": 8.82519284765099e-06, + "loss": 0.4742, + "step": 1641 + }, + { + "epoch": 0.3008427995602785, + "grad_norm": 0.4756153230731036, + "learning_rate": 8.823132683454662e-06, + "loss": 0.4858, + "step": 1642 + }, + { + "epoch": 0.30102601685599123, + "grad_norm": 0.4105739310138263, + "learning_rate": 8.821070955382082e-06, + "loss": 0.4748, + "step": 1643 + }, + { + "epoch": 0.3012092341517039, + "grad_norm": 0.42461593671764813, + "learning_rate": 8.819007664276614e-06, + "loss": 0.4747, + "step": 1644 + }, + { + "epoch": 0.3013924514474166, + "grad_norm": 0.42619757195470465, + "learning_rate": 8.816942810982258e-06, + "loss": 0.4783, + "step": 1645 + }, + { + "epoch": 0.30157566874312935, + "grad_norm": 0.441815860292698, + "learning_rate": 8.814876396343655e-06, + "loss": 0.4822, + "step": 1646 + }, + { + "epoch": 0.30175888603884204, + "grad_norm": 0.39546125127248405, + "learning_rate": 8.812808421206083e-06, + "loss": 0.4751, + "step": 1647 + }, + { + "epoch": 0.3019421033345548, + "grad_norm": 0.4039788211552708, + "learning_rate": 8.810738886415464e-06, + "loss": 0.4774, + "step": 1648 + }, + { + "epoch": 0.3021253206302675, + "grad_norm": 0.4195550927374702, + "learning_rate": 8.80866779281835e-06, + "loss": 0.5184, + "step": 1649 + }, + { + "epoch": 0.3023085379259802, + "grad_norm": 0.4410446432137119, + "learning_rate": 8.80659514126194e-06, + "loss": 0.52, + "step": 1650 + }, + { + "epoch": 0.3024917552216929, + "grad_norm": 0.3910752829108485, + "learning_rate": 8.804520932594061e-06, + "loss": 0.4734, + "step": 1651 + }, + { + "epoch": 0.30267497251740566, + "grad_norm": 0.4084027351750214, + "learning_rate": 8.802445167663181e-06, + "loss": 0.4906, + "step": 1652 + }, + { + "epoch": 0.30285818981311835, + "grad_norm": 0.3779109317009322, + "learning_rate": 8.800367847318407e-06, + "loss": 0.4661, + "step": 1653 + }, + { + "epoch": 0.3030414071088311, + "grad_norm": 0.4434368130901037, + "learning_rate": 8.798288972409477e-06, + "loss": 0.4956, + "step": 1654 + }, + { + "epoch": 0.3032246244045438, + "grad_norm": 0.40653821037943955, + "learning_rate": 8.79620854378677e-06, + "loss": 0.5006, + "step": 1655 + }, + { + "epoch": 0.30340784170025653, + "grad_norm": 0.40448780548308255, + "learning_rate": 8.794126562301298e-06, + "loss": 0.5131, + "step": 1656 + }, + { + "epoch": 0.3035910589959692, + "grad_norm": 0.4185848391026118, + "learning_rate": 8.792043028804706e-06, + "loss": 0.4937, + "step": 1657 + }, + { + "epoch": 0.3037742762916819, + "grad_norm": 0.43416404958296123, + "learning_rate": 8.78995794414928e-06, + "loss": 0.4927, + "step": 1658 + }, + { + "epoch": 0.30395749358739466, + "grad_norm": 0.3941144660162759, + "learning_rate": 8.787871309187936e-06, + "loss": 0.4719, + "step": 1659 + }, + { + "epoch": 0.30414071088310735, + "grad_norm": 0.37126342352995245, + "learning_rate": 8.785783124774223e-06, + "loss": 0.4722, + "step": 1660 + }, + { + "epoch": 0.3043239281788201, + "grad_norm": 0.4103240007205328, + "learning_rate": 8.783693391762328e-06, + "loss": 0.4809, + "step": 1661 + }, + { + "epoch": 0.3045071454745328, + "grad_norm": 0.4249044003852677, + "learning_rate": 8.781602111007066e-06, + "loss": 0.4626, + "step": 1662 + }, + { + "epoch": 0.3046903627702455, + "grad_norm": 0.41638775694944496, + "learning_rate": 8.779509283363894e-06, + "loss": 0.4929, + "step": 1663 + }, + { + "epoch": 0.3048735800659582, + "grad_norm": 0.3861009808325714, + "learning_rate": 8.77741490968889e-06, + "loss": 0.4706, + "step": 1664 + }, + { + "epoch": 0.30505679736167096, + "grad_norm": 0.4284459852901781, + "learning_rate": 8.775318990838775e-06, + "loss": 0.542, + "step": 1665 + }, + { + "epoch": 0.30524001465738365, + "grad_norm": 0.39067748379098055, + "learning_rate": 8.773221527670896e-06, + "loss": 0.4921, + "step": 1666 + }, + { + "epoch": 0.3054232319530964, + "grad_norm": 0.4636902711928648, + "learning_rate": 8.771122521043236e-06, + "loss": 0.4847, + "step": 1667 + }, + { + "epoch": 0.3056064492488091, + "grad_norm": 0.4041218634315, + "learning_rate": 8.769021971814401e-06, + "loss": 0.4865, + "step": 1668 + }, + { + "epoch": 0.3057896665445218, + "grad_norm": 0.39459279955438104, + "learning_rate": 8.766919880843639e-06, + "loss": 0.519, + "step": 1669 + }, + { + "epoch": 0.3059728838402345, + "grad_norm": 0.40128357110983964, + "learning_rate": 8.764816248990822e-06, + "loss": 0.4841, + "step": 1670 + }, + { + "epoch": 0.3061561011359472, + "grad_norm": 0.36997768478257936, + "learning_rate": 8.762711077116453e-06, + "loss": 0.4771, + "step": 1671 + }, + { + "epoch": 0.30633931843165996, + "grad_norm": 0.3480828562287117, + "learning_rate": 8.760604366081665e-06, + "loss": 0.4876, + "step": 1672 + }, + { + "epoch": 0.30652253572737265, + "grad_norm": 0.417116733445613, + "learning_rate": 8.758496116748221e-06, + "loss": 0.4763, + "step": 1673 + }, + { + "epoch": 0.3067057530230854, + "grad_norm": 0.38990995104925313, + "learning_rate": 8.756386329978518e-06, + "loss": 0.4952, + "step": 1674 + }, + { + "epoch": 0.3068889703187981, + "grad_norm": 0.4102790428532168, + "learning_rate": 8.754275006635573e-06, + "loss": 0.4938, + "step": 1675 + }, + { + "epoch": 0.3070721876145108, + "grad_norm": 0.44424837968994635, + "learning_rate": 8.752162147583036e-06, + "loss": 0.4862, + "step": 1676 + }, + { + "epoch": 0.3072554049102235, + "grad_norm": 0.36829717881414353, + "learning_rate": 8.750047753685188e-06, + "loss": 0.4714, + "step": 1677 + }, + { + "epoch": 0.30743862220593626, + "grad_norm": 0.39758875548969896, + "learning_rate": 8.747931825806933e-06, + "loss": 0.4832, + "step": 1678 + }, + { + "epoch": 0.30762183950164895, + "grad_norm": 0.41272232708939577, + "learning_rate": 8.745814364813806e-06, + "loss": 0.4872, + "step": 1679 + }, + { + "epoch": 0.3078050567973617, + "grad_norm": 0.3854030542482913, + "learning_rate": 8.743695371571963e-06, + "loss": 0.484, + "step": 1680 + }, + { + "epoch": 0.3079882740930744, + "grad_norm": 0.36746498837241803, + "learning_rate": 8.741574846948198e-06, + "loss": 0.4711, + "step": 1681 + }, + { + "epoch": 0.3081714913887871, + "grad_norm": 0.411804382886457, + "learning_rate": 8.73945279180992e-06, + "loss": 0.4759, + "step": 1682 + }, + { + "epoch": 0.3083547086844998, + "grad_norm": 0.4026474390517101, + "learning_rate": 8.737329207025172e-06, + "loss": 0.502, + "step": 1683 + }, + { + "epoch": 0.3085379259802125, + "grad_norm": 0.3639512942893561, + "learning_rate": 8.735204093462617e-06, + "loss": 0.4647, + "step": 1684 + }, + { + "epoch": 0.30872114327592526, + "grad_norm": 0.4601114282657689, + "learning_rate": 8.733077451991546e-06, + "loss": 0.4794, + "step": 1685 + }, + { + "epoch": 0.30890436057163795, + "grad_norm": 0.4029114065064777, + "learning_rate": 8.730949283481877e-06, + "loss": 0.4581, + "step": 1686 + }, + { + "epoch": 0.3090875778673507, + "grad_norm": 0.47530481298922933, + "learning_rate": 8.72881958880415e-06, + "loss": 0.5358, + "step": 1687 + }, + { + "epoch": 0.3092707951630634, + "grad_norm": 0.46546993084658045, + "learning_rate": 8.72668836882953e-06, + "loss": 0.508, + "step": 1688 + }, + { + "epoch": 0.30945401245877613, + "grad_norm": 0.44880821238197843, + "learning_rate": 8.724555624429805e-06, + "loss": 0.494, + "step": 1689 + }, + { + "epoch": 0.3096372297544888, + "grad_norm": 0.39081723015666026, + "learning_rate": 8.722421356477389e-06, + "loss": 0.4845, + "step": 1690 + }, + { + "epoch": 0.30982044705020156, + "grad_norm": 0.3529090891828588, + "learning_rate": 8.720285565845313e-06, + "loss": 0.4672, + "step": 1691 + }, + { + "epoch": 0.31000366434591425, + "grad_norm": 0.39898106739463174, + "learning_rate": 8.718148253407242e-06, + "loss": 0.458, + "step": 1692 + }, + { + "epoch": 0.31018688164162694, + "grad_norm": 0.3651751506557362, + "learning_rate": 8.716009420037452e-06, + "loss": 0.4978, + "step": 1693 + }, + { + "epoch": 0.3103700989373397, + "grad_norm": 0.360403523409843, + "learning_rate": 8.713869066610847e-06, + "loss": 0.489, + "step": 1694 + }, + { + "epoch": 0.3105533162330524, + "grad_norm": 0.37968929495733417, + "learning_rate": 8.711727194002955e-06, + "loss": 0.493, + "step": 1695 + }, + { + "epoch": 0.3107365335287651, + "grad_norm": 0.38561595360319445, + "learning_rate": 8.709583803089919e-06, + "loss": 0.4864, + "step": 1696 + }, + { + "epoch": 0.3109197508244778, + "grad_norm": 0.3925407466954778, + "learning_rate": 8.707438894748508e-06, + "loss": 0.4592, + "step": 1697 + }, + { + "epoch": 0.31110296812019056, + "grad_norm": 0.38697234703046507, + "learning_rate": 8.70529246985611e-06, + "loss": 0.506, + "step": 1698 + }, + { + "epoch": 0.31128618541590325, + "grad_norm": 0.40545775127101236, + "learning_rate": 8.703144529290733e-06, + "loss": 0.498, + "step": 1699 + }, + { + "epoch": 0.311469402711616, + "grad_norm": 0.35555366552450657, + "learning_rate": 8.700995073931004e-06, + "loss": 0.4838, + "step": 1700 + }, + { + "epoch": 0.3116526200073287, + "grad_norm": 0.4022620468333227, + "learning_rate": 8.698844104656175e-06, + "loss": 0.4859, + "step": 1701 + }, + { + "epoch": 0.31183583730304143, + "grad_norm": 0.4370267955102546, + "learning_rate": 8.696691622346109e-06, + "loss": 0.4821, + "step": 1702 + }, + { + "epoch": 0.3120190545987541, + "grad_norm": 0.39148309791555397, + "learning_rate": 8.694537627881296e-06, + "loss": 0.4969, + "step": 1703 + }, + { + "epoch": 0.31220227189446687, + "grad_norm": 0.4149560429420508, + "learning_rate": 8.69238212214284e-06, + "loss": 0.4833, + "step": 1704 + }, + { + "epoch": 0.31238548919017955, + "grad_norm": 0.4012554167583925, + "learning_rate": 8.690225106012462e-06, + "loss": 0.4845, + "step": 1705 + }, + { + "epoch": 0.31256870648589224, + "grad_norm": 0.3880778144850858, + "learning_rate": 8.688066580372506e-06, + "loss": 0.4402, + "step": 1706 + }, + { + "epoch": 0.312751923781605, + "grad_norm": 0.3709906292066814, + "learning_rate": 8.685906546105925e-06, + "loss": 0.461, + "step": 1707 + }, + { + "epoch": 0.3129351410773177, + "grad_norm": 0.4038618996987991, + "learning_rate": 8.6837450040963e-06, + "loss": 0.4952, + "step": 1708 + }, + { + "epoch": 0.3131183583730304, + "grad_norm": 0.4498213583992615, + "learning_rate": 8.681581955227823e-06, + "loss": 0.463, + "step": 1709 + }, + { + "epoch": 0.3133015756687431, + "grad_norm": 0.6063342768243897, + "learning_rate": 8.6794174003853e-06, + "loss": 0.4692, + "step": 1710 + }, + { + "epoch": 0.31348479296445586, + "grad_norm": 0.393078000699868, + "learning_rate": 8.677251340454155e-06, + "loss": 0.477, + "step": 1711 + }, + { + "epoch": 0.31366801026016855, + "grad_norm": 0.5208615924956094, + "learning_rate": 8.67508377632043e-06, + "loss": 0.4686, + "step": 1712 + }, + { + "epoch": 0.3138512275558813, + "grad_norm": 0.4156650220868597, + "learning_rate": 8.672914708870782e-06, + "loss": 0.489, + "step": 1713 + }, + { + "epoch": 0.314034444851594, + "grad_norm": 0.39229916132917864, + "learning_rate": 8.67074413899248e-06, + "loss": 0.4854, + "step": 1714 + }, + { + "epoch": 0.31421766214730673, + "grad_norm": 0.3432761699669145, + "learning_rate": 8.668572067573409e-06, + "loss": 0.5026, + "step": 1715 + }, + { + "epoch": 0.3144008794430194, + "grad_norm": 0.4024782274517547, + "learning_rate": 8.666398495502068e-06, + "loss": 0.5145, + "step": 1716 + }, + { + "epoch": 0.3145840967387321, + "grad_norm": 0.4193753630351778, + "learning_rate": 8.664223423667571e-06, + "loss": 0.5072, + "step": 1717 + }, + { + "epoch": 0.31476731403444486, + "grad_norm": 0.44033356668027723, + "learning_rate": 8.662046852959644e-06, + "loss": 0.4962, + "step": 1718 + }, + { + "epoch": 0.31495053133015755, + "grad_norm": 0.4301457201335492, + "learning_rate": 8.65986878426863e-06, + "loss": 0.5034, + "step": 1719 + }, + { + "epoch": 0.3151337486258703, + "grad_norm": 0.42488867042789086, + "learning_rate": 8.657689218485476e-06, + "loss": 0.5188, + "step": 1720 + }, + { + "epoch": 0.315316965921583, + "grad_norm": 0.41341522328402974, + "learning_rate": 8.65550815650175e-06, + "loss": 0.5216, + "step": 1721 + }, + { + "epoch": 0.3155001832172957, + "grad_norm": 0.3771662909597496, + "learning_rate": 8.65332559920963e-06, + "loss": 0.5106, + "step": 1722 + }, + { + "epoch": 0.3156834005130084, + "grad_norm": 0.39212468702748254, + "learning_rate": 8.651141547501904e-06, + "loss": 0.4766, + "step": 1723 + }, + { + "epoch": 0.31586661780872116, + "grad_norm": 0.36654865806495635, + "learning_rate": 8.64895600227197e-06, + "loss": 0.4802, + "step": 1724 + }, + { + "epoch": 0.31604983510443385, + "grad_norm": 0.4065034981518445, + "learning_rate": 8.64676896441384e-06, + "loss": 0.5288, + "step": 1725 + }, + { + "epoch": 0.3162330524001466, + "grad_norm": 0.3854499856808352, + "learning_rate": 8.644580434822136e-06, + "loss": 0.4699, + "step": 1726 + }, + { + "epoch": 0.3164162696958593, + "grad_norm": 0.4057249078954745, + "learning_rate": 8.64239041439209e-06, + "loss": 0.4759, + "step": 1727 + }, + { + "epoch": 0.316599486991572, + "grad_norm": 0.44500981118266, + "learning_rate": 8.640198904019544e-06, + "loss": 0.4822, + "step": 1728 + }, + { + "epoch": 0.3167827042872847, + "grad_norm": 0.3896776477436031, + "learning_rate": 8.638005904600948e-06, + "loss": 0.49, + "step": 1729 + }, + { + "epoch": 0.3169659215829974, + "grad_norm": 0.4432593842454478, + "learning_rate": 8.635811417033361e-06, + "loss": 0.4952, + "step": 1730 + }, + { + "epoch": 0.31714913887871016, + "grad_norm": 0.38053644866499986, + "learning_rate": 8.633615442214452e-06, + "loss": 0.4579, + "step": 1731 + }, + { + "epoch": 0.31733235617442285, + "grad_norm": 0.36140726326704303, + "learning_rate": 8.6314179810425e-06, + "loss": 0.469, + "step": 1732 + }, + { + "epoch": 0.3175155734701356, + "grad_norm": 0.387898862687194, + "learning_rate": 8.62921903441639e-06, + "loss": 0.4903, + "step": 1733 + }, + { + "epoch": 0.3176987907658483, + "grad_norm": 0.4221126995607179, + "learning_rate": 8.627018603235613e-06, + "loss": 0.4959, + "step": 1734 + }, + { + "epoch": 0.317882008061561, + "grad_norm": 0.46009818640085887, + "learning_rate": 8.624816688400271e-06, + "loss": 0.4563, + "step": 1735 + }, + { + "epoch": 0.3180652253572737, + "grad_norm": 0.4685918929067162, + "learning_rate": 8.62261329081107e-06, + "loss": 0.4714, + "step": 1736 + }, + { + "epoch": 0.31824844265298646, + "grad_norm": 0.4412482439531488, + "learning_rate": 8.620408411369323e-06, + "loss": 0.5007, + "step": 1737 + }, + { + "epoch": 0.31843165994869915, + "grad_norm": 0.4547503190879881, + "learning_rate": 8.61820205097695e-06, + "loss": 0.4581, + "step": 1738 + }, + { + "epoch": 0.3186148772444119, + "grad_norm": 0.43395651034382204, + "learning_rate": 8.615994210536479e-06, + "loss": 0.49, + "step": 1739 + }, + { + "epoch": 0.3187980945401246, + "grad_norm": 0.4428071176266843, + "learning_rate": 8.613784890951036e-06, + "loss": 0.4922, + "step": 1740 + }, + { + "epoch": 0.3189813118358373, + "grad_norm": 0.40849265010723296, + "learning_rate": 8.61157409312436e-06, + "loss": 0.4743, + "step": 1741 + }, + { + "epoch": 0.31916452913155, + "grad_norm": 0.42591237999179565, + "learning_rate": 8.609361817960794e-06, + "loss": 0.505, + "step": 1742 + }, + { + "epoch": 0.3193477464272627, + "grad_norm": 0.46495694756695405, + "learning_rate": 8.607148066365278e-06, + "loss": 0.5111, + "step": 1743 + }, + { + "epoch": 0.31953096372297546, + "grad_norm": 0.3979181415316983, + "learning_rate": 8.604932839243363e-06, + "loss": 0.4911, + "step": 1744 + }, + { + "epoch": 0.31971418101868815, + "grad_norm": 0.47254311835047524, + "learning_rate": 8.602716137501201e-06, + "loss": 0.5059, + "step": 1745 + }, + { + "epoch": 0.3198973983144009, + "grad_norm": 0.4110415786461455, + "learning_rate": 8.600497962045551e-06, + "loss": 0.4931, + "step": 1746 + }, + { + "epoch": 0.3200806156101136, + "grad_norm": 0.448385298002212, + "learning_rate": 8.598278313783765e-06, + "loss": 0.4816, + "step": 1747 + }, + { + "epoch": 0.32026383290582633, + "grad_norm": 0.564876624397995, + "learning_rate": 8.59605719362381e-06, + "loss": 0.4949, + "step": 1748 + }, + { + "epoch": 0.320447050201539, + "grad_norm": 0.35471940765133153, + "learning_rate": 8.593834602474248e-06, + "loss": 0.4794, + "step": 1749 + }, + { + "epoch": 0.32063026749725176, + "grad_norm": 0.4192300945251544, + "learning_rate": 8.59161054124424e-06, + "loss": 0.4519, + "step": 1750 + }, + { + "epoch": 0.32081348479296445, + "grad_norm": 0.5038523389603609, + "learning_rate": 8.589385010843557e-06, + "loss": 0.4729, + "step": 1751 + }, + { + "epoch": 0.32099670208867714, + "grad_norm": 0.7356309775332599, + "learning_rate": 8.587158012182561e-06, + "loss": 0.5082, + "step": 1752 + }, + { + "epoch": 0.3211799193843899, + "grad_norm": 0.3818652835002793, + "learning_rate": 8.584929546172224e-06, + "loss": 0.4756, + "step": 1753 + }, + { + "epoch": 0.3213631366801026, + "grad_norm": 0.4435070016683062, + "learning_rate": 8.582699613724111e-06, + "loss": 0.5102, + "step": 1754 + }, + { + "epoch": 0.3215463539758153, + "grad_norm": 0.5377483293815659, + "learning_rate": 8.580468215750392e-06, + "loss": 0.4844, + "step": 1755 + }, + { + "epoch": 0.321729571271528, + "grad_norm": 0.4451665769582748, + "learning_rate": 8.578235353163832e-06, + "loss": 0.4617, + "step": 1756 + }, + { + "epoch": 0.32191278856724076, + "grad_norm": 0.4135616911329195, + "learning_rate": 8.576001026877802e-06, + "loss": 0.4919, + "step": 1757 + }, + { + "epoch": 0.32209600586295345, + "grad_norm": 0.4319033591486567, + "learning_rate": 8.573765237806262e-06, + "loss": 0.4699, + "step": 1758 + }, + { + "epoch": 0.3222792231586662, + "grad_norm": 0.37755085037719216, + "learning_rate": 8.571527986863775e-06, + "loss": 0.4816, + "step": 1759 + }, + { + "epoch": 0.3224624404543789, + "grad_norm": 0.3515458496554082, + "learning_rate": 8.569289274965507e-06, + "loss": 0.4742, + "step": 1760 + }, + { + "epoch": 0.32264565775009163, + "grad_norm": 0.4448770787291807, + "learning_rate": 8.567049103027215e-06, + "loss": 0.498, + "step": 1761 + }, + { + "epoch": 0.3228288750458043, + "grad_norm": 0.3941180324286139, + "learning_rate": 8.564807471965253e-06, + "loss": 0.5405, + "step": 1762 + }, + { + "epoch": 0.32301209234151707, + "grad_norm": 0.3435676251293926, + "learning_rate": 8.562564382696578e-06, + "loss": 0.4743, + "step": 1763 + }, + { + "epoch": 0.32319530963722976, + "grad_norm": 0.37988171856202396, + "learning_rate": 8.560319836138737e-06, + "loss": 0.4674, + "step": 1764 + }, + { + "epoch": 0.32337852693294244, + "grad_norm": 0.3836135165007263, + "learning_rate": 8.558073833209878e-06, + "loss": 0.4519, + "step": 1765 + }, + { + "epoch": 0.3235617442286552, + "grad_norm": 0.39708836154368815, + "learning_rate": 8.555826374828737e-06, + "loss": 0.4669, + "step": 1766 + }, + { + "epoch": 0.3237449615243679, + "grad_norm": 0.3771156156458366, + "learning_rate": 8.553577461914658e-06, + "loss": 0.4518, + "step": 1767 + }, + { + "epoch": 0.3239281788200806, + "grad_norm": 0.3598271555441401, + "learning_rate": 8.55132709538757e-06, + "loss": 0.4729, + "step": 1768 + }, + { + "epoch": 0.3241113961157933, + "grad_norm": 0.3951522532510331, + "learning_rate": 8.549075276167999e-06, + "loss": 0.4848, + "step": 1769 + }, + { + "epoch": 0.32429461341150606, + "grad_norm": 0.4147560955415563, + "learning_rate": 8.546822005177065e-06, + "loss": 0.5051, + "step": 1770 + }, + { + "epoch": 0.32447783070721875, + "grad_norm": 0.37548763357253123, + "learning_rate": 8.544567283336484e-06, + "loss": 0.4887, + "step": 1771 + }, + { + "epoch": 0.3246610480029315, + "grad_norm": 0.3562817695493491, + "learning_rate": 8.542311111568564e-06, + "loss": 0.4627, + "step": 1772 + }, + { + "epoch": 0.3248442652986442, + "grad_norm": 0.3403427009030267, + "learning_rate": 8.540053490796204e-06, + "loss": 0.498, + "step": 1773 + }, + { + "epoch": 0.32502748259435693, + "grad_norm": 0.3618093246911228, + "learning_rate": 8.5377944219429e-06, + "loss": 0.4711, + "step": 1774 + }, + { + "epoch": 0.3252106998900696, + "grad_norm": 0.37156147331448447, + "learning_rate": 8.535533905932739e-06, + "loss": 0.4914, + "step": 1775 + }, + { + "epoch": 0.3253939171857823, + "grad_norm": 0.4104697698807329, + "learning_rate": 8.533271943690397e-06, + "loss": 0.4926, + "step": 1776 + }, + { + "epoch": 0.32557713448149506, + "grad_norm": 0.4221079282420862, + "learning_rate": 8.531008536141143e-06, + "loss": 0.494, + "step": 1777 + }, + { + "epoch": 0.32576035177720775, + "grad_norm": 0.44238661768749327, + "learning_rate": 8.528743684210842e-06, + "loss": 0.4679, + "step": 1778 + }, + { + "epoch": 0.3259435690729205, + "grad_norm": 0.37351954296984535, + "learning_rate": 8.52647738882594e-06, + "loss": 0.4743, + "step": 1779 + }, + { + "epoch": 0.3261267863686332, + "grad_norm": 0.39302981254445146, + "learning_rate": 8.524209650913487e-06, + "loss": 0.4817, + "step": 1780 + }, + { + "epoch": 0.3263100036643459, + "grad_norm": 0.3421628958137739, + "learning_rate": 8.521940471401106e-06, + "loss": 0.4663, + "step": 1781 + }, + { + "epoch": 0.3264932209600586, + "grad_norm": 0.5533622907062358, + "learning_rate": 8.519669851217028e-06, + "loss": 0.4918, + "step": 1782 + }, + { + "epoch": 0.32667643825577136, + "grad_norm": 0.3970061640019179, + "learning_rate": 8.517397791290059e-06, + "loss": 0.4914, + "step": 1783 + }, + { + "epoch": 0.32685965555148405, + "grad_norm": 0.435520424701832, + "learning_rate": 8.5151242925496e-06, + "loss": 0.5093, + "step": 1784 + }, + { + "epoch": 0.3270428728471968, + "grad_norm": 0.3465454680201401, + "learning_rate": 8.512849355925641e-06, + "loss": 0.4718, + "step": 1785 + }, + { + "epoch": 0.3272260901429095, + "grad_norm": 0.4166371179539154, + "learning_rate": 8.510572982348759e-06, + "loss": 0.508, + "step": 1786 + }, + { + "epoch": 0.32740930743862223, + "grad_norm": 0.4192663642287507, + "learning_rate": 8.508295172750116e-06, + "loss": 0.4802, + "step": 1787 + }, + { + "epoch": 0.3275925247343349, + "grad_norm": 0.3935083715874497, + "learning_rate": 8.506015928061468e-06, + "loss": 0.4748, + "step": 1788 + }, + { + "epoch": 0.3277757420300476, + "grad_norm": 0.4149102123371367, + "learning_rate": 8.50373524921515e-06, + "loss": 0.4927, + "step": 1789 + }, + { + "epoch": 0.32795895932576036, + "grad_norm": 0.3867478956169408, + "learning_rate": 8.501453137144093e-06, + "loss": 0.5062, + "step": 1790 + }, + { + "epoch": 0.32814217662147305, + "grad_norm": 0.4017651304087876, + "learning_rate": 8.499169592781807e-06, + "loss": 0.5199, + "step": 1791 + }, + { + "epoch": 0.3283253939171858, + "grad_norm": 0.365617732104251, + "learning_rate": 8.496884617062389e-06, + "loss": 0.4777, + "step": 1792 + }, + { + "epoch": 0.3285086112128985, + "grad_norm": 0.5935783428527378, + "learning_rate": 8.494598210920522e-06, + "loss": 0.5327, + "step": 1793 + }, + { + "epoch": 0.32869182850861123, + "grad_norm": 0.4461319964673768, + "learning_rate": 8.492310375291478e-06, + "loss": 0.4783, + "step": 1794 + }, + { + "epoch": 0.3288750458043239, + "grad_norm": 0.4548047555505239, + "learning_rate": 8.490021111111108e-06, + "loss": 0.4793, + "step": 1795 + }, + { + "epoch": 0.32905826310003666, + "grad_norm": 0.3837539231745361, + "learning_rate": 8.48773041931585e-06, + "loss": 0.4656, + "step": 1796 + }, + { + "epoch": 0.32924148039574935, + "grad_norm": 0.38081008240354797, + "learning_rate": 8.485438300842725e-06, + "loss": 0.4538, + "step": 1797 + }, + { + "epoch": 0.3294246976914621, + "grad_norm": 0.381006774669353, + "learning_rate": 8.483144756629342e-06, + "loss": 0.4678, + "step": 1798 + }, + { + "epoch": 0.3296079149871748, + "grad_norm": 4.393546461902215, + "learning_rate": 8.480849787613883e-06, + "loss": 0.48, + "step": 1799 + }, + { + "epoch": 0.3297911322828875, + "grad_norm": 0.4216378380286295, + "learning_rate": 8.478553394735126e-06, + "loss": 0.4664, + "step": 1800 + }, + { + "epoch": 0.3299743495786002, + "grad_norm": 0.40689364633375075, + "learning_rate": 8.47625557893242e-06, + "loss": 0.4687, + "step": 1801 + }, + { + "epoch": 0.3301575668743129, + "grad_norm": 0.4232729156542768, + "learning_rate": 8.473956341145706e-06, + "loss": 0.4684, + "step": 1802 + }, + { + "epoch": 0.33034078417002566, + "grad_norm": 0.4289741259876096, + "learning_rate": 8.471655682315496e-06, + "loss": 0.4716, + "step": 1803 + }, + { + "epoch": 0.33052400146573835, + "grad_norm": 0.5275080199063381, + "learning_rate": 8.469353603382892e-06, + "loss": 0.4757, + "step": 1804 + }, + { + "epoch": 0.3307072187614511, + "grad_norm": 0.43974415262075855, + "learning_rate": 8.467050105289572e-06, + "loss": 0.4865, + "step": 1805 + }, + { + "epoch": 0.3308904360571638, + "grad_norm": 0.4814298309864084, + "learning_rate": 8.4647451889778e-06, + "loss": 0.4797, + "step": 1806 + }, + { + "epoch": 0.33107365335287653, + "grad_norm": 0.42391918940288414, + "learning_rate": 8.462438855390409e-06, + "loss": 0.4823, + "step": 1807 + }, + { + "epoch": 0.3312568706485892, + "grad_norm": 0.3665799179835301, + "learning_rate": 8.460131105470829e-06, + "loss": 0.4649, + "step": 1808 + }, + { + "epoch": 0.33144008794430196, + "grad_norm": 0.4035555626403075, + "learning_rate": 8.45782194016305e-06, + "loss": 0.4758, + "step": 1809 + }, + { + "epoch": 0.33162330524001465, + "grad_norm": 0.4425161709111113, + "learning_rate": 8.455511360411657e-06, + "loss": 0.4961, + "step": 1810 + }, + { + "epoch": 0.3318065225357274, + "grad_norm": 0.4297733230595355, + "learning_rate": 8.453199367161804e-06, + "loss": 0.4801, + "step": 1811 + }, + { + "epoch": 0.3319897398314401, + "grad_norm": 0.407563982418168, + "learning_rate": 8.450885961359227e-06, + "loss": 0.5109, + "step": 1812 + }, + { + "epoch": 0.3321729571271528, + "grad_norm": 0.40167352725847427, + "learning_rate": 8.448571143950239e-06, + "loss": 0.5056, + "step": 1813 + }, + { + "epoch": 0.3323561744228655, + "grad_norm": 0.41581106857716255, + "learning_rate": 8.446254915881733e-06, + "loss": 0.4432, + "step": 1814 + }, + { + "epoch": 0.3325393917185782, + "grad_norm": 0.42034315457885957, + "learning_rate": 8.443937278101171e-06, + "loss": 0.5129, + "step": 1815 + }, + { + "epoch": 0.33272260901429096, + "grad_norm": 0.49697266318929034, + "learning_rate": 8.441618231556604e-06, + "loss": 0.4914, + "step": 1816 + }, + { + "epoch": 0.33290582631000365, + "grad_norm": 0.45473540523372535, + "learning_rate": 8.439297777196645e-06, + "loss": 0.5046, + "step": 1817 + }, + { + "epoch": 0.3330890436057164, + "grad_norm": 0.38498739235497825, + "learning_rate": 8.436975915970496e-06, + "loss": 0.5125, + "step": 1818 + }, + { + "epoch": 0.3332722609014291, + "grad_norm": 0.3856310516540707, + "learning_rate": 8.434652648827925e-06, + "loss": 0.4877, + "step": 1819 + }, + { + "epoch": 0.33345547819714183, + "grad_norm": 0.3545408916699545, + "learning_rate": 8.432327976719281e-06, + "loss": 0.4819, + "step": 1820 + }, + { + "epoch": 0.3336386954928545, + "grad_norm": 0.3816671046542913, + "learning_rate": 8.430001900595487e-06, + "loss": 0.5074, + "step": 1821 + }, + { + "epoch": 0.33382191278856727, + "grad_norm": 0.4236478252674396, + "learning_rate": 8.427674421408037e-06, + "loss": 0.4746, + "step": 1822 + }, + { + "epoch": 0.33400513008427996, + "grad_norm": 0.4093466909489272, + "learning_rate": 8.425345540109e-06, + "loss": 0.4727, + "step": 1823 + }, + { + "epoch": 0.33418834737999265, + "grad_norm": 0.37909373859606427, + "learning_rate": 8.423015257651022e-06, + "loss": 0.482, + "step": 1824 + }, + { + "epoch": 0.3343715646757054, + "grad_norm": 0.3801638934135252, + "learning_rate": 8.420683574987319e-06, + "loss": 0.4763, + "step": 1825 + }, + { + "epoch": 0.3345547819714181, + "grad_norm": 1.0336015888999988, + "learning_rate": 8.418350493071677e-06, + "loss": 0.4797, + "step": 1826 + }, + { + "epoch": 0.3347379992671308, + "grad_norm": 0.35984342130185926, + "learning_rate": 8.41601601285846e-06, + "loss": 0.4868, + "step": 1827 + }, + { + "epoch": 0.3349212165628435, + "grad_norm": 0.3550613764764406, + "learning_rate": 8.413680135302604e-06, + "loss": 0.4706, + "step": 1828 + }, + { + "epoch": 0.33510443385855626, + "grad_norm": 0.3623869235380865, + "learning_rate": 8.411342861359612e-06, + "loss": 0.4602, + "step": 1829 + }, + { + "epoch": 0.33528765115426895, + "grad_norm": 0.4364870755091544, + "learning_rate": 8.40900419198556e-06, + "loss": 0.4832, + "step": 1830 + }, + { + "epoch": 0.3354708684499817, + "grad_norm": 0.4636618121853265, + "learning_rate": 8.406664128137095e-06, + "loss": 0.453, + "step": 1831 + }, + { + "epoch": 0.3356540857456944, + "grad_norm": 0.36221754354794844, + "learning_rate": 8.404322670771436e-06, + "loss": 0.4849, + "step": 1832 + }, + { + "epoch": 0.33583730304140713, + "grad_norm": 0.39602504845526065, + "learning_rate": 8.40197982084637e-06, + "loss": 0.4755, + "step": 1833 + }, + { + "epoch": 0.3360205203371198, + "grad_norm": 0.4091319003361619, + "learning_rate": 8.399635579320259e-06, + "loss": 0.5015, + "step": 1834 + }, + { + "epoch": 0.33620373763283257, + "grad_norm": 0.3415400300998229, + "learning_rate": 8.397289947152021e-06, + "loss": 0.4645, + "step": 1835 + }, + { + "epoch": 0.33638695492854526, + "grad_norm": 0.4295090419145507, + "learning_rate": 8.394942925301155e-06, + "loss": 0.4385, + "step": 1836 + }, + { + "epoch": 0.33657017222425795, + "grad_norm": 0.3919892057064269, + "learning_rate": 8.392594514727728e-06, + "loss": 0.5009, + "step": 1837 + }, + { + "epoch": 0.3367533895199707, + "grad_norm": 0.4276728302114804, + "learning_rate": 8.390244716392369e-06, + "loss": 0.5113, + "step": 1838 + }, + { + "epoch": 0.3369366068156834, + "grad_norm": 0.4027722155599858, + "learning_rate": 8.387893531256278e-06, + "loss": 0.4825, + "step": 1839 + }, + { + "epoch": 0.3371198241113961, + "grad_norm": 0.34856854492446304, + "learning_rate": 8.385540960281223e-06, + "loss": 0.4614, + "step": 1840 + }, + { + "epoch": 0.3373030414071088, + "grad_norm": 0.33714083093594926, + "learning_rate": 8.383187004429536e-06, + "loss": 0.4723, + "step": 1841 + }, + { + "epoch": 0.33748625870282156, + "grad_norm": 0.5015806620678309, + "learning_rate": 8.38083166466412e-06, + "loss": 0.4804, + "step": 1842 + }, + { + "epoch": 0.33766947599853425, + "grad_norm": 0.40706034865044627, + "learning_rate": 8.378474941948437e-06, + "loss": 0.4814, + "step": 1843 + }, + { + "epoch": 0.337852693294247, + "grad_norm": 0.42150894844205045, + "learning_rate": 8.376116837246525e-06, + "loss": 0.4959, + "step": 1844 + }, + { + "epoch": 0.3380359105899597, + "grad_norm": 0.37939627732061587, + "learning_rate": 8.373757351522976e-06, + "loss": 0.4737, + "step": 1845 + }, + { + "epoch": 0.33821912788567243, + "grad_norm": 0.37334004265146, + "learning_rate": 8.371396485742956e-06, + "loss": 0.4583, + "step": 1846 + }, + { + "epoch": 0.3384023451813851, + "grad_norm": 0.5502578818261286, + "learning_rate": 8.36903424087219e-06, + "loss": 0.5031, + "step": 1847 + }, + { + "epoch": 0.3385855624770978, + "grad_norm": 0.4305389526404514, + "learning_rate": 8.366670617876969e-06, + "loss": 0.5081, + "step": 1848 + }, + { + "epoch": 0.33876877977281056, + "grad_norm": 0.373003871725479, + "learning_rate": 8.36430561772415e-06, + "loss": 0.4674, + "step": 1849 + }, + { + "epoch": 0.33895199706852325, + "grad_norm": 0.36910312348131236, + "learning_rate": 8.361939241381148e-06, + "loss": 0.4824, + "step": 1850 + }, + { + "epoch": 0.339135214364236, + "grad_norm": 0.38233912239924234, + "learning_rate": 8.359571489815946e-06, + "loss": 0.5155, + "step": 1851 + }, + { + "epoch": 0.3393184316599487, + "grad_norm": 0.42080088842006386, + "learning_rate": 8.357202363997085e-06, + "loss": 0.4932, + "step": 1852 + }, + { + "epoch": 0.33950164895566143, + "grad_norm": 0.40024242741960253, + "learning_rate": 8.354831864893675e-06, + "loss": 0.4889, + "step": 1853 + }, + { + "epoch": 0.3396848662513741, + "grad_norm": 0.3818263254638861, + "learning_rate": 8.352459993475379e-06, + "loss": 0.4489, + "step": 1854 + }, + { + "epoch": 0.33986808354708686, + "grad_norm": 0.45691270540100465, + "learning_rate": 8.35008675071243e-06, + "loss": 0.445, + "step": 1855 + }, + { + "epoch": 0.34005130084279955, + "grad_norm": 0.40956822564503537, + "learning_rate": 8.347712137575614e-06, + "loss": 0.4875, + "step": 1856 + }, + { + "epoch": 0.3402345181385123, + "grad_norm": 0.4223915335148929, + "learning_rate": 8.34533615503628e-06, + "loss": 0.5017, + "step": 1857 + }, + { + "epoch": 0.340417735434225, + "grad_norm": 0.4339594623834613, + "learning_rate": 8.342958804066345e-06, + "loss": 0.5349, + "step": 1858 + }, + { + "epoch": 0.34060095272993773, + "grad_norm": 0.38617828743973426, + "learning_rate": 8.340580085638275e-06, + "loss": 0.4858, + "step": 1859 + }, + { + "epoch": 0.3407841700256504, + "grad_norm": 0.39545092131685405, + "learning_rate": 8.3382000007251e-06, + "loss": 0.4844, + "step": 1860 + }, + { + "epoch": 0.3409673873213631, + "grad_norm": 0.4257990961795386, + "learning_rate": 8.33581855030041e-06, + "loss": 0.5055, + "step": 1861 + }, + { + "epoch": 0.34115060461707586, + "grad_norm": 0.3528695250715933, + "learning_rate": 8.333435735338346e-06, + "loss": 0.4875, + "step": 1862 + }, + { + "epoch": 0.34133382191278855, + "grad_norm": 0.37091916482456505, + "learning_rate": 8.331051556813623e-06, + "loss": 0.4984, + "step": 1863 + }, + { + "epoch": 0.3415170392085013, + "grad_norm": 0.4264847608510263, + "learning_rate": 8.328666015701496e-06, + "loss": 0.4867, + "step": 1864 + }, + { + "epoch": 0.341700256504214, + "grad_norm": 0.4099601844664771, + "learning_rate": 8.326279112977791e-06, + "loss": 0.4994, + "step": 1865 + }, + { + "epoch": 0.34188347379992673, + "grad_norm": 0.3677590723427411, + "learning_rate": 8.323890849618882e-06, + "loss": 0.5027, + "step": 1866 + }, + { + "epoch": 0.3420666910956394, + "grad_norm": 0.4409098617591673, + "learning_rate": 8.321501226601702e-06, + "loss": 0.496, + "step": 1867 + }, + { + "epoch": 0.34224990839135216, + "grad_norm": 0.3926664857139486, + "learning_rate": 8.319110244903748e-06, + "loss": 0.5018, + "step": 1868 + }, + { + "epoch": 0.34243312568706485, + "grad_norm": 0.3981457421214149, + "learning_rate": 8.316717905503058e-06, + "loss": 0.4803, + "step": 1869 + }, + { + "epoch": 0.3426163429827776, + "grad_norm": 0.4111803238988, + "learning_rate": 8.314324209378237e-06, + "loss": 0.4944, + "step": 1870 + }, + { + "epoch": 0.3427995602784903, + "grad_norm": 0.3837210327181565, + "learning_rate": 8.31192915750844e-06, + "loss": 0.4637, + "step": 1871 + }, + { + "epoch": 0.342982777574203, + "grad_norm": 0.36141808942109716, + "learning_rate": 8.309532750873381e-06, + "loss": 0.5022, + "step": 1872 + }, + { + "epoch": 0.3431659948699157, + "grad_norm": 0.4135486922923667, + "learning_rate": 8.30713499045332e-06, + "loss": 0.5429, + "step": 1873 + }, + { + "epoch": 0.3433492121656284, + "grad_norm": 0.4100838578777171, + "learning_rate": 8.304735877229082e-06, + "loss": 0.4814, + "step": 1874 + }, + { + "epoch": 0.34353242946134116, + "grad_norm": 0.4419756386117236, + "learning_rate": 8.302335412182034e-06, + "loss": 0.506, + "step": 1875 + }, + { + "epoch": 0.34371564675705385, + "grad_norm": 0.4023978864041945, + "learning_rate": 8.299933596294104e-06, + "loss": 0.4835, + "step": 1876 + }, + { + "epoch": 0.3438988640527666, + "grad_norm": 0.37938282086115893, + "learning_rate": 8.297530430547767e-06, + "loss": 0.4802, + "step": 1877 + }, + { + "epoch": 0.3440820813484793, + "grad_norm": 0.44808143362476516, + "learning_rate": 8.295125915926057e-06, + "loss": 0.4948, + "step": 1878 + }, + { + "epoch": 0.34426529864419203, + "grad_norm": 0.38559125969803254, + "learning_rate": 8.292720053412553e-06, + "loss": 0.4685, + "step": 1879 + }, + { + "epoch": 0.3444485159399047, + "grad_norm": 0.33369724182793525, + "learning_rate": 8.290312843991388e-06, + "loss": 0.4395, + "step": 1880 + }, + { + "epoch": 0.34463173323561747, + "grad_norm": 0.36560725000329003, + "learning_rate": 8.287904288647246e-06, + "loss": 0.4769, + "step": 1881 + }, + { + "epoch": 0.34481495053133016, + "grad_norm": 0.40109807099240397, + "learning_rate": 8.285494388365364e-06, + "loss": 0.4896, + "step": 1882 + }, + { + "epoch": 0.34499816782704285, + "grad_norm": 0.3758614423812384, + "learning_rate": 8.283083144131523e-06, + "loss": 0.4591, + "step": 1883 + }, + { + "epoch": 0.3451813851227556, + "grad_norm": 0.4045825714686613, + "learning_rate": 8.28067055693206e-06, + "loss": 0.4983, + "step": 1884 + }, + { + "epoch": 0.3453646024184683, + "grad_norm": 0.5653539162553293, + "learning_rate": 8.278256627753857e-06, + "loss": 0.4912, + "step": 1885 + }, + { + "epoch": 0.345547819714181, + "grad_norm": 0.4120278659351058, + "learning_rate": 8.27584135758435e-06, + "loss": 0.5053, + "step": 1886 + }, + { + "epoch": 0.3457310370098937, + "grad_norm": 0.4087683124654251, + "learning_rate": 8.273424747411519e-06, + "loss": 0.4768, + "step": 1887 + }, + { + "epoch": 0.34591425430560646, + "grad_norm": 0.3868920894610502, + "learning_rate": 8.27100679822389e-06, + "loss": 0.4872, + "step": 1888 + }, + { + "epoch": 0.34609747160131915, + "grad_norm": 0.3672508442805955, + "learning_rate": 8.268587511010546e-06, + "loss": 0.4492, + "step": 1889 + }, + { + "epoch": 0.3462806888970319, + "grad_norm": 0.3226405339400022, + "learning_rate": 8.266166886761106e-06, + "loss": 0.447, + "step": 1890 + }, + { + "epoch": 0.3464639061927446, + "grad_norm": 0.3896492518387063, + "learning_rate": 8.263744926465744e-06, + "loss": 0.4952, + "step": 1891 + }, + { + "epoch": 0.34664712348845733, + "grad_norm": 0.40806210827621486, + "learning_rate": 8.26132163111518e-06, + "loss": 0.4942, + "step": 1892 + }, + { + "epoch": 0.34683034078417, + "grad_norm": 0.5064133110397161, + "learning_rate": 8.258897001700673e-06, + "loss": 0.4914, + "step": 1893 + }, + { + "epoch": 0.34701355807988277, + "grad_norm": 0.37350719921589426, + "learning_rate": 8.256471039214036e-06, + "loss": 0.4667, + "step": 1894 + }, + { + "epoch": 0.34719677537559546, + "grad_norm": 0.4472694649954096, + "learning_rate": 8.254043744647625e-06, + "loss": 0.4768, + "step": 1895 + }, + { + "epoch": 0.34737999267130815, + "grad_norm": 0.5451471608365284, + "learning_rate": 8.251615118994338e-06, + "loss": 0.4833, + "step": 1896 + }, + { + "epoch": 0.3475632099670209, + "grad_norm": 0.43194887420538414, + "learning_rate": 8.249185163247621e-06, + "loss": 0.5041, + "step": 1897 + }, + { + "epoch": 0.3477464272627336, + "grad_norm": 0.4440434179645274, + "learning_rate": 8.24675387840146e-06, + "loss": 0.4928, + "step": 1898 + }, + { + "epoch": 0.3479296445584463, + "grad_norm": 0.4470574728369111, + "learning_rate": 8.24432126545039e-06, + "loss": 0.4721, + "step": 1899 + }, + { + "epoch": 0.348112861854159, + "grad_norm": 0.37811511446638624, + "learning_rate": 8.241887325389486e-06, + "loss": 0.4916, + "step": 1900 + }, + { + "epoch": 0.34829607914987176, + "grad_norm": 0.401475995772846, + "learning_rate": 8.239452059214367e-06, + "loss": 0.4922, + "step": 1901 + }, + { + "epoch": 0.34847929644558445, + "grad_norm": 0.36588179778579716, + "learning_rate": 8.23701546792119e-06, + "loss": 0.4769, + "step": 1902 + }, + { + "epoch": 0.3486625137412972, + "grad_norm": 0.37419447204674516, + "learning_rate": 8.234577552506662e-06, + "loss": 0.4683, + "step": 1903 + }, + { + "epoch": 0.3488457310370099, + "grad_norm": 0.44910824354183104, + "learning_rate": 8.232138313968025e-06, + "loss": 0.5002, + "step": 1904 + }, + { + "epoch": 0.34902894833272263, + "grad_norm": 0.37853030952448913, + "learning_rate": 8.229697753303067e-06, + "loss": 0.499, + "step": 1905 + }, + { + "epoch": 0.3492121656284353, + "grad_norm": 0.3787302843131875, + "learning_rate": 8.227255871510111e-06, + "loss": 0.4652, + "step": 1906 + }, + { + "epoch": 0.349395382924148, + "grad_norm": 0.42294805152134374, + "learning_rate": 8.224812669588028e-06, + "loss": 0.4761, + "step": 1907 + }, + { + "epoch": 0.34957860021986076, + "grad_norm": 0.4251808716753822, + "learning_rate": 8.222368148536223e-06, + "loss": 0.4781, + "step": 1908 + }, + { + "epoch": 0.34976181751557345, + "grad_norm": 0.41741877899455326, + "learning_rate": 8.219922309354643e-06, + "loss": 0.4979, + "step": 1909 + }, + { + "epoch": 0.3499450348112862, + "grad_norm": 0.4528858990840802, + "learning_rate": 8.217475153043772e-06, + "loss": 0.4841, + "step": 1910 + }, + { + "epoch": 0.3501282521069989, + "grad_norm": 0.37317689057150794, + "learning_rate": 8.215026680604637e-06, + "loss": 0.4713, + "step": 1911 + }, + { + "epoch": 0.35031146940271163, + "grad_norm": 0.3667470219941718, + "learning_rate": 8.212576893038799e-06, + "loss": 0.4689, + "step": 1912 + }, + { + "epoch": 0.3504946866984243, + "grad_norm": 0.3979696962375165, + "learning_rate": 8.21012579134836e-06, + "loss": 0.4538, + "step": 1913 + }, + { + "epoch": 0.35067790399413706, + "grad_norm": 0.43331562093452664, + "learning_rate": 8.20767337653596e-06, + "loss": 0.4878, + "step": 1914 + }, + { + "epoch": 0.35086112128984975, + "grad_norm": 0.36256762359906813, + "learning_rate": 8.20521964960477e-06, + "loss": 0.4462, + "step": 1915 + }, + { + "epoch": 0.3510443385855625, + "grad_norm": 0.43213117928406414, + "learning_rate": 8.202764611558507e-06, + "loss": 0.5066, + "step": 1916 + }, + { + "epoch": 0.3512275558812752, + "grad_norm": 0.4293633236485711, + "learning_rate": 8.200308263401417e-06, + "loss": 0.5351, + "step": 1917 + }, + { + "epoch": 0.35141077317698793, + "grad_norm": 0.5287407111719371, + "learning_rate": 8.197850606138286e-06, + "loss": 0.4705, + "step": 1918 + }, + { + "epoch": 0.3515939904727006, + "grad_norm": 0.48855534444525384, + "learning_rate": 8.195391640774433e-06, + "loss": 0.4852, + "step": 1919 + }, + { + "epoch": 0.3517772077684133, + "grad_norm": 0.4148812235933253, + "learning_rate": 8.192931368315715e-06, + "loss": 0.4785, + "step": 1920 + }, + { + "epoch": 0.35196042506412606, + "grad_norm": 0.44919257408205365, + "learning_rate": 8.190469789768517e-06, + "loss": 0.4851, + "step": 1921 + }, + { + "epoch": 0.35214364235983875, + "grad_norm": 0.4359193627784823, + "learning_rate": 8.188006906139767e-06, + "loss": 0.5077, + "step": 1922 + }, + { + "epoch": 0.3523268596555515, + "grad_norm": 0.43124736358934584, + "learning_rate": 8.185542718436923e-06, + "loss": 0.4705, + "step": 1923 + }, + { + "epoch": 0.3525100769512642, + "grad_norm": 0.4270332132061218, + "learning_rate": 8.183077227667975e-06, + "loss": 0.5061, + "step": 1924 + }, + { + "epoch": 0.35269329424697693, + "grad_norm": 0.3709113632371592, + "learning_rate": 8.180610434841448e-06, + "loss": 0.4926, + "step": 1925 + }, + { + "epoch": 0.3528765115426896, + "grad_norm": 0.3686143335119551, + "learning_rate": 8.1781423409664e-06, + "loss": 0.4736, + "step": 1926 + }, + { + "epoch": 0.35305972883840236, + "grad_norm": 0.4768757941328196, + "learning_rate": 8.175672947052416e-06, + "loss": 0.4899, + "step": 1927 + }, + { + "epoch": 0.35324294613411505, + "grad_norm": 0.37435018310100715, + "learning_rate": 8.173202254109622e-06, + "loss": 0.5019, + "step": 1928 + }, + { + "epoch": 0.3534261634298278, + "grad_norm": 0.5886108613911327, + "learning_rate": 8.170730263148668e-06, + "loss": 0.5208, + "step": 1929 + }, + { + "epoch": 0.3536093807255405, + "grad_norm": 0.42394082862501326, + "learning_rate": 8.168256975180737e-06, + "loss": 0.5124, + "step": 1930 + }, + { + "epoch": 0.3537925980212532, + "grad_norm": 0.39473180614386977, + "learning_rate": 8.165782391217543e-06, + "loss": 0.4863, + "step": 1931 + }, + { + "epoch": 0.3539758153169659, + "grad_norm": 0.4022544573376633, + "learning_rate": 8.163306512271334e-06, + "loss": 0.4989, + "step": 1932 + }, + { + "epoch": 0.3541590326126786, + "grad_norm": 0.4154530765976519, + "learning_rate": 8.160829339354876e-06, + "loss": 0.4661, + "step": 1933 + }, + { + "epoch": 0.35434224990839136, + "grad_norm": 0.3921468409429114, + "learning_rate": 8.158350873481478e-06, + "loss": 0.4738, + "step": 1934 + }, + { + "epoch": 0.35452546720410405, + "grad_norm": 0.3954474064023121, + "learning_rate": 8.155871115664968e-06, + "loss": 0.51, + "step": 1935 + }, + { + "epoch": 0.3547086844998168, + "grad_norm": 0.3882950367284415, + "learning_rate": 8.15339006691971e-06, + "loss": 0.4642, + "step": 1936 + }, + { + "epoch": 0.3548919017955295, + "grad_norm": 0.37897867702653376, + "learning_rate": 8.150907728260592e-06, + "loss": 0.5044, + "step": 1937 + }, + { + "epoch": 0.35507511909124223, + "grad_norm": 0.3859422166696926, + "learning_rate": 8.14842410070303e-06, + "loss": 0.4551, + "step": 1938 + }, + { + "epoch": 0.3552583363869549, + "grad_norm": 0.38875341973643246, + "learning_rate": 8.145939185262963e-06, + "loss": 0.4944, + "step": 1939 + }, + { + "epoch": 0.35544155368266767, + "grad_norm": 0.3983931627212232, + "learning_rate": 8.143452982956866e-06, + "loss": 0.5008, + "step": 1940 + }, + { + "epoch": 0.35562477097838036, + "grad_norm": 0.39775757772326403, + "learning_rate": 8.140965494801733e-06, + "loss": 0.5263, + "step": 1941 + }, + { + "epoch": 0.3558079882740931, + "grad_norm": 0.3944293856829923, + "learning_rate": 8.13847672181509e-06, + "loss": 0.4771, + "step": 1942 + }, + { + "epoch": 0.3559912055698058, + "grad_norm": 0.3817860113481927, + "learning_rate": 8.13598666501498e-06, + "loss": 0.4666, + "step": 1943 + }, + { + "epoch": 0.3561744228655185, + "grad_norm": 0.42632726063844545, + "learning_rate": 8.133495325419983e-06, + "loss": 0.4663, + "step": 1944 + }, + { + "epoch": 0.3563576401612312, + "grad_norm": 0.3572805692583185, + "learning_rate": 8.131002704049189e-06, + "loss": 0.4708, + "step": 1945 + }, + { + "epoch": 0.3565408574569439, + "grad_norm": 0.404852300953083, + "learning_rate": 8.128508801922226e-06, + "loss": 0.5325, + "step": 1946 + }, + { + "epoch": 0.35672407475265666, + "grad_norm": 0.35876468020600716, + "learning_rate": 8.126013620059236e-06, + "loss": 0.4949, + "step": 1947 + }, + { + "epoch": 0.35690729204836935, + "grad_norm": 0.38545171520037624, + "learning_rate": 8.123517159480894e-06, + "loss": 0.4764, + "step": 1948 + }, + { + "epoch": 0.3570905093440821, + "grad_norm": 0.40541828483545733, + "learning_rate": 8.12101942120839e-06, + "loss": 0.4807, + "step": 1949 + }, + { + "epoch": 0.3572737266397948, + "grad_norm": 0.4243415026003622, + "learning_rate": 8.118520406263437e-06, + "loss": 0.5003, + "step": 1950 + }, + { + "epoch": 0.35745694393550753, + "grad_norm": 0.4384025995443247, + "learning_rate": 8.116020115668278e-06, + "loss": 0.4916, + "step": 1951 + }, + { + "epoch": 0.3576401612312202, + "grad_norm": 0.356464489561591, + "learning_rate": 8.113518550445667e-06, + "loss": 0.5068, + "step": 1952 + }, + { + "epoch": 0.35782337852693297, + "grad_norm": 0.36237464089967664, + "learning_rate": 8.111015711618888e-06, + "loss": 0.4691, + "step": 1953 + }, + { + "epoch": 0.35800659582264566, + "grad_norm": 0.41004625226837194, + "learning_rate": 8.108511600211741e-06, + "loss": 0.495, + "step": 1954 + }, + { + "epoch": 0.35818981311835835, + "grad_norm": 0.37519930557205516, + "learning_rate": 8.106006217248552e-06, + "loss": 0.527, + "step": 1955 + }, + { + "epoch": 0.3583730304140711, + "grad_norm": 0.3930139446184139, + "learning_rate": 8.103499563754159e-06, + "loss": 0.49, + "step": 1956 + }, + { + "epoch": 0.3585562477097838, + "grad_norm": 0.4049543770945572, + "learning_rate": 8.100991640753926e-06, + "loss": 0.4918, + "step": 1957 + }, + { + "epoch": 0.3587394650054965, + "grad_norm": 0.41598194124283755, + "learning_rate": 8.098482449273737e-06, + "loss": 0.4583, + "step": 1958 + }, + { + "epoch": 0.3589226823012092, + "grad_norm": 0.3497094599476565, + "learning_rate": 8.095971990339987e-06, + "loss": 0.441, + "step": 1959 + }, + { + "epoch": 0.35910589959692196, + "grad_norm": 0.857453529704147, + "learning_rate": 8.0934602649796e-06, + "loss": 0.4862, + "step": 1960 + }, + { + "epoch": 0.35928911689263465, + "grad_norm": 0.3948051526194389, + "learning_rate": 8.090947274220011e-06, + "loss": 0.4696, + "step": 1961 + }, + { + "epoch": 0.3594723341883474, + "grad_norm": 0.4025724340675242, + "learning_rate": 8.088433019089174e-06, + "loss": 0.4784, + "step": 1962 + }, + { + "epoch": 0.3596555514840601, + "grad_norm": 0.4250158054819026, + "learning_rate": 8.08591750061556e-06, + "loss": 0.4887, + "step": 1963 + }, + { + "epoch": 0.35983876877977283, + "grad_norm": 0.3987052277731631, + "learning_rate": 8.083400719828161e-06, + "loss": 0.486, + "step": 1964 + }, + { + "epoch": 0.3600219860754855, + "grad_norm": 0.3818650890686896, + "learning_rate": 8.080882677756479e-06, + "loss": 0.4961, + "step": 1965 + }, + { + "epoch": 0.36020520337119827, + "grad_norm": 0.37338678994677815, + "learning_rate": 8.078363375430534e-06, + "loss": 0.4849, + "step": 1966 + }, + { + "epoch": 0.36038842066691096, + "grad_norm": 0.4033555520259992, + "learning_rate": 8.075842813880865e-06, + "loss": 0.5092, + "step": 1967 + }, + { + "epoch": 0.36057163796262365, + "grad_norm": 0.3904145734640018, + "learning_rate": 8.073320994138522e-06, + "loss": 0.498, + "step": 1968 + }, + { + "epoch": 0.3607548552583364, + "grad_norm": 0.41968588248870525, + "learning_rate": 8.070797917235071e-06, + "loss": 0.4592, + "step": 1969 + }, + { + "epoch": 0.3609380725540491, + "grad_norm": 0.40225333028901955, + "learning_rate": 8.068273584202593e-06, + "loss": 0.498, + "step": 1970 + }, + { + "epoch": 0.36112128984976183, + "grad_norm": 0.4779749898506442, + "learning_rate": 8.065747996073681e-06, + "loss": 0.4852, + "step": 1971 + }, + { + "epoch": 0.3613045071454745, + "grad_norm": 0.3801653362164248, + "learning_rate": 8.063221153881443e-06, + "loss": 0.4821, + "step": 1972 + }, + { + "epoch": 0.36148772444118726, + "grad_norm": 0.4360824877290448, + "learning_rate": 8.0606930586595e-06, + "loss": 0.4867, + "step": 1973 + }, + { + "epoch": 0.36167094173689995, + "grad_norm": 0.4253700547037495, + "learning_rate": 8.058163711441986e-06, + "loss": 0.4829, + "step": 1974 + }, + { + "epoch": 0.3618541590326127, + "grad_norm": 0.4341340818854911, + "learning_rate": 8.055633113263543e-06, + "loss": 0.5169, + "step": 1975 + }, + { + "epoch": 0.3620373763283254, + "grad_norm": 0.4424513726038396, + "learning_rate": 8.053101265159331e-06, + "loss": 0.5005, + "step": 1976 + }, + { + "epoch": 0.36222059362403813, + "grad_norm": 0.40080376851886934, + "learning_rate": 8.050568168165018e-06, + "loss": 0.4831, + "step": 1977 + }, + { + "epoch": 0.3624038109197508, + "grad_norm": 0.4092773451489212, + "learning_rate": 8.048033823316784e-06, + "loss": 0.4886, + "step": 1978 + }, + { + "epoch": 0.3625870282154635, + "grad_norm": 0.5750732156961504, + "learning_rate": 8.045498231651314e-06, + "loss": 0.4924, + "step": 1979 + }, + { + "epoch": 0.36277024551117626, + "grad_norm": 0.4000642911770352, + "learning_rate": 8.042961394205812e-06, + "loss": 0.4856, + "step": 1980 + }, + { + "epoch": 0.36295346280688895, + "grad_norm": 0.38938145758582615, + "learning_rate": 8.040423312017986e-06, + "loss": 0.4778, + "step": 1981 + }, + { + "epoch": 0.3631366801026017, + "grad_norm": 0.3643427586018773, + "learning_rate": 8.037883986126054e-06, + "loss": 0.4909, + "step": 1982 + }, + { + "epoch": 0.3633198973983144, + "grad_norm": 0.3939262969339654, + "learning_rate": 8.035343417568742e-06, + "loss": 0.4714, + "step": 1983 + }, + { + "epoch": 0.36350311469402713, + "grad_norm": 0.4016181460823361, + "learning_rate": 8.032801607385288e-06, + "loss": 0.4931, + "step": 1984 + }, + { + "epoch": 0.3636863319897398, + "grad_norm": 0.44751708009914176, + "learning_rate": 8.030258556615433e-06, + "loss": 0.4894, + "step": 1985 + }, + { + "epoch": 0.36386954928545256, + "grad_norm": 0.34958387993124196, + "learning_rate": 8.02771426629943e-06, + "loss": 0.4598, + "step": 1986 + }, + { + "epoch": 0.36405276658116525, + "grad_norm": 0.348786463445316, + "learning_rate": 8.025168737478034e-06, + "loss": 0.447, + "step": 1987 + }, + { + "epoch": 0.364235983876878, + "grad_norm": 0.3854079222539858, + "learning_rate": 8.022621971192513e-06, + "loss": 0.4812, + "step": 1988 + }, + { + "epoch": 0.3644192011725907, + "grad_norm": 0.38734200128527324, + "learning_rate": 8.020073968484632e-06, + "loss": 0.4862, + "step": 1989 + }, + { + "epoch": 0.36460241846830344, + "grad_norm": 0.411711055657578, + "learning_rate": 8.017524730396673e-06, + "loss": 0.4907, + "step": 1990 + }, + { + "epoch": 0.3647856357640161, + "grad_norm": 0.440618505246582, + "learning_rate": 8.014974257971415e-06, + "loss": 0.4588, + "step": 1991 + }, + { + "epoch": 0.3649688530597288, + "grad_norm": 0.4841735582737877, + "learning_rate": 8.012422552252148e-06, + "loss": 0.4988, + "step": 1992 + }, + { + "epoch": 0.36515207035544156, + "grad_norm": 0.35566001438725936, + "learning_rate": 8.009869614282657e-06, + "loss": 0.5083, + "step": 1993 + }, + { + "epoch": 0.36533528765115425, + "grad_norm": 0.3607042141519043, + "learning_rate": 8.007315445107242e-06, + "loss": 0.4634, + "step": 1994 + }, + { + "epoch": 0.365518504946867, + "grad_norm": 0.3841642715938066, + "learning_rate": 8.004760045770702e-06, + "loss": 0.4972, + "step": 1995 + }, + { + "epoch": 0.3657017222425797, + "grad_norm": 0.39468083856708197, + "learning_rate": 8.002203417318335e-06, + "loss": 0.4779, + "step": 1996 + }, + { + "epoch": 0.36588493953829243, + "grad_norm": 0.419151699921075, + "learning_rate": 7.999645560795947e-06, + "loss": 0.5039, + "step": 1997 + }, + { + "epoch": 0.3660681568340051, + "grad_norm": 0.4148076872613475, + "learning_rate": 7.99708647724985e-06, + "loss": 0.4969, + "step": 1998 + }, + { + "epoch": 0.36625137412971787, + "grad_norm": 0.37720877965942756, + "learning_rate": 7.994526167726847e-06, + "loss": 0.4657, + "step": 1999 + }, + { + "epoch": 0.36643459142543056, + "grad_norm": 0.3456057298421402, + "learning_rate": 7.991964633274255e-06, + "loss": 0.5015, + "step": 2000 + }, + { + "epoch": 0.3666178087211433, + "grad_norm": 0.44797110483125663, + "learning_rate": 7.989401874939877e-06, + "loss": 0.5106, + "step": 2001 + }, + { + "epoch": 0.366801026016856, + "grad_norm": 0.4272501663842639, + "learning_rate": 7.986837893772031e-06, + "loss": 0.4783, + "step": 2002 + }, + { + "epoch": 0.3669842433125687, + "grad_norm": 0.5436758459964995, + "learning_rate": 7.98427269081953e-06, + "loss": 0.535, + "step": 2003 + }, + { + "epoch": 0.3671674606082814, + "grad_norm": 0.4193212630668118, + "learning_rate": 7.981706267131684e-06, + "loss": 0.4893, + "step": 2004 + }, + { + "epoch": 0.3673506779039941, + "grad_norm": 0.42948677992300865, + "learning_rate": 7.979138623758305e-06, + "loss": 0.4925, + "step": 2005 + }, + { + "epoch": 0.36753389519970686, + "grad_norm": 0.38906501628138734, + "learning_rate": 7.976569761749704e-06, + "loss": 0.4864, + "step": 2006 + }, + { + "epoch": 0.36771711249541955, + "grad_norm": 0.3905681742731238, + "learning_rate": 7.973999682156689e-06, + "loss": 0.4814, + "step": 2007 + }, + { + "epoch": 0.3679003297911323, + "grad_norm": 0.384284775034963, + "learning_rate": 7.97142838603057e-06, + "loss": 0.4903, + "step": 2008 + }, + { + "epoch": 0.368083547086845, + "grad_norm": 0.42047182046573356, + "learning_rate": 7.968855874423146e-06, + "loss": 0.4874, + "step": 2009 + }, + { + "epoch": 0.36826676438255773, + "grad_norm": 0.45925338817929307, + "learning_rate": 7.966282148386726e-06, + "loss": 0.4921, + "step": 2010 + }, + { + "epoch": 0.3684499816782704, + "grad_norm": 0.4327334130084159, + "learning_rate": 7.963707208974103e-06, + "loss": 0.4939, + "step": 2011 + }, + { + "epoch": 0.36863319897398317, + "grad_norm": 0.3885802329431291, + "learning_rate": 7.961131057238576e-06, + "loss": 0.4747, + "step": 2012 + }, + { + "epoch": 0.36881641626969586, + "grad_norm": 0.421099560605058, + "learning_rate": 7.958553694233933e-06, + "loss": 0.4765, + "step": 2013 + }, + { + "epoch": 0.36899963356540855, + "grad_norm": 0.388202635805495, + "learning_rate": 7.955975121014462e-06, + "loss": 0.4649, + "step": 2014 + }, + { + "epoch": 0.3691828508611213, + "grad_norm": 0.3501587036877818, + "learning_rate": 7.953395338634943e-06, + "loss": 0.4592, + "step": 2015 + }, + { + "epoch": 0.369366068156834, + "grad_norm": 0.3889685256407414, + "learning_rate": 7.950814348150655e-06, + "loss": 0.4884, + "step": 2016 + }, + { + "epoch": 0.3695492854525467, + "grad_norm": 0.3726532989515501, + "learning_rate": 7.948232150617365e-06, + "loss": 0.4809, + "step": 2017 + }, + { + "epoch": 0.3697325027482594, + "grad_norm": 0.4258326153743124, + "learning_rate": 7.945648747091339e-06, + "loss": 0.5255, + "step": 2018 + }, + { + "epoch": 0.36991572004397216, + "grad_norm": 0.422873775241114, + "learning_rate": 7.943064138629332e-06, + "loss": 0.4835, + "step": 2019 + }, + { + "epoch": 0.37009893733968485, + "grad_norm": 0.3854940689794936, + "learning_rate": 7.940478326288597e-06, + "loss": 0.5199, + "step": 2020 + }, + { + "epoch": 0.3702821546353976, + "grad_norm": 0.4090502879792355, + "learning_rate": 7.937891311126875e-06, + "loss": 0.4968, + "step": 2021 + }, + { + "epoch": 0.3704653719311103, + "grad_norm": 0.5672840308934909, + "learning_rate": 7.935303094202402e-06, + "loss": 0.5074, + "step": 2022 + }, + { + "epoch": 0.37064858922682303, + "grad_norm": 0.43313528860057604, + "learning_rate": 7.932713676573902e-06, + "loss": 0.5185, + "step": 2023 + }, + { + "epoch": 0.3708318065225357, + "grad_norm": 0.3999275079366991, + "learning_rate": 7.930123059300595e-06, + "loss": 0.4662, + "step": 2024 + }, + { + "epoch": 0.37101502381824847, + "grad_norm": 0.4016179992674339, + "learning_rate": 7.927531243442188e-06, + "loss": 0.484, + "step": 2025 + }, + { + "epoch": 0.37119824111396116, + "grad_norm": 0.4280944544354096, + "learning_rate": 7.924938230058878e-06, + "loss": 0.4851, + "step": 2026 + }, + { + "epoch": 0.37138145840967385, + "grad_norm": 0.43747500513006565, + "learning_rate": 7.922344020211357e-06, + "loss": 0.4981, + "step": 2027 + }, + { + "epoch": 0.3715646757053866, + "grad_norm": 0.36930540160986264, + "learning_rate": 7.919748614960798e-06, + "loss": 0.4593, + "step": 2028 + }, + { + "epoch": 0.3717478930010993, + "grad_norm": 0.37696526354057724, + "learning_rate": 7.917152015368872e-06, + "loss": 0.4713, + "step": 2029 + }, + { + "epoch": 0.37193111029681203, + "grad_norm": 0.3512686254761771, + "learning_rate": 7.914554222497732e-06, + "loss": 0.4803, + "step": 2030 + }, + { + "epoch": 0.3721143275925247, + "grad_norm": 0.3999987608614825, + "learning_rate": 7.91195523741002e-06, + "loss": 0.4616, + "step": 2031 + }, + { + "epoch": 0.37229754488823746, + "grad_norm": 0.3724358403073566, + "learning_rate": 7.90935506116887e-06, + "loss": 0.4765, + "step": 2032 + }, + { + "epoch": 0.37248076218395015, + "grad_norm": 0.5231754999712253, + "learning_rate": 7.906753694837898e-06, + "loss": 0.4577, + "step": 2033 + }, + { + "epoch": 0.3726639794796629, + "grad_norm": 0.41538678668159756, + "learning_rate": 7.904151139481213e-06, + "loss": 0.4695, + "step": 2034 + }, + { + "epoch": 0.3728471967753756, + "grad_norm": 0.3669187862199872, + "learning_rate": 7.9015473961634e-06, + "loss": 0.4457, + "step": 2035 + }, + { + "epoch": 0.37303041407108833, + "grad_norm": 0.39716872566616657, + "learning_rate": 7.898942465949543e-06, + "loss": 0.4769, + "step": 2036 + }, + { + "epoch": 0.373213631366801, + "grad_norm": 0.39707124935919325, + "learning_rate": 7.8963363499052e-06, + "loss": 0.5038, + "step": 2037 + }, + { + "epoch": 0.3733968486625137, + "grad_norm": 0.5407284529620036, + "learning_rate": 7.893729049096423e-06, + "loss": 0.4883, + "step": 2038 + }, + { + "epoch": 0.37358006595822646, + "grad_norm": 0.4434140600327088, + "learning_rate": 7.891120564589741e-06, + "loss": 0.4797, + "step": 2039 + }, + { + "epoch": 0.37376328325393915, + "grad_norm": 0.3669111491685124, + "learning_rate": 7.888510897452173e-06, + "loss": 0.4542, + "step": 2040 + }, + { + "epoch": 0.3739465005496519, + "grad_norm": 0.39292395564183696, + "learning_rate": 7.885900048751219e-06, + "loss": 0.4843, + "step": 2041 + }, + { + "epoch": 0.3741297178453646, + "grad_norm": 0.42693187771524804, + "learning_rate": 7.883288019554864e-06, + "loss": 0.5163, + "step": 2042 + }, + { + "epoch": 0.37431293514107733, + "grad_norm": 0.4166082295492774, + "learning_rate": 7.880674810931572e-06, + "loss": 0.5167, + "step": 2043 + }, + { + "epoch": 0.37449615243679, + "grad_norm": 0.388587145431864, + "learning_rate": 7.878060423950296e-06, + "loss": 0.4696, + "step": 2044 + }, + { + "epoch": 0.37467936973250277, + "grad_norm": 0.3907542736094849, + "learning_rate": 7.875444859680463e-06, + "loss": 0.5188, + "step": 2045 + }, + { + "epoch": 0.37486258702821545, + "grad_norm": 0.35105779035513274, + "learning_rate": 7.872828119191991e-06, + "loss": 0.4819, + "step": 2046 + }, + { + "epoch": 0.3750458043239282, + "grad_norm": 0.387372551594969, + "learning_rate": 7.87021020355527e-06, + "loss": 0.4888, + "step": 2047 + }, + { + "epoch": 0.3752290216196409, + "grad_norm": 0.3925107894377394, + "learning_rate": 7.867591113841177e-06, + "loss": 0.4784, + "step": 2048 + }, + { + "epoch": 0.37541223891535364, + "grad_norm": 0.3765661581212424, + "learning_rate": 7.864970851121065e-06, + "loss": 0.4718, + "step": 2049 + }, + { + "epoch": 0.3755954562110663, + "grad_norm": 0.3358655051277223, + "learning_rate": 7.86234941646677e-06, + "loss": 0.4422, + "step": 2050 + }, + { + "epoch": 0.375778673506779, + "grad_norm": 0.40689174811822976, + "learning_rate": 7.859726810950606e-06, + "loss": 0.5066, + "step": 2051 + }, + { + "epoch": 0.37596189080249176, + "grad_norm": 0.3400880833533205, + "learning_rate": 7.857103035645365e-06, + "loss": 0.4768, + "step": 2052 + }, + { + "epoch": 0.37614510809820445, + "grad_norm": 0.4277337291100283, + "learning_rate": 7.854478091624322e-06, + "loss": 0.5055, + "step": 2053 + }, + { + "epoch": 0.3763283253939172, + "grad_norm": 0.40260423661772543, + "learning_rate": 7.851851979961222e-06, + "loss": 0.4816, + "step": 2054 + }, + { + "epoch": 0.3765115426896299, + "grad_norm": 0.37218413297346103, + "learning_rate": 7.849224701730296e-06, + "loss": 0.4606, + "step": 2055 + }, + { + "epoch": 0.37669475998534263, + "grad_norm": 0.41491062083730507, + "learning_rate": 7.846596258006249e-06, + "loss": 0.4602, + "step": 2056 + }, + { + "epoch": 0.3768779772810553, + "grad_norm": 0.3960576653962031, + "learning_rate": 7.843966649864258e-06, + "loss": 0.4771, + "step": 2057 + }, + { + "epoch": 0.37706119457676807, + "grad_norm": 0.44222626036262375, + "learning_rate": 7.841335878379985e-06, + "loss": 0.4898, + "step": 2058 + }, + { + "epoch": 0.37724441187248076, + "grad_norm": 0.4342374882463006, + "learning_rate": 7.83870394462956e-06, + "loss": 0.4766, + "step": 2059 + }, + { + "epoch": 0.3774276291681935, + "grad_norm": 0.47449752687835595, + "learning_rate": 7.836070849689597e-06, + "loss": 0.5311, + "step": 2060 + }, + { + "epoch": 0.3776108464639062, + "grad_norm": 0.36364778037510276, + "learning_rate": 7.833436594637172e-06, + "loss": 0.5042, + "step": 2061 + }, + { + "epoch": 0.3777940637596189, + "grad_norm": 0.41142606251077274, + "learning_rate": 7.83080118054985e-06, + "loss": 0.5032, + "step": 2062 + }, + { + "epoch": 0.3779772810553316, + "grad_norm": 0.37574847969260744, + "learning_rate": 7.828164608505661e-06, + "loss": 0.509, + "step": 2063 + }, + { + "epoch": 0.3781604983510443, + "grad_norm": 0.37268467475303474, + "learning_rate": 7.825526879583115e-06, + "loss": 0.5024, + "step": 2064 + }, + { + "epoch": 0.37834371564675706, + "grad_norm": 0.37860686677862815, + "learning_rate": 7.822887994861183e-06, + "loss": 0.4625, + "step": 2065 + }, + { + "epoch": 0.37852693294246975, + "grad_norm": 0.38924369717801605, + "learning_rate": 7.820247955419326e-06, + "loss": 0.4725, + "step": 2066 + }, + { + "epoch": 0.3787101502381825, + "grad_norm": 0.3934868236020354, + "learning_rate": 7.817606762337465e-06, + "loss": 0.4573, + "step": 2067 + }, + { + "epoch": 0.3788933675338952, + "grad_norm": 0.40539217938905403, + "learning_rate": 7.814964416695995e-06, + "loss": 0.4652, + "step": 2068 + }, + { + "epoch": 0.37907658482960793, + "grad_norm": 0.34378698165362287, + "learning_rate": 7.812320919575786e-06, + "loss": 0.4559, + "step": 2069 + }, + { + "epoch": 0.3792598021253206, + "grad_norm": 0.44129404302636444, + "learning_rate": 7.809676272058177e-06, + "loss": 0.4913, + "step": 2070 + }, + { + "epoch": 0.37944301942103337, + "grad_norm": 0.43087188582145197, + "learning_rate": 7.807030475224974e-06, + "loss": 0.5004, + "step": 2071 + }, + { + "epoch": 0.37962623671674606, + "grad_norm": 0.47291970091792523, + "learning_rate": 7.804383530158463e-06, + "loss": 0.4864, + "step": 2072 + }, + { + "epoch": 0.3798094540124588, + "grad_norm": 0.4283612152851097, + "learning_rate": 7.80173543794139e-06, + "loss": 0.4998, + "step": 2073 + }, + { + "epoch": 0.3799926713081715, + "grad_norm": 0.37778750598499644, + "learning_rate": 7.799086199656973e-06, + "loss": 0.4804, + "step": 2074 + }, + { + "epoch": 0.3801758886038842, + "grad_norm": 0.3943558730390511, + "learning_rate": 7.796435816388899e-06, + "loss": 0.4801, + "step": 2075 + }, + { + "epoch": 0.3803591058995969, + "grad_norm": 0.3489087190074218, + "learning_rate": 7.793784289221327e-06, + "loss": 0.4739, + "step": 2076 + }, + { + "epoch": 0.3805423231953096, + "grad_norm": 0.46217929374869915, + "learning_rate": 7.791131619238876e-06, + "loss": 0.4969, + "step": 2077 + }, + { + "epoch": 0.38072554049102236, + "grad_norm": 0.41121895668172287, + "learning_rate": 7.788477807526642e-06, + "loss": 0.5032, + "step": 2078 + }, + { + "epoch": 0.38090875778673505, + "grad_norm": 0.47930768983705513, + "learning_rate": 7.78582285517018e-06, + "loss": 0.4639, + "step": 2079 + }, + { + "epoch": 0.3810919750824478, + "grad_norm": 0.47120906791011313, + "learning_rate": 7.783166763255516e-06, + "loss": 0.5224, + "step": 2080 + }, + { + "epoch": 0.3812751923781605, + "grad_norm": 0.4205960711178245, + "learning_rate": 7.780509532869141e-06, + "loss": 0.4979, + "step": 2081 + }, + { + "epoch": 0.38145840967387323, + "grad_norm": 0.3887963496226228, + "learning_rate": 7.777851165098012e-06, + "loss": 0.4834, + "step": 2082 + }, + { + "epoch": 0.3816416269695859, + "grad_norm": 0.4018128309846247, + "learning_rate": 7.77519166102955e-06, + "loss": 0.506, + "step": 2083 + }, + { + "epoch": 0.38182484426529867, + "grad_norm": 0.3794711195512327, + "learning_rate": 7.772531021751642e-06, + "loss": 0.4766, + "step": 2084 + }, + { + "epoch": 0.38200806156101136, + "grad_norm": 0.4002770093735582, + "learning_rate": 7.769869248352639e-06, + "loss": 0.4521, + "step": 2085 + }, + { + "epoch": 0.38219127885672405, + "grad_norm": 0.40267292791276327, + "learning_rate": 7.767206341921355e-06, + "loss": 0.5018, + "step": 2086 + }, + { + "epoch": 0.3823744961524368, + "grad_norm": 0.383846702167805, + "learning_rate": 7.764542303547072e-06, + "loss": 0.477, + "step": 2087 + }, + { + "epoch": 0.3825577134481495, + "grad_norm": 0.389093812397585, + "learning_rate": 7.761877134319527e-06, + "loss": 0.5249, + "step": 2088 + }, + { + "epoch": 0.38274093074386223, + "grad_norm": 0.3517488994665113, + "learning_rate": 7.759210835328928e-06, + "loss": 0.4485, + "step": 2089 + }, + { + "epoch": 0.3829241480395749, + "grad_norm": 0.4275804985293704, + "learning_rate": 7.756543407665938e-06, + "loss": 0.4793, + "step": 2090 + }, + { + "epoch": 0.38310736533528766, + "grad_norm": 0.3767475363274418, + "learning_rate": 7.753874852421685e-06, + "loss": 0.4789, + "step": 2091 + }, + { + "epoch": 0.38329058263100035, + "grad_norm": 0.38527193603900556, + "learning_rate": 7.751205170687762e-06, + "loss": 0.4594, + "step": 2092 + }, + { + "epoch": 0.3834737999267131, + "grad_norm": 0.33802023090956396, + "learning_rate": 7.748534363556218e-06, + "loss": 0.4583, + "step": 2093 + }, + { + "epoch": 0.3836570172224258, + "grad_norm": 0.3895945308329752, + "learning_rate": 7.745862432119558e-06, + "loss": 0.4686, + "step": 2094 + }, + { + "epoch": 0.38384023451813853, + "grad_norm": 0.39516215803159543, + "learning_rate": 7.743189377470757e-06, + "loss": 0.4742, + "step": 2095 + }, + { + "epoch": 0.3840234518138512, + "grad_norm": 0.39723290350325086, + "learning_rate": 7.740515200703243e-06, + "loss": 0.4882, + "step": 2096 + }, + { + "epoch": 0.38420666910956397, + "grad_norm": 0.37756507867324496, + "learning_rate": 7.737839902910907e-06, + "loss": 0.4697, + "step": 2097 + }, + { + "epoch": 0.38438988640527666, + "grad_norm": 0.3879992387325463, + "learning_rate": 7.735163485188094e-06, + "loss": 0.4744, + "step": 2098 + }, + { + "epoch": 0.38457310370098935, + "grad_norm": 0.45772551256710314, + "learning_rate": 7.73248594862961e-06, + "loss": 0.4686, + "step": 2099 + }, + { + "epoch": 0.3847563209967021, + "grad_norm": 0.3717857609113189, + "learning_rate": 7.729807294330718e-06, + "loss": 0.4891, + "step": 2100 + }, + { + "epoch": 0.3849395382924148, + "grad_norm": 0.5207031925479648, + "learning_rate": 7.727127523387137e-06, + "loss": 0.5163, + "step": 2101 + }, + { + "epoch": 0.38512275558812753, + "grad_norm": 0.38532398981684535, + "learning_rate": 7.724446636895046e-06, + "loss": 0.4912, + "step": 2102 + }, + { + "epoch": 0.3853059728838402, + "grad_norm": 0.3872603076306824, + "learning_rate": 7.721764635951077e-06, + "loss": 0.4683, + "step": 2103 + }, + { + "epoch": 0.38548919017955297, + "grad_norm": 0.3744508595734281, + "learning_rate": 7.71908152165232e-06, + "loss": 0.463, + "step": 2104 + }, + { + "epoch": 0.38567240747526566, + "grad_norm": 0.37074138714238275, + "learning_rate": 7.71639729509632e-06, + "loss": 0.455, + "step": 2105 + }, + { + "epoch": 0.3858556247709784, + "grad_norm": 6.78934868665371, + "learning_rate": 7.713711957381075e-06, + "loss": 0.4645, + "step": 2106 + }, + { + "epoch": 0.3860388420666911, + "grad_norm": 0.3486518549122761, + "learning_rate": 7.711025509605041e-06, + "loss": 0.4554, + "step": 2107 + }, + { + "epoch": 0.38622205936240384, + "grad_norm": 0.39662412739089803, + "learning_rate": 7.708337952867124e-06, + "loss": 0.4766, + "step": 2108 + }, + { + "epoch": 0.3864052766581165, + "grad_norm": 0.39678336687354887, + "learning_rate": 7.705649288266688e-06, + "loss": 0.4809, + "step": 2109 + }, + { + "epoch": 0.3865884939538292, + "grad_norm": 0.41426444886861935, + "learning_rate": 7.702959516903545e-06, + "loss": 0.4412, + "step": 2110 + }, + { + "epoch": 0.38677171124954196, + "grad_norm": 0.4254268214393484, + "learning_rate": 7.700268639877967e-06, + "loss": 0.5163, + "step": 2111 + }, + { + "epoch": 0.38695492854525465, + "grad_norm": 0.3850160297329316, + "learning_rate": 7.69757665829067e-06, + "loss": 0.4522, + "step": 2112 + }, + { + "epoch": 0.3871381458409674, + "grad_norm": 0.40665372424080226, + "learning_rate": 7.69488357324283e-06, + "loss": 0.4778, + "step": 2113 + }, + { + "epoch": 0.3873213631366801, + "grad_norm": 0.3734000274211507, + "learning_rate": 7.692189385836066e-06, + "loss": 0.504, + "step": 2114 + }, + { + "epoch": 0.38750458043239283, + "grad_norm": 0.38369481344162354, + "learning_rate": 7.689494097172457e-06, + "loss": 0.4819, + "step": 2115 + }, + { + "epoch": 0.3876877977281055, + "grad_norm": 0.39840029988152165, + "learning_rate": 7.68679770835452e-06, + "loss": 0.4762, + "step": 2116 + }, + { + "epoch": 0.38787101502381827, + "grad_norm": 0.4950377112241031, + "learning_rate": 7.684100220485239e-06, + "loss": 0.4542, + "step": 2117 + }, + { + "epoch": 0.38805423231953096, + "grad_norm": 0.41270189068249064, + "learning_rate": 7.681401634668032e-06, + "loss": 0.5096, + "step": 2118 + }, + { + "epoch": 0.3882374496152437, + "grad_norm": 0.4081024716177487, + "learning_rate": 7.678701952006773e-06, + "loss": 0.4965, + "step": 2119 + }, + { + "epoch": 0.3884206669109564, + "grad_norm": 0.36403093696511857, + "learning_rate": 7.676001173605784e-06, + "loss": 0.4849, + "step": 2120 + }, + { + "epoch": 0.38860388420666914, + "grad_norm": 0.4211817424530919, + "learning_rate": 7.673299300569838e-06, + "loss": 0.4765, + "step": 2121 + }, + { + "epoch": 0.3887871015023818, + "grad_norm": 0.35600532346699054, + "learning_rate": 7.670596334004152e-06, + "loss": 0.4456, + "step": 2122 + }, + { + "epoch": 0.3889703187980945, + "grad_norm": 0.38487262294629787, + "learning_rate": 7.66789227501439e-06, + "loss": 0.4717, + "step": 2123 + }, + { + "epoch": 0.38915353609380726, + "grad_norm": 0.42008503947816606, + "learning_rate": 7.665187124706664e-06, + "loss": 0.4904, + "step": 2124 + }, + { + "epoch": 0.38933675338951995, + "grad_norm": 0.3883906836870506, + "learning_rate": 7.662480884187533e-06, + "loss": 0.5033, + "step": 2125 + }, + { + "epoch": 0.3895199706852327, + "grad_norm": 0.34677206427001805, + "learning_rate": 7.659773554564e-06, + "loss": 0.4211, + "step": 2126 + }, + { + "epoch": 0.3897031879809454, + "grad_norm": 0.3886521107281691, + "learning_rate": 7.65706513694352e-06, + "loss": 0.4657, + "step": 2127 + }, + { + "epoch": 0.38988640527665813, + "grad_norm": 0.3956908204358445, + "learning_rate": 7.654355632433983e-06, + "loss": 0.4798, + "step": 2128 + }, + { + "epoch": 0.3900696225723708, + "grad_norm": 0.39174045933217894, + "learning_rate": 7.651645042143732e-06, + "loss": 0.4719, + "step": 2129 + }, + { + "epoch": 0.39025283986808357, + "grad_norm": 0.3893797813627037, + "learning_rate": 7.64893336718155e-06, + "loss": 0.4792, + "step": 2130 + }, + { + "epoch": 0.39043605716379626, + "grad_norm": 0.4020267142996408, + "learning_rate": 7.646220608656662e-06, + "loss": 0.4919, + "step": 2131 + }, + { + "epoch": 0.390619274459509, + "grad_norm": 0.4355175322607962, + "learning_rate": 7.643506767678744e-06, + "loss": 0.491, + "step": 2132 + }, + { + "epoch": 0.3908024917552217, + "grad_norm": 0.3594415535306099, + "learning_rate": 7.640791845357902e-06, + "loss": 0.4689, + "step": 2133 + }, + { + "epoch": 0.3909857090509344, + "grad_norm": 0.3629079494776069, + "learning_rate": 7.6380758428047e-06, + "loss": 0.483, + "step": 2134 + }, + { + "epoch": 0.39116892634664713, + "grad_norm": 0.40466076269530643, + "learning_rate": 7.635358761130132e-06, + "loss": 0.4684, + "step": 2135 + }, + { + "epoch": 0.3913521436423598, + "grad_norm": 0.3673513096544063, + "learning_rate": 7.632640601445636e-06, + "loss": 0.4457, + "step": 2136 + }, + { + "epoch": 0.39153536093807256, + "grad_norm": 0.4115880305227066, + "learning_rate": 7.629921364863093e-06, + "loss": 0.4416, + "step": 2137 + }, + { + "epoch": 0.39171857823378525, + "grad_norm": 0.40835494877531947, + "learning_rate": 7.627201052494826e-06, + "loss": 0.509, + "step": 2138 + }, + { + "epoch": 0.391901795529498, + "grad_norm": 0.390879632879768, + "learning_rate": 7.624479665453593e-06, + "loss": 0.5069, + "step": 2139 + }, + { + "epoch": 0.3920850128252107, + "grad_norm": 0.3926518851167653, + "learning_rate": 7.6217572048525935e-06, + "loss": 0.4801, + "step": 2140 + }, + { + "epoch": 0.39226823012092343, + "grad_norm": 0.3605648124653004, + "learning_rate": 7.619033671805472e-06, + "loss": 0.4694, + "step": 2141 + }, + { + "epoch": 0.3924514474166361, + "grad_norm": 0.40842922029409745, + "learning_rate": 7.616309067426299e-06, + "loss": 0.4857, + "step": 2142 + }, + { + "epoch": 0.39263466471234887, + "grad_norm": 0.369533925890746, + "learning_rate": 7.6135833928296e-06, + "loss": 0.4918, + "step": 2143 + }, + { + "epoch": 0.39281788200806156, + "grad_norm": 0.3825237185857143, + "learning_rate": 7.61085664913032e-06, + "loss": 0.4534, + "step": 2144 + }, + { + "epoch": 0.3930010993037743, + "grad_norm": 0.38017004429876705, + "learning_rate": 7.608128837443859e-06, + "loss": 0.4842, + "step": 2145 + }, + { + "epoch": 0.393184316599487, + "grad_norm": 0.3449268373210601, + "learning_rate": 7.605399958886038e-06, + "loss": 0.4686, + "step": 2146 + }, + { + "epoch": 0.3933675338951997, + "grad_norm": 0.40247350440016877, + "learning_rate": 7.602670014573128e-06, + "loss": 0.51, + "step": 2147 + }, + { + "epoch": 0.39355075119091243, + "grad_norm": 0.40267803112415795, + "learning_rate": 7.599939005621824e-06, + "loss": 0.483, + "step": 2148 + }, + { + "epoch": 0.3937339684866251, + "grad_norm": 0.37001298207862, + "learning_rate": 7.597206933149268e-06, + "loss": 0.4733, + "step": 2149 + }, + { + "epoch": 0.39391718578233786, + "grad_norm": 0.38585830693759504, + "learning_rate": 7.594473798273027e-06, + "loss": 0.5061, + "step": 2150 + }, + { + "epoch": 0.39410040307805055, + "grad_norm": 0.4018233458442901, + "learning_rate": 7.591739602111108e-06, + "loss": 0.4896, + "step": 2151 + }, + { + "epoch": 0.3942836203737633, + "grad_norm": 0.3792984268443907, + "learning_rate": 7.589004345781951e-06, + "loss": 0.5053, + "step": 2152 + }, + { + "epoch": 0.394466837669476, + "grad_norm": 0.36337707550131254, + "learning_rate": 7.5862680304044286e-06, + "loss": 0.4865, + "step": 2153 + }, + { + "epoch": 0.39465005496518873, + "grad_norm": 0.41298620701906524, + "learning_rate": 7.583530657097851e-06, + "loss": 0.4727, + "step": 2154 + }, + { + "epoch": 0.3948332722609014, + "grad_norm": 0.37563122770443724, + "learning_rate": 7.580792226981954e-06, + "loss": 0.4821, + "step": 2155 + }, + { + "epoch": 0.39501648955661417, + "grad_norm": 0.37054526425400464, + "learning_rate": 7.578052741176914e-06, + "loss": 0.4653, + "step": 2156 + }, + { + "epoch": 0.39519970685232686, + "grad_norm": 0.4039521209284209, + "learning_rate": 7.575312200803329e-06, + "loss": 0.4726, + "step": 2157 + }, + { + "epoch": 0.39538292414803955, + "grad_norm": 0.4394954934863309, + "learning_rate": 7.572570606982239e-06, + "loss": 0.4933, + "step": 2158 + }, + { + "epoch": 0.3955661414437523, + "grad_norm": 0.38465023174159757, + "learning_rate": 7.5698279608351064e-06, + "loss": 0.4888, + "step": 2159 + }, + { + "epoch": 0.395749358739465, + "grad_norm": 0.3560885282118001, + "learning_rate": 7.5670842634838304e-06, + "loss": 0.5009, + "step": 2160 + }, + { + "epoch": 0.39593257603517773, + "grad_norm": 0.3589853290162964, + "learning_rate": 7.564339516050736e-06, + "loss": 0.5129, + "step": 2161 + }, + { + "epoch": 0.3961157933308904, + "grad_norm": 0.35558865750409446, + "learning_rate": 7.5615937196585795e-06, + "loss": 0.4494, + "step": 2162 + }, + { + "epoch": 0.39629901062660317, + "grad_norm": 0.37425144469863364, + "learning_rate": 7.558846875430548e-06, + "loss": 0.4783, + "step": 2163 + }, + { + "epoch": 0.39648222792231586, + "grad_norm": 0.5576051549727741, + "learning_rate": 7.556098984490252e-06, + "loss": 0.4901, + "step": 2164 + }, + { + "epoch": 0.3966654452180286, + "grad_norm": 0.3923037940354051, + "learning_rate": 7.553350047961736e-06, + "loss": 0.4707, + "step": 2165 + }, + { + "epoch": 0.3968486625137413, + "grad_norm": 0.3967667690205941, + "learning_rate": 7.550600066969467e-06, + "loss": 0.4612, + "step": 2166 + }, + { + "epoch": 0.39703187980945404, + "grad_norm": 0.4126790308950325, + "learning_rate": 7.547849042638346e-06, + "loss": 0.4755, + "step": 2167 + }, + { + "epoch": 0.3972150971051667, + "grad_norm": 0.32410941657150816, + "learning_rate": 7.545096976093691e-06, + "loss": 0.4431, + "step": 2168 + }, + { + "epoch": 0.3973983144008794, + "grad_norm": 0.35328756787903626, + "learning_rate": 7.542343868461257e-06, + "loss": 0.461, + "step": 2169 + }, + { + "epoch": 0.39758153169659216, + "grad_norm": 0.40535237947349906, + "learning_rate": 7.5395897208672175e-06, + "loss": 0.4395, + "step": 2170 + }, + { + "epoch": 0.39776474899230485, + "grad_norm": 0.3895685994822475, + "learning_rate": 7.536834534438174e-06, + "loss": 0.4647, + "step": 2171 + }, + { + "epoch": 0.3979479662880176, + "grad_norm": 0.4200201633472651, + "learning_rate": 7.5340783103011514e-06, + "loss": 0.5078, + "step": 2172 + }, + { + "epoch": 0.3981311835837303, + "grad_norm": 0.4220605089377968, + "learning_rate": 7.531321049583604e-06, + "loss": 0.5153, + "step": 2173 + }, + { + "epoch": 0.39831440087944303, + "grad_norm": 0.46540844259324055, + "learning_rate": 7.528562753413403e-06, + "loss": 0.4587, + "step": 2174 + }, + { + "epoch": 0.3984976181751557, + "grad_norm": 0.38752820563099566, + "learning_rate": 7.525803422918849e-06, + "loss": 0.5002, + "step": 2175 + }, + { + "epoch": 0.39868083547086847, + "grad_norm": 0.4048295498281393, + "learning_rate": 7.523043059228663e-06, + "loss": 0.4754, + "step": 2176 + }, + { + "epoch": 0.39886405276658116, + "grad_norm": 0.39440387714054603, + "learning_rate": 7.520281663471987e-06, + "loss": 0.4617, + "step": 2177 + }, + { + "epoch": 0.3990472700622939, + "grad_norm": 0.3735891747445075, + "learning_rate": 7.517519236778388e-06, + "loss": 0.4791, + "step": 2178 + }, + { + "epoch": 0.3992304873580066, + "grad_norm": 0.37520297038166667, + "learning_rate": 7.514755780277854e-06, + "loss": 0.4665, + "step": 2179 + }, + { + "epoch": 0.39941370465371934, + "grad_norm": 0.39124000876509607, + "learning_rate": 7.511991295100793e-06, + "loss": 0.487, + "step": 2180 + }, + { + "epoch": 0.399596921949432, + "grad_norm": 0.37390493837079697, + "learning_rate": 7.509225782378039e-06, + "loss": 0.4857, + "step": 2181 + }, + { + "epoch": 0.3997801392451447, + "grad_norm": 0.4449963874189541, + "learning_rate": 7.506459243240838e-06, + "loss": 0.4337, + "step": 2182 + }, + { + "epoch": 0.39996335654085746, + "grad_norm": 0.3574874059229643, + "learning_rate": 7.503691678820862e-06, + "loss": 0.486, + "step": 2183 + }, + { + "epoch": 0.40014657383657015, + "grad_norm": 0.4093212385745805, + "learning_rate": 7.500923090250199e-06, + "loss": 0.4724, + "step": 2184 + }, + { + "epoch": 0.4003297911322829, + "grad_norm": 0.37204853752647493, + "learning_rate": 7.49815347866136e-06, + "loss": 0.4763, + "step": 2185 + }, + { + "epoch": 0.4005130084279956, + "grad_norm": 0.3977741195665693, + "learning_rate": 7.495382845187269e-06, + "loss": 0.4948, + "step": 2186 + }, + { + "epoch": 0.40069622572370833, + "grad_norm": 0.4355144605708253, + "learning_rate": 7.492611190961272e-06, + "loss": 0.5168, + "step": 2187 + }, + { + "epoch": 0.400879443019421, + "grad_norm": 0.38786474481273503, + "learning_rate": 7.4898385171171325e-06, + "loss": 0.4885, + "step": 2188 + }, + { + "epoch": 0.40106266031513377, + "grad_norm": 0.39820590534589656, + "learning_rate": 7.487064824789029e-06, + "loss": 0.4931, + "step": 2189 + }, + { + "epoch": 0.40124587761084646, + "grad_norm": 0.4369066265711082, + "learning_rate": 7.484290115111556e-06, + "loss": 0.5038, + "step": 2190 + }, + { + "epoch": 0.4014290949065592, + "grad_norm": 0.3707270116501878, + "learning_rate": 7.48151438921973e-06, + "loss": 0.5141, + "step": 2191 + }, + { + "epoch": 0.4016123122022719, + "grad_norm": 0.412393618517063, + "learning_rate": 7.478737648248976e-06, + "loss": 0.4083, + "step": 2192 + }, + { + "epoch": 0.4017955294979846, + "grad_norm": 0.3454576812197135, + "learning_rate": 7.475959893335137e-06, + "loss": 0.4898, + "step": 2193 + }, + { + "epoch": 0.40197874679369733, + "grad_norm": 0.36298388895956907, + "learning_rate": 7.4731811256144715e-06, + "loss": 0.477, + "step": 2194 + }, + { + "epoch": 0.40216196408941, + "grad_norm": 0.42911123820279806, + "learning_rate": 7.470401346223653e-06, + "loss": 0.4856, + "step": 2195 + }, + { + "epoch": 0.40234518138512276, + "grad_norm": 0.337394372757646, + "learning_rate": 7.467620556299765e-06, + "loss": 0.4874, + "step": 2196 + }, + { + "epoch": 0.40252839868083545, + "grad_norm": 0.403125636190336, + "learning_rate": 7.464838756980311e-06, + "loss": 0.4818, + "step": 2197 + }, + { + "epoch": 0.4027116159765482, + "grad_norm": 0.42854909357038723, + "learning_rate": 7.462055949403197e-06, + "loss": 0.4707, + "step": 2198 + }, + { + "epoch": 0.4028948332722609, + "grad_norm": 0.3976349637207099, + "learning_rate": 7.459272134706756e-06, + "loss": 0.4872, + "step": 2199 + }, + { + "epoch": 0.40307805056797363, + "grad_norm": 0.39149725720719286, + "learning_rate": 7.4564873140297165e-06, + "loss": 0.4547, + "step": 2200 + }, + { + "epoch": 0.4032612678636863, + "grad_norm": 0.40865931439064734, + "learning_rate": 7.4537014885112324e-06, + "loss": 0.4935, + "step": 2201 + }, + { + "epoch": 0.40344448515939907, + "grad_norm": 0.4015472011822115, + "learning_rate": 7.450914659290861e-06, + "loss": 0.4799, + "step": 2202 + }, + { + "epoch": 0.40362770245511176, + "grad_norm": 0.4091493415252484, + "learning_rate": 7.448126827508573e-06, + "loss": 0.5062, + "step": 2203 + }, + { + "epoch": 0.4038109197508245, + "grad_norm": 0.9070067666667128, + "learning_rate": 7.4453379943047464e-06, + "loss": 0.472, + "step": 2204 + }, + { + "epoch": 0.4039941370465372, + "grad_norm": 0.4073603354252967, + "learning_rate": 7.442548160820173e-06, + "loss": 0.4688, + "step": 2205 + }, + { + "epoch": 0.4041773543422499, + "grad_norm": 0.40144644953179104, + "learning_rate": 7.43975732819605e-06, + "loss": 0.4832, + "step": 2206 + }, + { + "epoch": 0.40436057163796263, + "grad_norm": 0.4170873209324686, + "learning_rate": 7.436965497573986e-06, + "loss": 0.4828, + "step": 2207 + }, + { + "epoch": 0.4045437889336753, + "grad_norm": 0.40847429220932524, + "learning_rate": 7.434172670095996e-06, + "loss": 0.5074, + "step": 2208 + }, + { + "epoch": 0.40472700622938806, + "grad_norm": 0.40814229613422337, + "learning_rate": 7.431378846904504e-06, + "loss": 0.4771, + "step": 2209 + }, + { + "epoch": 0.40491022352510075, + "grad_norm": 0.3679576716810249, + "learning_rate": 7.42858402914234e-06, + "loss": 0.4333, + "step": 2210 + }, + { + "epoch": 0.4050934408208135, + "grad_norm": 0.37846692009685645, + "learning_rate": 7.425788217952744e-06, + "loss": 0.4749, + "step": 2211 + }, + { + "epoch": 0.4052766581165262, + "grad_norm": 0.3837302497036784, + "learning_rate": 7.422991414479357e-06, + "loss": 0.5077, + "step": 2212 + }, + { + "epoch": 0.40545987541223893, + "grad_norm": 0.3780801221538502, + "learning_rate": 7.42019361986623e-06, + "loss": 0.4957, + "step": 2213 + }, + { + "epoch": 0.4056430927079516, + "grad_norm": 0.40118620548139255, + "learning_rate": 7.4173948352578185e-06, + "loss": 0.4906, + "step": 2214 + }, + { + "epoch": 0.40582631000366437, + "grad_norm": 0.39428856095280773, + "learning_rate": 7.414595061798982e-06, + "loss": 0.4848, + "step": 2215 + }, + { + "epoch": 0.40600952729937706, + "grad_norm": 0.38252867751426445, + "learning_rate": 7.411794300634987e-06, + "loss": 0.4712, + "step": 2216 + }, + { + "epoch": 0.40619274459508975, + "grad_norm": 0.36880701937182847, + "learning_rate": 7.408992552911501e-06, + "loss": 0.4655, + "step": 2217 + }, + { + "epoch": 0.4063759618908025, + "grad_norm": 0.3940805182917263, + "learning_rate": 7.406189819774597e-06, + "loss": 0.4833, + "step": 2218 + }, + { + "epoch": 0.4065591791865152, + "grad_norm": 0.4354726470234921, + "learning_rate": 7.403386102370751e-06, + "loss": 0.4963, + "step": 2219 + }, + { + "epoch": 0.40674239648222793, + "grad_norm": 0.35800143118316946, + "learning_rate": 7.400581401846838e-06, + "loss": 0.4886, + "step": 2220 + }, + { + "epoch": 0.4069256137779406, + "grad_norm": 0.38756053513470157, + "learning_rate": 7.397775719350142e-06, + "loss": 0.4883, + "step": 2221 + }, + { + "epoch": 0.40710883107365337, + "grad_norm": 0.3978384125978718, + "learning_rate": 7.394969056028342e-06, + "loss": 0.5035, + "step": 2222 + }, + { + "epoch": 0.40729204836936606, + "grad_norm": 0.4098220600424564, + "learning_rate": 7.392161413029524e-06, + "loss": 0.5013, + "step": 2223 + }, + { + "epoch": 0.4074752656650788, + "grad_norm": 0.3913210982172656, + "learning_rate": 7.389352791502169e-06, + "loss": 0.5028, + "step": 2224 + }, + { + "epoch": 0.4076584829607915, + "grad_norm": 0.4018866869596013, + "learning_rate": 7.386543192595164e-06, + "loss": 0.4944, + "step": 2225 + }, + { + "epoch": 0.40784170025650424, + "grad_norm": 0.44556717505388216, + "learning_rate": 7.3837326174577885e-06, + "loss": 0.5003, + "step": 2226 + }, + { + "epoch": 0.4080249175522169, + "grad_norm": 0.39726874380074945, + "learning_rate": 7.380921067239732e-06, + "loss": 0.4853, + "step": 2227 + }, + { + "epoch": 0.40820813484792967, + "grad_norm": 0.3604307128401756, + "learning_rate": 7.378108543091071e-06, + "loss": 0.5015, + "step": 2228 + }, + { + "epoch": 0.40839135214364236, + "grad_norm": 0.3745391205200884, + "learning_rate": 7.375295046162289e-06, + "loss": 0.503, + "step": 2229 + }, + { + "epoch": 0.40857456943935505, + "grad_norm": 0.5671616342121636, + "learning_rate": 7.372480577604265e-06, + "loss": 0.4645, + "step": 2230 + }, + { + "epoch": 0.4087577867350678, + "grad_norm": 0.4233519035995331, + "learning_rate": 7.369665138568273e-06, + "loss": 0.4917, + "step": 2231 + }, + { + "epoch": 0.4089410040307805, + "grad_norm": 0.4160654952014177, + "learning_rate": 7.3668487302059864e-06, + "loss": 0.4968, + "step": 2232 + }, + { + "epoch": 0.40912422132649323, + "grad_norm": 0.39647388057983846, + "learning_rate": 7.364031353669474e-06, + "loss": 0.4905, + "step": 2233 + }, + { + "epoch": 0.4093074386222059, + "grad_norm": 0.35780078289468054, + "learning_rate": 7.361213010111203e-06, + "loss": 0.4691, + "step": 2234 + }, + { + "epoch": 0.40949065591791867, + "grad_norm": 0.3810189187242057, + "learning_rate": 7.3583937006840335e-06, + "loss": 0.478, + "step": 2235 + }, + { + "epoch": 0.40967387321363136, + "grad_norm": 0.3638125779573905, + "learning_rate": 7.355573426541221e-06, + "loss": 0.4403, + "step": 2236 + }, + { + "epoch": 0.4098570905093441, + "grad_norm": 0.3879941333719194, + "learning_rate": 7.352752188836417e-06, + "loss": 0.48, + "step": 2237 + }, + { + "epoch": 0.4100403078050568, + "grad_norm": 0.35223422724381886, + "learning_rate": 7.349929988723667e-06, + "loss": 0.4382, + "step": 2238 + }, + { + "epoch": 0.41022352510076954, + "grad_norm": 0.43035086728389904, + "learning_rate": 7.347106827357408e-06, + "loss": 0.5003, + "step": 2239 + }, + { + "epoch": 0.4104067423964822, + "grad_norm": 0.3697650980170332, + "learning_rate": 7.344282705892472e-06, + "loss": 0.4812, + "step": 2240 + }, + { + "epoch": 0.4105899596921949, + "grad_norm": 0.3688535347430096, + "learning_rate": 7.341457625484084e-06, + "loss": 0.4755, + "step": 2241 + }, + { + "epoch": 0.41077317698790766, + "grad_norm": 0.37729895938975644, + "learning_rate": 7.3386315872878614e-06, + "loss": 0.4508, + "step": 2242 + }, + { + "epoch": 0.41095639428362035, + "grad_norm": 0.36893825292436533, + "learning_rate": 7.335804592459811e-06, + "loss": 0.4581, + "step": 2243 + }, + { + "epoch": 0.4111396115793331, + "grad_norm": 0.36619333950680655, + "learning_rate": 7.332976642156334e-06, + "loss": 0.493, + "step": 2244 + }, + { + "epoch": 0.4113228288750458, + "grad_norm": 0.3837821710857386, + "learning_rate": 7.330147737534221e-06, + "loss": 0.4666, + "step": 2245 + }, + { + "epoch": 0.41150604617075853, + "grad_norm": 1.2145896374263545, + "learning_rate": 7.327317879750653e-06, + "loss": 0.4731, + "step": 2246 + }, + { + "epoch": 0.4116892634664712, + "grad_norm": 0.38417244763549135, + "learning_rate": 7.324487069963201e-06, + "loss": 0.4802, + "step": 2247 + }, + { + "epoch": 0.41187248076218397, + "grad_norm": 0.36398444912010497, + "learning_rate": 7.321655309329823e-06, + "loss": 0.4899, + "step": 2248 + }, + { + "epoch": 0.41205569805789666, + "grad_norm": 0.3577648813102727, + "learning_rate": 7.318822599008872e-06, + "loss": 0.503, + "step": 2249 + }, + { + "epoch": 0.4122389153536094, + "grad_norm": 0.38814144311402826, + "learning_rate": 7.315988940159085e-06, + "loss": 0.4619, + "step": 2250 + }, + { + "epoch": 0.4124221326493221, + "grad_norm": 0.37076356579116254, + "learning_rate": 7.313154333939587e-06, + "loss": 0.494, + "step": 2251 + }, + { + "epoch": 0.41260534994503484, + "grad_norm": 0.35795968217997975, + "learning_rate": 7.31031878150989e-06, + "loss": 0.4542, + "step": 2252 + }, + { + "epoch": 0.41278856724074753, + "grad_norm": 0.34611454726130503, + "learning_rate": 7.307482284029897e-06, + "loss": 0.4896, + "step": 2253 + }, + { + "epoch": 0.4129717845364602, + "grad_norm": 0.4006663715053596, + "learning_rate": 7.304644842659894e-06, + "loss": 0.4808, + "step": 2254 + }, + { + "epoch": 0.41315500183217296, + "grad_norm": 0.6073873668548115, + "learning_rate": 7.301806458560553e-06, + "loss": 0.4588, + "step": 2255 + }, + { + "epoch": 0.41333821912788565, + "grad_norm": 0.38804921937120407, + "learning_rate": 7.298967132892934e-06, + "loss": 0.483, + "step": 2256 + }, + { + "epoch": 0.4135214364235984, + "grad_norm": 0.414915648982362, + "learning_rate": 7.2961268668184805e-06, + "loss": 0.4633, + "step": 2257 + }, + { + "epoch": 0.4137046537193111, + "grad_norm": 0.4036416573411531, + "learning_rate": 7.293285661499021e-06, + "loss": 0.489, + "step": 2258 + }, + { + "epoch": 0.41388787101502383, + "grad_norm": 0.3856693171158019, + "learning_rate": 7.2904435180967695e-06, + "loss": 0.4943, + "step": 2259 + }, + { + "epoch": 0.4140710883107365, + "grad_norm": 0.37781814896390287, + "learning_rate": 7.2876004377743195e-06, + "loss": 0.4711, + "step": 2260 + }, + { + "epoch": 0.41425430560644927, + "grad_norm": 0.42598022297248983, + "learning_rate": 7.284756421694653e-06, + "loss": 0.4635, + "step": 2261 + }, + { + "epoch": 0.41443752290216196, + "grad_norm": 0.3843619562770163, + "learning_rate": 7.281911471021132e-06, + "loss": 0.4613, + "step": 2262 + }, + { + "epoch": 0.4146207401978747, + "grad_norm": 0.3719571330491704, + "learning_rate": 7.279065586917501e-06, + "loss": 0.4831, + "step": 2263 + }, + { + "epoch": 0.4148039574935874, + "grad_norm": 0.36099595629868814, + "learning_rate": 7.276218770547887e-06, + "loss": 0.4745, + "step": 2264 + }, + { + "epoch": 0.4149871747893001, + "grad_norm": 0.3710349117267771, + "learning_rate": 7.273371023076798e-06, + "loss": 0.4522, + "step": 2265 + }, + { + "epoch": 0.41517039208501283, + "grad_norm": 0.3955654732668277, + "learning_rate": 7.270522345669123e-06, + "loss": 0.4777, + "step": 2266 + }, + { + "epoch": 0.4153536093807255, + "grad_norm": 0.3497920558334845, + "learning_rate": 7.26767273949013e-06, + "loss": 0.4808, + "step": 2267 + }, + { + "epoch": 0.41553682667643826, + "grad_norm": 0.3768858557614057, + "learning_rate": 7.26482220570547e-06, + "loss": 0.4813, + "step": 2268 + }, + { + "epoch": 0.41572004397215095, + "grad_norm": 0.3846462900161903, + "learning_rate": 7.261970745481171e-06, + "loss": 0.5131, + "step": 2269 + }, + { + "epoch": 0.4159032612678637, + "grad_norm": 0.33914797999173163, + "learning_rate": 7.25911835998364e-06, + "loss": 0.4859, + "step": 2270 + }, + { + "epoch": 0.4160864785635764, + "grad_norm": 0.40245926200014415, + "learning_rate": 7.256265050379665e-06, + "loss": 0.441, + "step": 2271 + }, + { + "epoch": 0.41626969585928913, + "grad_norm": 0.4076739044981689, + "learning_rate": 7.253410817836405e-06, + "loss": 0.508, + "step": 2272 + }, + { + "epoch": 0.4164529131550018, + "grad_norm": 0.3549816696082439, + "learning_rate": 7.250555663521408e-06, + "loss": 0.4419, + "step": 2273 + }, + { + "epoch": 0.41663613045071457, + "grad_norm": 0.42523114420007596, + "learning_rate": 7.247699588602588e-06, + "loss": 0.4637, + "step": 2274 + }, + { + "epoch": 0.41681934774642726, + "grad_norm": 0.43818992996209233, + "learning_rate": 7.244842594248244e-06, + "loss": 0.4551, + "step": 2275 + }, + { + "epoch": 0.41700256504214, + "grad_norm": 0.39560111046216095, + "learning_rate": 7.2419846816270425e-06, + "loss": 0.4361, + "step": 2276 + }, + { + "epoch": 0.4171857823378527, + "grad_norm": 0.38787100181152556, + "learning_rate": 7.239125851908035e-06, + "loss": 0.4737, + "step": 2277 + }, + { + "epoch": 0.4173689996335654, + "grad_norm": 0.3557365091431396, + "learning_rate": 7.236266106260641e-06, + "loss": 0.4992, + "step": 2278 + }, + { + "epoch": 0.41755221692927813, + "grad_norm": 0.4203070336395134, + "learning_rate": 7.23340544585466e-06, + "loss": 0.4909, + "step": 2279 + }, + { + "epoch": 0.4177354342249908, + "grad_norm": 0.4108396938981641, + "learning_rate": 7.2305438718602594e-06, + "loss": 0.4695, + "step": 2280 + }, + { + "epoch": 0.41791865152070357, + "grad_norm": 0.4620346664763805, + "learning_rate": 7.227681385447988e-06, + "loss": 0.4918, + "step": 2281 + }, + { + "epoch": 0.41810186881641626, + "grad_norm": 0.41033815242464133, + "learning_rate": 7.22481798778876e-06, + "loss": 0.4571, + "step": 2282 + }, + { + "epoch": 0.418285086112129, + "grad_norm": 0.40104297809788464, + "learning_rate": 7.221953680053867e-06, + "loss": 0.4381, + "step": 2283 + }, + { + "epoch": 0.4184683034078417, + "grad_norm": 0.40968469354164905, + "learning_rate": 7.219088463414972e-06, + "loss": 0.4844, + "step": 2284 + }, + { + "epoch": 0.41865152070355444, + "grad_norm": 0.39955381671097556, + "learning_rate": 7.216222339044111e-06, + "loss": 0.4835, + "step": 2285 + }, + { + "epoch": 0.4188347379992671, + "grad_norm": 0.5918823635482916, + "learning_rate": 7.213355308113688e-06, + "loss": 0.481, + "step": 2286 + }, + { + "epoch": 0.41901795529497987, + "grad_norm": 0.4438898450190298, + "learning_rate": 7.210487371796482e-06, + "loss": 0.5296, + "step": 2287 + }, + { + "epoch": 0.41920117259069256, + "grad_norm": 0.42513041758375775, + "learning_rate": 7.207618531265639e-06, + "loss": 0.4787, + "step": 2288 + }, + { + "epoch": 0.41938438988640525, + "grad_norm": 0.4344461669669104, + "learning_rate": 7.204748787694676e-06, + "loss": 0.5039, + "step": 2289 + }, + { + "epoch": 0.419567607182118, + "grad_norm": 0.362299437631823, + "learning_rate": 7.201878142257481e-06, + "loss": 0.4858, + "step": 2290 + }, + { + "epoch": 0.4197508244778307, + "grad_norm": 0.38326936079136137, + "learning_rate": 7.1990065961283075e-06, + "loss": 0.4999, + "step": 2291 + }, + { + "epoch": 0.41993404177354343, + "grad_norm": 0.37084577713617844, + "learning_rate": 7.19613415048178e-06, + "loss": 0.4744, + "step": 2292 + }, + { + "epoch": 0.4201172590692561, + "grad_norm": 0.3857222120456293, + "learning_rate": 7.19326080649289e-06, + "loss": 0.4567, + "step": 2293 + }, + { + "epoch": 0.42030047636496887, + "grad_norm": 0.41124445894901135, + "learning_rate": 7.190386565336997e-06, + "loss": 0.445, + "step": 2294 + }, + { + "epoch": 0.42048369366068156, + "grad_norm": 0.3888927715893213, + "learning_rate": 7.187511428189828e-06, + "loss": 0.4659, + "step": 2295 + }, + { + "epoch": 0.4206669109563943, + "grad_norm": 0.40097744869064067, + "learning_rate": 7.184635396227475e-06, + "loss": 0.4925, + "step": 2296 + }, + { + "epoch": 0.420850128252107, + "grad_norm": 0.34895768769998925, + "learning_rate": 7.181758470626398e-06, + "loss": 0.4621, + "step": 2297 + }, + { + "epoch": 0.42103334554781974, + "grad_norm": 0.42929749713876525, + "learning_rate": 7.178880652563418e-06, + "loss": 0.4495, + "step": 2298 + }, + { + "epoch": 0.4212165628435324, + "grad_norm": 0.4370392489135322, + "learning_rate": 7.1760019432157295e-06, + "loss": 0.5151, + "step": 2299 + }, + { + "epoch": 0.4213997801392451, + "grad_norm": 0.4050632338683901, + "learning_rate": 7.173122343760881e-06, + "loss": 0.4686, + "step": 2300 + }, + { + "epoch": 0.42158299743495786, + "grad_norm": 0.39620071670775564, + "learning_rate": 7.170241855376795e-06, + "loss": 0.4549, + "step": 2301 + }, + { + "epoch": 0.42176621473067055, + "grad_norm": 0.376717314708779, + "learning_rate": 7.16736047924175e-06, + "loss": 0.4609, + "step": 2302 + }, + { + "epoch": 0.4219494320263833, + "grad_norm": 0.39539845287388553, + "learning_rate": 7.164478216534394e-06, + "loss": 0.4487, + "step": 2303 + }, + { + "epoch": 0.422132649322096, + "grad_norm": 0.3977551665287066, + "learning_rate": 7.161595068433732e-06, + "loss": 0.4967, + "step": 2304 + }, + { + "epoch": 0.42231586661780873, + "grad_norm": 0.4238609498668758, + "learning_rate": 7.158711036119136e-06, + "loss": 0.4842, + "step": 2305 + }, + { + "epoch": 0.4224990839135214, + "grad_norm": 0.3816935398139279, + "learning_rate": 7.155826120770334e-06, + "loss": 0.4945, + "step": 2306 + }, + { + "epoch": 0.42268230120923417, + "grad_norm": 0.40787570736995066, + "learning_rate": 7.1529403235674236e-06, + "loss": 0.4384, + "step": 2307 + }, + { + "epoch": 0.42286551850494686, + "grad_norm": 0.39353745511283, + "learning_rate": 7.150053645690852e-06, + "loss": 0.4675, + "step": 2308 + }, + { + "epoch": 0.4230487358006596, + "grad_norm": 0.37196245750947515, + "learning_rate": 7.147166088321438e-06, + "loss": 0.4757, + "step": 2309 + }, + { + "epoch": 0.4232319530963723, + "grad_norm": 0.35562310599814084, + "learning_rate": 7.144277652640351e-06, + "loss": 0.4746, + "step": 2310 + }, + { + "epoch": 0.42341517039208504, + "grad_norm": 0.4033914255134625, + "learning_rate": 7.141388339829128e-06, + "loss": 0.4582, + "step": 2311 + }, + { + "epoch": 0.42359838768779773, + "grad_norm": 0.39959402533230987, + "learning_rate": 7.138498151069656e-06, + "loss": 0.4529, + "step": 2312 + }, + { + "epoch": 0.4237816049835104, + "grad_norm": 0.4330351824355995, + "learning_rate": 7.135607087544187e-06, + "loss": 0.511, + "step": 2313 + }, + { + "epoch": 0.42396482227922316, + "grad_norm": 0.37271850156382125, + "learning_rate": 7.132715150435328e-06, + "loss": 0.4896, + "step": 2314 + }, + { + "epoch": 0.42414803957493585, + "grad_norm": 0.3317719970934139, + "learning_rate": 7.129822340926045e-06, + "loss": 0.4821, + "step": 2315 + }, + { + "epoch": 0.4243312568706486, + "grad_norm": 0.33830240266972705, + "learning_rate": 7.126928660199658e-06, + "loss": 0.4634, + "step": 2316 + }, + { + "epoch": 0.4245144741663613, + "grad_norm": 0.399219687554885, + "learning_rate": 7.1240341094398454e-06, + "loss": 0.4598, + "step": 2317 + }, + { + "epoch": 0.42469769146207403, + "grad_norm": 0.5150508709340539, + "learning_rate": 7.121138689830643e-06, + "loss": 0.4851, + "step": 2318 + }, + { + "epoch": 0.4248809087577867, + "grad_norm": 0.3970274185491194, + "learning_rate": 7.118242402556439e-06, + "loss": 0.447, + "step": 2319 + }, + { + "epoch": 0.42506412605349947, + "grad_norm": 0.3946847066468069, + "learning_rate": 7.115345248801978e-06, + "loss": 0.499, + "step": 2320 + }, + { + "epoch": 0.42524734334921216, + "grad_norm": 0.4466655731547931, + "learning_rate": 7.112447229752358e-06, + "loss": 0.514, + "step": 2321 + }, + { + "epoch": 0.4254305606449249, + "grad_norm": 0.43460768309632875, + "learning_rate": 7.1095483465930324e-06, + "loss": 0.4769, + "step": 2322 + }, + { + "epoch": 0.4256137779406376, + "grad_norm": 0.3941251324343338, + "learning_rate": 7.106648600509809e-06, + "loss": 0.463, + "step": 2323 + }, + { + "epoch": 0.4257969952363503, + "grad_norm": 0.35038780729862357, + "learning_rate": 7.1037479926888435e-06, + "loss": 0.4433, + "step": 2324 + }, + { + "epoch": 0.42598021253206303, + "grad_norm": 0.3908500139385616, + "learning_rate": 7.100846524316651e-06, + "loss": 0.4712, + "step": 2325 + }, + { + "epoch": 0.4261634298277757, + "grad_norm": 0.3846701226032806, + "learning_rate": 7.0979441965800914e-06, + "loss": 0.4534, + "step": 2326 + }, + { + "epoch": 0.42634664712348846, + "grad_norm": 0.34990602751139527, + "learning_rate": 7.095041010666385e-06, + "loss": 0.4621, + "step": 2327 + }, + { + "epoch": 0.42652986441920115, + "grad_norm": 0.37881854495740264, + "learning_rate": 7.092136967763092e-06, + "loss": 0.4724, + "step": 2328 + }, + { + "epoch": 0.4267130817149139, + "grad_norm": 0.43968136919390277, + "learning_rate": 7.089232069058136e-06, + "loss": 0.5014, + "step": 2329 + }, + { + "epoch": 0.4268962990106266, + "grad_norm": 0.37304517337745097, + "learning_rate": 7.086326315739779e-06, + "loss": 0.4529, + "step": 2330 + }, + { + "epoch": 0.42707951630633934, + "grad_norm": 0.40182316420986464, + "learning_rate": 7.083419708996641e-06, + "loss": 0.4991, + "step": 2331 + }, + { + "epoch": 0.427262733602052, + "grad_norm": 0.37377079348812675, + "learning_rate": 7.080512250017682e-06, + "loss": 0.4797, + "step": 2332 + }, + { + "epoch": 0.42744595089776477, + "grad_norm": 0.372732844224778, + "learning_rate": 7.0776039399922235e-06, + "loss": 0.5096, + "step": 2333 + }, + { + "epoch": 0.42762916819347746, + "grad_norm": 0.4343243599651058, + "learning_rate": 7.074694780109923e-06, + "loss": 0.4912, + "step": 2334 + }, + { + "epoch": 0.4278123854891902, + "grad_norm": 0.4198262565591371, + "learning_rate": 7.07178477156079e-06, + "loss": 0.488, + "step": 2335 + }, + { + "epoch": 0.4279956027849029, + "grad_norm": 0.41905722971607356, + "learning_rate": 7.068873915535186e-06, + "loss": 0.5009, + "step": 2336 + }, + { + "epoch": 0.4281788200806156, + "grad_norm": 0.4198816783375255, + "learning_rate": 7.0659622132238105e-06, + "loss": 0.5005, + "step": 2337 + }, + { + "epoch": 0.42836203737632833, + "grad_norm": 0.3985114381144264, + "learning_rate": 7.063049665817715e-06, + "loss": 0.4938, + "step": 2338 + }, + { + "epoch": 0.428545254672041, + "grad_norm": 0.3846276120934009, + "learning_rate": 7.060136274508296e-06, + "loss": 0.4399, + "step": 2339 + }, + { + "epoch": 0.42872847196775377, + "grad_norm": 0.35050377226240487, + "learning_rate": 7.057222040487292e-06, + "loss": 0.4731, + "step": 2340 + }, + { + "epoch": 0.42891168926346646, + "grad_norm": 0.3758851287952108, + "learning_rate": 7.054306964946793e-06, + "loss": 0.4955, + "step": 2341 + }, + { + "epoch": 0.4290949065591792, + "grad_norm": 0.41910557353111433, + "learning_rate": 7.051391049079225e-06, + "loss": 0.4535, + "step": 2342 + }, + { + "epoch": 0.4292781238548919, + "grad_norm": 0.34980515305649706, + "learning_rate": 7.048474294077362e-06, + "loss": 0.475, + "step": 2343 + }, + { + "epoch": 0.42946134115060464, + "grad_norm": 0.46683734709027264, + "learning_rate": 7.045556701134322e-06, + "loss": 0.5025, + "step": 2344 + }, + { + "epoch": 0.4296445584463173, + "grad_norm": 0.39201870728174676, + "learning_rate": 7.042638271443564e-06, + "loss": 0.4706, + "step": 2345 + }, + { + "epoch": 0.42982777574203007, + "grad_norm": 0.4294298317206374, + "learning_rate": 7.039719006198889e-06, + "loss": 0.508, + "step": 2346 + }, + { + "epoch": 0.43001099303774276, + "grad_norm": 0.4378542511044781, + "learning_rate": 7.036798906594442e-06, + "loss": 0.5178, + "step": 2347 + }, + { + "epoch": 0.43019421033345545, + "grad_norm": 0.38517984572050734, + "learning_rate": 7.033877973824708e-06, + "loss": 0.4893, + "step": 2348 + }, + { + "epoch": 0.4303774276291682, + "grad_norm": 0.38827074079168944, + "learning_rate": 7.030956209084512e-06, + "loss": 0.4446, + "step": 2349 + }, + { + "epoch": 0.4305606449248809, + "grad_norm": 0.7716576786129257, + "learning_rate": 7.02803361356902e-06, + "loss": 0.488, + "step": 2350 + }, + { + "epoch": 0.43074386222059363, + "grad_norm": 0.440955197231558, + "learning_rate": 7.02511018847374e-06, + "loss": 0.5003, + "step": 2351 + }, + { + "epoch": 0.4309270795163063, + "grad_norm": 0.4300046727583612, + "learning_rate": 7.022185934994512e-06, + "loss": 0.4669, + "step": 2352 + }, + { + "epoch": 0.43111029681201907, + "grad_norm": 0.37622768004399426, + "learning_rate": 7.019260854327527e-06, + "loss": 0.4614, + "step": 2353 + }, + { + "epoch": 0.43129351410773176, + "grad_norm": 0.41243518545225494, + "learning_rate": 7.016334947669301e-06, + "loss": 0.4778, + "step": 2354 + }, + { + "epoch": 0.4314767314034445, + "grad_norm": 0.38779305226028166, + "learning_rate": 7.0134082162167e-06, + "loss": 0.4373, + "step": 2355 + }, + { + "epoch": 0.4316599486991572, + "grad_norm": 0.414920440559208, + "learning_rate": 7.010480661166916e-06, + "loss": 0.472, + "step": 2356 + }, + { + "epoch": 0.43184316599486994, + "grad_norm": 0.44396932472750594, + "learning_rate": 7.0075522837174905e-06, + "loss": 0.4997, + "step": 2357 + }, + { + "epoch": 0.4320263832905826, + "grad_norm": 0.37434944716162305, + "learning_rate": 7.0046230850662875e-06, + "loss": 0.4431, + "step": 2358 + }, + { + "epoch": 0.4322096005862954, + "grad_norm": 0.3896111851669577, + "learning_rate": 7.00169306641152e-06, + "loss": 0.5243, + "step": 2359 + }, + { + "epoch": 0.43239281788200806, + "grad_norm": 0.48856853232800274, + "learning_rate": 6.9987622289517275e-06, + "loss": 0.4943, + "step": 2360 + }, + { + "epoch": 0.43257603517772075, + "grad_norm": 0.4283382340422475, + "learning_rate": 6.995830573885787e-06, + "loss": 0.4708, + "step": 2361 + }, + { + "epoch": 0.4327592524734335, + "grad_norm": 0.38077833149258267, + "learning_rate": 6.99289810241291e-06, + "loss": 0.463, + "step": 2362 + }, + { + "epoch": 0.4329424697691462, + "grad_norm": 0.39433666825831887, + "learning_rate": 6.989964815732643e-06, + "loss": 0.4608, + "step": 2363 + }, + { + "epoch": 0.43312568706485893, + "grad_norm": 0.39945011068065406, + "learning_rate": 6.9870307150448655e-06, + "loss": 0.4544, + "step": 2364 + }, + { + "epoch": 0.4333089043605716, + "grad_norm": 0.3819624777398398, + "learning_rate": 6.98409580154979e-06, + "loss": 0.4914, + "step": 2365 + }, + { + "epoch": 0.43349212165628437, + "grad_norm": 0.3729087458748407, + "learning_rate": 6.98116007644796e-06, + "loss": 0.4751, + "step": 2366 + }, + { + "epoch": 0.43367533895199706, + "grad_norm": 0.3793765431857637, + "learning_rate": 6.978223540940253e-06, + "loss": 0.4368, + "step": 2367 + }, + { + "epoch": 0.4338585562477098, + "grad_norm": 0.33810640239705714, + "learning_rate": 6.9752861962278775e-06, + "loss": 0.472, + "step": 2368 + }, + { + "epoch": 0.4340417735434225, + "grad_norm": 0.43051317835138775, + "learning_rate": 6.97234804351237e-06, + "loss": 0.4945, + "step": 2369 + }, + { + "epoch": 0.43422499083913524, + "grad_norm": 0.3961468902292914, + "learning_rate": 6.969409083995604e-06, + "loss": 0.5028, + "step": 2370 + }, + { + "epoch": 0.43440820813484793, + "grad_norm": 0.36626866660201934, + "learning_rate": 6.9664693188797776e-06, + "loss": 0.468, + "step": 2371 + }, + { + "epoch": 0.4345914254305606, + "grad_norm": 0.3840127265100837, + "learning_rate": 6.963528749367419e-06, + "loss": 0.4791, + "step": 2372 + }, + { + "epoch": 0.43477464272627336, + "grad_norm": 0.4233350086872284, + "learning_rate": 6.960587376661388e-06, + "loss": 0.4973, + "step": 2373 + }, + { + "epoch": 0.43495786002198605, + "grad_norm": 0.39660597892499455, + "learning_rate": 6.957645201964869e-06, + "loss": 0.5158, + "step": 2374 + }, + { + "epoch": 0.4351410773176988, + "grad_norm": 0.4125283665972468, + "learning_rate": 6.95470222648138e-06, + "loss": 0.4674, + "step": 2375 + }, + { + "epoch": 0.4353242946134115, + "grad_norm": 0.3510950157002611, + "learning_rate": 6.951758451414763e-06, + "loss": 0.4642, + "step": 2376 + }, + { + "epoch": 0.43550751190912423, + "grad_norm": 0.4182866500256308, + "learning_rate": 6.948813877969186e-06, + "loss": 0.4471, + "step": 2377 + }, + { + "epoch": 0.4356907292048369, + "grad_norm": 0.3981987861546714, + "learning_rate": 6.945868507349145e-06, + "loss": 0.4634, + "step": 2378 + }, + { + "epoch": 0.43587394650054967, + "grad_norm": 0.3873084451270221, + "learning_rate": 6.942922340759465e-06, + "loss": 0.4658, + "step": 2379 + }, + { + "epoch": 0.43605716379626236, + "grad_norm": 0.4252831559416119, + "learning_rate": 6.9399753794052885e-06, + "loss": 0.4913, + "step": 2380 + }, + { + "epoch": 0.4362403810919751, + "grad_norm": 0.40795500548552754, + "learning_rate": 6.937027624492096e-06, + "loss": 0.495, + "step": 2381 + }, + { + "epoch": 0.4364235983876878, + "grad_norm": 0.3925888436033216, + "learning_rate": 6.934079077225677e-06, + "loss": 0.4763, + "step": 2382 + }, + { + "epoch": 0.43660681568340054, + "grad_norm": 0.39941871004367224, + "learning_rate": 6.931129738812159e-06, + "loss": 0.4787, + "step": 2383 + }, + { + "epoch": 0.43679003297911323, + "grad_norm": 0.3996234463219394, + "learning_rate": 6.9281796104579835e-06, + "loss": 0.4753, + "step": 2384 + }, + { + "epoch": 0.4369732502748259, + "grad_norm": 0.3922661120163237, + "learning_rate": 6.925228693369923e-06, + "loss": 0.4771, + "step": 2385 + }, + { + "epoch": 0.43715646757053866, + "grad_norm": 0.404326588298471, + "learning_rate": 6.9222769887550635e-06, + "loss": 0.4699, + "step": 2386 + }, + { + "epoch": 0.43733968486625135, + "grad_norm": 0.3655927936043432, + "learning_rate": 6.9193244978208226e-06, + "loss": 0.4693, + "step": 2387 + }, + { + "epoch": 0.4375229021619641, + "grad_norm": 0.3988742818248961, + "learning_rate": 6.916371221774933e-06, + "loss": 0.4987, + "step": 2388 + }, + { + "epoch": 0.4377061194576768, + "grad_norm": 0.3693227091150787, + "learning_rate": 6.913417161825449e-06, + "loss": 0.4807, + "step": 2389 + }, + { + "epoch": 0.43788933675338954, + "grad_norm": 0.3743760464575296, + "learning_rate": 6.9104623191807504e-06, + "loss": 0.455, + "step": 2390 + }, + { + "epoch": 0.4380725540491022, + "grad_norm": 0.4005752769269041, + "learning_rate": 6.907506695049534e-06, + "loss": 0.4968, + "step": 2391 + }, + { + "epoch": 0.43825577134481497, + "grad_norm": 0.40888615444760756, + "learning_rate": 6.904550290640815e-06, + "loss": 0.4893, + "step": 2392 + }, + { + "epoch": 0.43843898864052766, + "grad_norm": 0.7348387573263818, + "learning_rate": 6.901593107163926e-06, + "loss": 0.4669, + "step": 2393 + }, + { + "epoch": 0.4386222059362404, + "grad_norm": 0.39657513243027714, + "learning_rate": 6.898635145828528e-06, + "loss": 0.4635, + "step": 2394 + }, + { + "epoch": 0.4388054232319531, + "grad_norm": 0.43761475845246406, + "learning_rate": 6.895676407844587e-06, + "loss": 0.4616, + "step": 2395 + }, + { + "epoch": 0.4389886405276658, + "grad_norm": 0.41601703075771596, + "learning_rate": 6.8927168944224e-06, + "loss": 0.4825, + "step": 2396 + }, + { + "epoch": 0.43917185782337853, + "grad_norm": 0.420359912271575, + "learning_rate": 6.889756606772566e-06, + "loss": 0.4992, + "step": 2397 + }, + { + "epoch": 0.4393550751190912, + "grad_norm": 0.33481929474990846, + "learning_rate": 6.886795546106018e-06, + "loss": 0.4711, + "step": 2398 + }, + { + "epoch": 0.43953829241480397, + "grad_norm": 0.38476666672073384, + "learning_rate": 6.883833713633988e-06, + "loss": 0.4819, + "step": 2399 + }, + { + "epoch": 0.43972150971051666, + "grad_norm": 0.3827469135624022, + "learning_rate": 6.880871110568039e-06, + "loss": 0.4773, + "step": 2400 + }, + { + "epoch": 0.4399047270062294, + "grad_norm": 0.36455987354043057, + "learning_rate": 6.877907738120038e-06, + "loss": 0.4485, + "step": 2401 + }, + { + "epoch": 0.4400879443019421, + "grad_norm": 0.40927725664501297, + "learning_rate": 6.8749435975021725e-06, + "loss": 0.4673, + "step": 2402 + }, + { + "epoch": 0.44027116159765484, + "grad_norm": 0.3635877075757829, + "learning_rate": 6.8719786899269426e-06, + "loss": 0.4514, + "step": 2403 + }, + { + "epoch": 0.4404543788933675, + "grad_norm": 0.3775825683457327, + "learning_rate": 6.869013016607161e-06, + "loss": 0.4672, + "step": 2404 + }, + { + "epoch": 0.44063759618908027, + "grad_norm": 0.41615535751605043, + "learning_rate": 6.866046578755957e-06, + "loss": 0.5103, + "step": 2405 + }, + { + "epoch": 0.44082081348479296, + "grad_norm": 0.37936631222961226, + "learning_rate": 6.86307937758677e-06, + "loss": 0.5207, + "step": 2406 + }, + { + "epoch": 0.4410040307805057, + "grad_norm": 0.4482772891450156, + "learning_rate": 6.8601114143133495e-06, + "loss": 0.4744, + "step": 2407 + }, + { + "epoch": 0.4411872480762184, + "grad_norm": 0.43659529780579315, + "learning_rate": 6.857142690149763e-06, + "loss": 0.4547, + "step": 2408 + }, + { + "epoch": 0.4413704653719311, + "grad_norm": 0.3999276749098566, + "learning_rate": 6.854173206310384e-06, + "loss": 0.4556, + "step": 2409 + }, + { + "epoch": 0.44155368266764383, + "grad_norm": 0.3656325688650321, + "learning_rate": 6.8512029640098975e-06, + "loss": 0.4776, + "step": 2410 + }, + { + "epoch": 0.4417368999633565, + "grad_norm": 0.3997280896111769, + "learning_rate": 6.848231964463301e-06, + "loss": 0.5314, + "step": 2411 + }, + { + "epoch": 0.44192011725906927, + "grad_norm": 0.404062762092106, + "learning_rate": 6.8452602088859e-06, + "loss": 0.5038, + "step": 2412 + }, + { + "epoch": 0.44210333455478196, + "grad_norm": 0.3896279134336174, + "learning_rate": 6.842287698493311e-06, + "loss": 0.5022, + "step": 2413 + }, + { + "epoch": 0.4422865518504947, + "grad_norm": 0.4009423784359649, + "learning_rate": 6.839314434501454e-06, + "loss": 0.4735, + "step": 2414 + }, + { + "epoch": 0.4424697691462074, + "grad_norm": 0.34610422596550333, + "learning_rate": 6.8363404181265645e-06, + "loss": 0.4381, + "step": 2415 + }, + { + "epoch": 0.44265298644192014, + "grad_norm": 0.3859453907975738, + "learning_rate": 6.833365650585182e-06, + "loss": 0.4704, + "step": 2416 + }, + { + "epoch": 0.4428362037376328, + "grad_norm": 0.41704048460990845, + "learning_rate": 6.830390133094153e-06, + "loss": 0.5059, + "step": 2417 + }, + { + "epoch": 0.4430194210333456, + "grad_norm": 0.37249261818803026, + "learning_rate": 6.8274138668706315e-06, + "loss": 0.4698, + "step": 2418 + }, + { + "epoch": 0.44320263832905826, + "grad_norm": 0.35619151499682244, + "learning_rate": 6.8244368531320795e-06, + "loss": 0.48, + "step": 2419 + }, + { + "epoch": 0.44338585562477095, + "grad_norm": 0.3485686263911084, + "learning_rate": 6.821459093096261e-06, + "loss": 0.4666, + "step": 2420 + }, + { + "epoch": 0.4435690729204837, + "grad_norm": 0.41358337388795335, + "learning_rate": 6.818480587981245e-06, + "loss": 0.4306, + "step": 2421 + }, + { + "epoch": 0.4437522902161964, + "grad_norm": 0.47634180403797133, + "learning_rate": 6.815501339005415e-06, + "loss": 0.4952, + "step": 2422 + }, + { + "epoch": 0.44393550751190913, + "grad_norm": 0.4176110671648082, + "learning_rate": 6.812521347387443e-06, + "loss": 0.4968, + "step": 2423 + }, + { + "epoch": 0.4441187248076218, + "grad_norm": 0.47871777541154165, + "learning_rate": 6.8095406143463195e-06, + "loss": 0.4857, + "step": 2424 + }, + { + "epoch": 0.44430194210333457, + "grad_norm": 0.3775258231394918, + "learning_rate": 6.806559141101326e-06, + "loss": 0.4664, + "step": 2425 + }, + { + "epoch": 0.44448515939904726, + "grad_norm": 0.38874193711392035, + "learning_rate": 6.80357692887206e-06, + "loss": 0.4461, + "step": 2426 + }, + { + "epoch": 0.44466837669476, + "grad_norm": 0.3850454530745506, + "learning_rate": 6.800593978878407e-06, + "loss": 0.4615, + "step": 2427 + }, + { + "epoch": 0.4448515939904727, + "grad_norm": 0.37462058976986146, + "learning_rate": 6.797610292340566e-06, + "loss": 0.4811, + "step": 2428 + }, + { + "epoch": 0.44503481128618544, + "grad_norm": 0.3961801018185967, + "learning_rate": 6.79462587047903e-06, + "loss": 0.4867, + "step": 2429 + }, + { + "epoch": 0.44521802858189813, + "grad_norm": 0.39615959643714926, + "learning_rate": 6.791640714514598e-06, + "loss": 0.5156, + "step": 2430 + }, + { + "epoch": 0.4454012458776109, + "grad_norm": 0.38164096282868487, + "learning_rate": 6.788654825668364e-06, + "loss": 0.4656, + "step": 2431 + }, + { + "epoch": 0.44558446317332356, + "grad_norm": 0.402858239640485, + "learning_rate": 6.785668205161724e-06, + "loss": 0.4768, + "step": 2432 + }, + { + "epoch": 0.44576768046903625, + "grad_norm": 0.40282157120383244, + "learning_rate": 6.782680854216377e-06, + "loss": 0.4768, + "step": 2433 + }, + { + "epoch": 0.445950897764749, + "grad_norm": 0.40626512419257327, + "learning_rate": 6.779692774054314e-06, + "loss": 0.498, + "step": 2434 + }, + { + "epoch": 0.4461341150604617, + "grad_norm": 0.36389289871667907, + "learning_rate": 6.776703965897831e-06, + "loss": 0.4616, + "step": 2435 + }, + { + "epoch": 0.44631733235617443, + "grad_norm": 0.4096643653171753, + "learning_rate": 6.773714430969516e-06, + "loss": 0.4809, + "step": 2436 + }, + { + "epoch": 0.4465005496518871, + "grad_norm": 0.3824306753102932, + "learning_rate": 6.770724170492258e-06, + "loss": 0.4629, + "step": 2437 + }, + { + "epoch": 0.44668376694759987, + "grad_norm": 0.4436204478768461, + "learning_rate": 6.767733185689242e-06, + "loss": 0.4931, + "step": 2438 + }, + { + "epoch": 0.44686698424331256, + "grad_norm": 0.44700888421170765, + "learning_rate": 6.764741477783948e-06, + "loss": 0.4804, + "step": 2439 + }, + { + "epoch": 0.4470502015390253, + "grad_norm": 0.3952398209671749, + "learning_rate": 6.761749048000155e-06, + "loss": 0.4834, + "step": 2440 + }, + { + "epoch": 0.447233418834738, + "grad_norm": 0.37493057776442534, + "learning_rate": 6.758755897561932e-06, + "loss": 0.4957, + "step": 2441 + }, + { + "epoch": 0.44741663613045074, + "grad_norm": 0.5006060287012225, + "learning_rate": 6.7557620276936465e-06, + "loss": 0.5233, + "step": 2442 + }, + { + "epoch": 0.44759985342616343, + "grad_norm": 0.4134212031355089, + "learning_rate": 6.752767439619961e-06, + "loss": 0.4584, + "step": 2443 + }, + { + "epoch": 0.4477830707218761, + "grad_norm": 0.3999250364659294, + "learning_rate": 6.749772134565831e-06, + "loss": 0.4946, + "step": 2444 + }, + { + "epoch": 0.44796628801758887, + "grad_norm": 0.3737698467337494, + "learning_rate": 6.746776113756503e-06, + "loss": 0.4788, + "step": 2445 + }, + { + "epoch": 0.44814950531330155, + "grad_norm": 0.4223216096535037, + "learning_rate": 6.743779378417519e-06, + "loss": 0.4876, + "step": 2446 + }, + { + "epoch": 0.4483327226090143, + "grad_norm": 0.431613620476394, + "learning_rate": 6.74078192977471e-06, + "loss": 0.4873, + "step": 2447 + }, + { + "epoch": 0.448515939904727, + "grad_norm": 0.38431685544468647, + "learning_rate": 6.737783769054205e-06, + "loss": 0.4825, + "step": 2448 + }, + { + "epoch": 0.44869915720043974, + "grad_norm": 0.3911098376394661, + "learning_rate": 6.734784897482416e-06, + "loss": 0.4725, + "step": 2449 + }, + { + "epoch": 0.4488823744961524, + "grad_norm": 0.3915504629811401, + "learning_rate": 6.731785316286055e-06, + "loss": 0.4983, + "step": 2450 + }, + { + "epoch": 0.44906559179186517, + "grad_norm": 0.4148113185798, + "learning_rate": 6.728785026692113e-06, + "loss": 0.4817, + "step": 2451 + }, + { + "epoch": 0.44924880908757786, + "grad_norm": 0.624108531595354, + "learning_rate": 6.725784029927884e-06, + "loss": 0.4835, + "step": 2452 + }, + { + "epoch": 0.4494320263832906, + "grad_norm": 0.4069043276199038, + "learning_rate": 6.722782327220939e-06, + "loss": 0.487, + "step": 2453 + }, + { + "epoch": 0.4496152436790033, + "grad_norm": 0.40931666347912105, + "learning_rate": 6.719779919799148e-06, + "loss": 0.4585, + "step": 2454 + }, + { + "epoch": 0.449798460974716, + "grad_norm": 0.4057761872265912, + "learning_rate": 6.716776808890659e-06, + "loss": 0.468, + "step": 2455 + }, + { + "epoch": 0.44998167827042873, + "grad_norm": 0.4042031306804991, + "learning_rate": 6.713772995723919e-06, + "loss": 0.4949, + "step": 2456 + }, + { + "epoch": 0.4501648955661414, + "grad_norm": 0.38813893626821355, + "learning_rate": 6.710768481527652e-06, + "loss": 0.4987, + "step": 2457 + }, + { + "epoch": 0.45034811286185417, + "grad_norm": 0.3590174575597566, + "learning_rate": 6.707763267530875e-06, + "loss": 0.475, + "step": 2458 + }, + { + "epoch": 0.45053133015756686, + "grad_norm": 0.4198445880391634, + "learning_rate": 6.704757354962888e-06, + "loss": 0.49, + "step": 2459 + }, + { + "epoch": 0.4507145474532796, + "grad_norm": 0.43834942507027247, + "learning_rate": 6.701750745053281e-06, + "loss": 0.4558, + "step": 2460 + }, + { + "epoch": 0.4508977647489923, + "grad_norm": 0.38619095225610495, + "learning_rate": 6.698743439031926e-06, + "loss": 0.4656, + "step": 2461 + }, + { + "epoch": 0.45108098204470504, + "grad_norm": 0.40078588702517964, + "learning_rate": 6.695735438128979e-06, + "loss": 0.463, + "step": 2462 + }, + { + "epoch": 0.4512641993404177, + "grad_norm": 0.3609087881626896, + "learning_rate": 6.6927267435748826e-06, + "loss": 0.4986, + "step": 2463 + }, + { + "epoch": 0.45144741663613047, + "grad_norm": 0.4351226156707901, + "learning_rate": 6.6897173566003595e-06, + "loss": 0.4983, + "step": 2464 + }, + { + "epoch": 0.45163063393184316, + "grad_norm": 0.37576273348695177, + "learning_rate": 6.686707278436422e-06, + "loss": 0.492, + "step": 2465 + }, + { + "epoch": 0.4518138512275559, + "grad_norm": 0.44002135284894817, + "learning_rate": 6.683696510314358e-06, + "loss": 0.4524, + "step": 2466 + }, + { + "epoch": 0.4519970685232686, + "grad_norm": 0.39162403492310033, + "learning_rate": 6.680685053465743e-06, + "loss": 0.4884, + "step": 2467 + }, + { + "epoch": 0.4521802858189813, + "grad_norm": 0.44367901990535713, + "learning_rate": 6.677672909122432e-06, + "loss": 0.5004, + "step": 2468 + }, + { + "epoch": 0.45236350311469403, + "grad_norm": 0.37838149865303566, + "learning_rate": 6.6746600785165605e-06, + "loss": 0.4672, + "step": 2469 + }, + { + "epoch": 0.4525467204104067, + "grad_norm": 0.398874364176312, + "learning_rate": 6.671646562880547e-06, + "loss": 0.4942, + "step": 2470 + }, + { + "epoch": 0.45272993770611947, + "grad_norm": 0.38110520416545984, + "learning_rate": 6.6686323634470875e-06, + "loss": 0.4677, + "step": 2471 + }, + { + "epoch": 0.45291315500183216, + "grad_norm": 0.41232886013451225, + "learning_rate": 6.66561748144916e-06, + "loss": 0.4956, + "step": 2472 + }, + { + "epoch": 0.4530963722975449, + "grad_norm": 0.354539591646497, + "learning_rate": 6.662601918120019e-06, + "loss": 0.4488, + "step": 2473 + }, + { + "epoch": 0.4532795895932576, + "grad_norm": 0.35518785769248845, + "learning_rate": 6.659585674693203e-06, + "loss": 0.4467, + "step": 2474 + }, + { + "epoch": 0.45346280688897034, + "grad_norm": 0.4420732987691351, + "learning_rate": 6.656568752402521e-06, + "loss": 0.4995, + "step": 2475 + }, + { + "epoch": 0.453646024184683, + "grad_norm": 0.34062803748892734, + "learning_rate": 6.6535511524820684e-06, + "loss": 0.4956, + "step": 2476 + }, + { + "epoch": 0.4538292414803958, + "grad_norm": 0.3672786509720195, + "learning_rate": 6.650532876166208e-06, + "loss": 0.4948, + "step": 2477 + }, + { + "epoch": 0.45401245877610846, + "grad_norm": 0.45120405587126444, + "learning_rate": 6.64751392468959e-06, + "loss": 0.4688, + "step": 2478 + }, + { + "epoch": 0.45419567607182115, + "grad_norm": 0.3661120556835148, + "learning_rate": 6.64449429928713e-06, + "loss": 0.4725, + "step": 2479 + }, + { + "epoch": 0.4543788933675339, + "grad_norm": 0.5179467767248828, + "learning_rate": 6.64147400119403e-06, + "loss": 0.4738, + "step": 2480 + }, + { + "epoch": 0.4545621106632466, + "grad_norm": 0.36739695479670653, + "learning_rate": 6.6384530316457555e-06, + "loss": 0.4652, + "step": 2481 + }, + { + "epoch": 0.45474532795895933, + "grad_norm": 0.36407763117656494, + "learning_rate": 6.63543139187806e-06, + "loss": 0.464, + "step": 2482 + }, + { + "epoch": 0.454928545254672, + "grad_norm": 0.5281645622956309, + "learning_rate": 6.632409083126959e-06, + "loss": 0.4929, + "step": 2483 + }, + { + "epoch": 0.45511176255038477, + "grad_norm": 0.3628507640542516, + "learning_rate": 6.629386106628746e-06, + "loss": 0.4623, + "step": 2484 + }, + { + "epoch": 0.45529497984609746, + "grad_norm": 0.3751108165654219, + "learning_rate": 6.626362463619993e-06, + "loss": 0.4918, + "step": 2485 + }, + { + "epoch": 0.4554781971418102, + "grad_norm": 0.37070836465790696, + "learning_rate": 6.623338155337538e-06, + "loss": 0.4798, + "step": 2486 + }, + { + "epoch": 0.4556614144375229, + "grad_norm": 0.38911661141673254, + "learning_rate": 6.620313183018491e-06, + "loss": 0.458, + "step": 2487 + }, + { + "epoch": 0.45584463173323564, + "grad_norm": 0.4162114279088436, + "learning_rate": 6.6172875479002385e-06, + "loss": 0.5018, + "step": 2488 + }, + { + "epoch": 0.45602784902894833, + "grad_norm": 0.36400542608541064, + "learning_rate": 6.614261251220434e-06, + "loss": 0.48, + "step": 2489 + }, + { + "epoch": 0.4562110663246611, + "grad_norm": 0.41647564044810204, + "learning_rate": 6.6112342942170035e-06, + "loss": 0.4899, + "step": 2490 + }, + { + "epoch": 0.45639428362037376, + "grad_norm": 0.3628507499775755, + "learning_rate": 6.608206678128143e-06, + "loss": 0.4594, + "step": 2491 + }, + { + "epoch": 0.45657750091608645, + "grad_norm": 0.38712578647221024, + "learning_rate": 6.605178404192317e-06, + "loss": 0.4879, + "step": 2492 + }, + { + "epoch": 0.4567607182117992, + "grad_norm": 0.37518676056759626, + "learning_rate": 6.602149473648261e-06, + "loss": 0.479, + "step": 2493 + }, + { + "epoch": 0.4569439355075119, + "grad_norm": 0.480134849540764, + "learning_rate": 6.599119887734976e-06, + "loss": 0.5144, + "step": 2494 + }, + { + "epoch": 0.45712715280322463, + "grad_norm": 0.3942091182072091, + "learning_rate": 6.596089647691735e-06, + "loss": 0.4587, + "step": 2495 + }, + { + "epoch": 0.4573103700989373, + "grad_norm": 0.361810255366327, + "learning_rate": 6.593058754758075e-06, + "loss": 0.4577, + "step": 2496 + }, + { + "epoch": 0.45749358739465007, + "grad_norm": 0.39602059798478256, + "learning_rate": 6.590027210173802e-06, + "loss": 0.4993, + "step": 2497 + }, + { + "epoch": 0.45767680469036276, + "grad_norm": 0.34772691807907075, + "learning_rate": 6.58699501517899e-06, + "loss": 0.4591, + "step": 2498 + }, + { + "epoch": 0.4578600219860755, + "grad_norm": 0.40415781626108666, + "learning_rate": 6.583962171013974e-06, + "loss": 0.4695, + "step": 2499 + }, + { + "epoch": 0.4580432392817882, + "grad_norm": 0.3521153681401539, + "learning_rate": 6.58092867891936e-06, + "loss": 0.4527, + "step": 2500 + }, + { + "epoch": 0.45822645657750094, + "grad_norm": 0.35566184115624544, + "learning_rate": 6.577894540136015e-06, + "loss": 0.4541, + "step": 2501 + }, + { + "epoch": 0.45840967387321363, + "grad_norm": 0.4110067670808069, + "learning_rate": 6.574859755905076e-06, + "loss": 0.4934, + "step": 2502 + }, + { + "epoch": 0.4585928911689263, + "grad_norm": 0.36639060831174014, + "learning_rate": 6.571824327467938e-06, + "loss": 0.4998, + "step": 2503 + }, + { + "epoch": 0.45877610846463907, + "grad_norm": 0.37909030226200563, + "learning_rate": 6.568788256066262e-06, + "loss": 0.5128, + "step": 2504 + }, + { + "epoch": 0.45895932576035176, + "grad_norm": 0.3904517350360066, + "learning_rate": 6.56575154294197e-06, + "loss": 0.4934, + "step": 2505 + }, + { + "epoch": 0.4591425430560645, + "grad_norm": 0.38239361159761764, + "learning_rate": 6.562714189337253e-06, + "loss": 0.4737, + "step": 2506 + }, + { + "epoch": 0.4593257603517772, + "grad_norm": 0.36309795150733687, + "learning_rate": 6.559676196494555e-06, + "loss": 0.4673, + "step": 2507 + }, + { + "epoch": 0.45950897764748994, + "grad_norm": 0.40413199928868393, + "learning_rate": 6.556637565656589e-06, + "loss": 0.5066, + "step": 2508 + }, + { + "epoch": 0.4596921949432026, + "grad_norm": 0.4284520932886585, + "learning_rate": 6.553598298066325e-06, + "loss": 0.4768, + "step": 2509 + }, + { + "epoch": 0.45987541223891537, + "grad_norm": 0.39173052675534187, + "learning_rate": 6.550558394966994e-06, + "loss": 0.492, + "step": 2510 + }, + { + "epoch": 0.46005862953462806, + "grad_norm": 0.38296744058544785, + "learning_rate": 6.547517857602086e-06, + "loss": 0.4647, + "step": 2511 + }, + { + "epoch": 0.4602418468303408, + "grad_norm": 0.37207617566583406, + "learning_rate": 6.544476687215354e-06, + "loss": 0.4941, + "step": 2512 + }, + { + "epoch": 0.4604250641260535, + "grad_norm": 0.3941001722011083, + "learning_rate": 6.541434885050807e-06, + "loss": 0.4959, + "step": 2513 + }, + { + "epoch": 0.46060828142176624, + "grad_norm": 0.37798568910896285, + "learning_rate": 6.538392452352713e-06, + "loss": 0.4903, + "step": 2514 + }, + { + "epoch": 0.46079149871747893, + "grad_norm": 0.3493915523131383, + "learning_rate": 6.535349390365597e-06, + "loss": 0.4699, + "step": 2515 + }, + { + "epoch": 0.4609747160131916, + "grad_norm": 0.427403925768062, + "learning_rate": 6.532305700334245e-06, + "loss": 0.5078, + "step": 2516 + }, + { + "epoch": 0.46115793330890437, + "grad_norm": 0.3298808391165227, + "learning_rate": 6.529261383503695e-06, + "loss": 0.4363, + "step": 2517 + }, + { + "epoch": 0.46134115060461706, + "grad_norm": 0.40879910077462445, + "learning_rate": 6.526216441119244e-06, + "loss": 0.5028, + "step": 2518 + }, + { + "epoch": 0.4615243679003298, + "grad_norm": 0.3882155693456558, + "learning_rate": 6.523170874426445e-06, + "loss": 0.4744, + "step": 2519 + }, + { + "epoch": 0.4617075851960425, + "grad_norm": 0.5524216348949658, + "learning_rate": 6.520124684671107e-06, + "loss": 0.4689, + "step": 2520 + }, + { + "epoch": 0.46189080249175524, + "grad_norm": 0.3857013987127986, + "learning_rate": 6.517077873099294e-06, + "loss": 0.4799, + "step": 2521 + }, + { + "epoch": 0.4620740197874679, + "grad_norm": 0.469996829805299, + "learning_rate": 6.514030440957319e-06, + "loss": 0.4902, + "step": 2522 + }, + { + "epoch": 0.46225723708318067, + "grad_norm": 0.37990408912020956, + "learning_rate": 6.510982389491756e-06, + "loss": 0.4738, + "step": 2523 + }, + { + "epoch": 0.46244045437889336, + "grad_norm": 0.43046593737024863, + "learning_rate": 6.507933719949431e-06, + "loss": 0.4866, + "step": 2524 + }, + { + "epoch": 0.4626236716746061, + "grad_norm": 0.7389353329433919, + "learning_rate": 6.504884433577417e-06, + "loss": 0.4587, + "step": 2525 + }, + { + "epoch": 0.4628068889703188, + "grad_norm": 0.3943412617039225, + "learning_rate": 6.501834531623048e-06, + "loss": 0.4794, + "step": 2526 + }, + { + "epoch": 0.4629901062660315, + "grad_norm": 0.38460652703629394, + "learning_rate": 6.498784015333901e-06, + "loss": 0.4749, + "step": 2527 + }, + { + "epoch": 0.46317332356174423, + "grad_norm": 0.4128915556666103, + "learning_rate": 6.495732885957813e-06, + "loss": 0.4791, + "step": 2528 + }, + { + "epoch": 0.4633565408574569, + "grad_norm": 0.6143189144340108, + "learning_rate": 6.492681144742864e-06, + "loss": 0.5017, + "step": 2529 + }, + { + "epoch": 0.46353975815316967, + "grad_norm": 0.38407626173050746, + "learning_rate": 6.489628792937391e-06, + "loss": 0.4744, + "step": 2530 + }, + { + "epoch": 0.46372297544888236, + "grad_norm": 0.38311124912656713, + "learning_rate": 6.486575831789974e-06, + "loss": 0.4989, + "step": 2531 + }, + { + "epoch": 0.4639061927445951, + "grad_norm": 0.3679740166911545, + "learning_rate": 6.483522262549449e-06, + "loss": 0.4721, + "step": 2532 + }, + { + "epoch": 0.4640894100403078, + "grad_norm": 0.35940980908602577, + "learning_rate": 6.480468086464895e-06, + "loss": 0.4714, + "step": 2533 + }, + { + "epoch": 0.46427262733602054, + "grad_norm": 0.36944546662854455, + "learning_rate": 6.477413304785645e-06, + "loss": 0.4458, + "step": 2534 + }, + { + "epoch": 0.46445584463173323, + "grad_norm": 0.4140734491534397, + "learning_rate": 6.474357918761271e-06, + "loss": 0.521, + "step": 2535 + }, + { + "epoch": 0.464639061927446, + "grad_norm": 0.38409259151293335, + "learning_rate": 6.471301929641604e-06, + "loss": 0.4746, + "step": 2536 + }, + { + "epoch": 0.46482227922315866, + "grad_norm": 0.3759175972151893, + "learning_rate": 6.468245338676711e-06, + "loss": 0.4806, + "step": 2537 + }, + { + "epoch": 0.4650054965188714, + "grad_norm": 0.7549471142146603, + "learning_rate": 6.465188147116911e-06, + "loss": 0.5046, + "step": 2538 + }, + { + "epoch": 0.4651887138145841, + "grad_norm": 0.44343946401890544, + "learning_rate": 6.462130356212768e-06, + "loss": 0.4855, + "step": 2539 + }, + { + "epoch": 0.4653719311102968, + "grad_norm": 0.38619943327242534, + "learning_rate": 6.459071967215089e-06, + "loss": 0.4798, + "step": 2540 + }, + { + "epoch": 0.46555514840600953, + "grad_norm": 0.4210369327718628, + "learning_rate": 6.456012981374928e-06, + "loss": 0.4811, + "step": 2541 + }, + { + "epoch": 0.4657383657017222, + "grad_norm": 0.42104823228565075, + "learning_rate": 6.452953399943582e-06, + "loss": 0.4998, + "step": 2542 + }, + { + "epoch": 0.46592158299743497, + "grad_norm": 0.3906699369342368, + "learning_rate": 6.449893224172593e-06, + "loss": 0.4725, + "step": 2543 + }, + { + "epoch": 0.46610480029314766, + "grad_norm": 0.3725120338780339, + "learning_rate": 6.446832455313743e-06, + "loss": 0.4725, + "step": 2544 + }, + { + "epoch": 0.4662880175888604, + "grad_norm": 0.44445192312634757, + "learning_rate": 6.44377109461906e-06, + "loss": 0.4641, + "step": 2545 + }, + { + "epoch": 0.4664712348845731, + "grad_norm": 0.3544016724918463, + "learning_rate": 6.440709143340811e-06, + "loss": 0.4927, + "step": 2546 + }, + { + "epoch": 0.46665445218028584, + "grad_norm": 0.4513705244339124, + "learning_rate": 6.437646602731509e-06, + "loss": 0.5023, + "step": 2547 + }, + { + "epoch": 0.46683766947599853, + "grad_norm": 0.3792756413426005, + "learning_rate": 6.434583474043902e-06, + "loss": 0.4856, + "step": 2548 + }, + { + "epoch": 0.4670208867717113, + "grad_norm": 0.42224727608203083, + "learning_rate": 6.431519758530984e-06, + "loss": 0.4611, + "step": 2549 + }, + { + "epoch": 0.46720410406742396, + "grad_norm": 0.39968269155459885, + "learning_rate": 6.428455457445987e-06, + "loss": 0.483, + "step": 2550 + }, + { + "epoch": 0.46738732136313665, + "grad_norm": 0.38028019402749325, + "learning_rate": 6.425390572042382e-06, + "loss": 0.4877, + "step": 2551 + }, + { + "epoch": 0.4675705386588494, + "grad_norm": 0.3666037694550172, + "learning_rate": 6.42232510357388e-06, + "loss": 0.5099, + "step": 2552 + }, + { + "epoch": 0.4677537559545621, + "grad_norm": 0.3968544348003642, + "learning_rate": 6.419259053294428e-06, + "loss": 0.4257, + "step": 2553 + }, + { + "epoch": 0.46793697325027483, + "grad_norm": 0.37481778140226907, + "learning_rate": 6.416192422458216e-06, + "loss": 0.4892, + "step": 2554 + }, + { + "epoch": 0.4681201905459875, + "grad_norm": 0.5218930688702873, + "learning_rate": 6.413125212319664e-06, + "loss": 0.4724, + "step": 2555 + }, + { + "epoch": 0.46830340784170027, + "grad_norm": 0.4386511012649629, + "learning_rate": 6.41005742413344e-06, + "loss": 0.4693, + "step": 2556 + }, + { + "epoch": 0.46848662513741296, + "grad_norm": 0.3759734757268749, + "learning_rate": 6.406989059154436e-06, + "loss": 0.4812, + "step": 2557 + }, + { + "epoch": 0.4686698424331257, + "grad_norm": 0.3741153619238794, + "learning_rate": 6.403920118637789e-06, + "loss": 0.4907, + "step": 2558 + }, + { + "epoch": 0.4688530597288384, + "grad_norm": 0.3870167404852689, + "learning_rate": 6.4008506038388655e-06, + "loss": 0.4815, + "step": 2559 + }, + { + "epoch": 0.46903627702455114, + "grad_norm": 0.42182053160245986, + "learning_rate": 6.397780516013273e-06, + "loss": 0.4932, + "step": 2560 + }, + { + "epoch": 0.46921949432026383, + "grad_norm": 0.3886751339233232, + "learning_rate": 6.394709856416849e-06, + "loss": 0.4735, + "step": 2561 + }, + { + "epoch": 0.4694027116159766, + "grad_norm": 0.33091152434712795, + "learning_rate": 6.391638626305663e-06, + "loss": 0.4748, + "step": 2562 + }, + { + "epoch": 0.46958592891168927, + "grad_norm": 0.3823377500939304, + "learning_rate": 6.388566826936025e-06, + "loss": 0.5109, + "step": 2563 + }, + { + "epoch": 0.46976914620740196, + "grad_norm": 0.41067411364371265, + "learning_rate": 6.385494459564472e-06, + "loss": 0.4773, + "step": 2564 + }, + { + "epoch": 0.4699523635031147, + "grad_norm": 0.42999298876833103, + "learning_rate": 6.382421525447774e-06, + "loss": 0.4823, + "step": 2565 + }, + { + "epoch": 0.4701355807988274, + "grad_norm": 0.4090212556220414, + "learning_rate": 6.379348025842935e-06, + "loss": 0.5059, + "step": 2566 + }, + { + "epoch": 0.47031879809454014, + "grad_norm": 0.3764731972964885, + "learning_rate": 6.376273962007189e-06, + "loss": 0.4707, + "step": 2567 + }, + { + "epoch": 0.4705020153902528, + "grad_norm": 0.345697680634986, + "learning_rate": 6.373199335198002e-06, + "loss": 0.4496, + "step": 2568 + }, + { + "epoch": 0.47068523268596557, + "grad_norm": 0.3547159369482873, + "learning_rate": 6.370124146673067e-06, + "loss": 0.4747, + "step": 2569 + }, + { + "epoch": 0.47086844998167826, + "grad_norm": 0.43160519732649605, + "learning_rate": 6.367048397690309e-06, + "loss": 0.4858, + "step": 2570 + }, + { + "epoch": 0.471051667277391, + "grad_norm": 0.39733398439483103, + "learning_rate": 6.363972089507886e-06, + "loss": 0.5075, + "step": 2571 + }, + { + "epoch": 0.4712348845731037, + "grad_norm": 0.3916262221059681, + "learning_rate": 6.360895223384177e-06, + "loss": 0.4828, + "step": 2572 + }, + { + "epoch": 0.47141810186881644, + "grad_norm": 0.40461249382290765, + "learning_rate": 6.357817800577795e-06, + "loss": 0.467, + "step": 2573 + }, + { + "epoch": 0.47160131916452913, + "grad_norm": 0.3973961980925733, + "learning_rate": 6.3547398223475784e-06, + "loss": 0.4836, + "step": 2574 + }, + { + "epoch": 0.4717845364602418, + "grad_norm": 0.35401313268032314, + "learning_rate": 6.3516612899525955e-06, + "loss": 0.4402, + "step": 2575 + }, + { + "epoch": 0.47196775375595457, + "grad_norm": 0.40391831341777373, + "learning_rate": 6.3485822046521365e-06, + "loss": 0.4662, + "step": 2576 + }, + { + "epoch": 0.47215097105166726, + "grad_norm": 0.4313861610200162, + "learning_rate": 6.34550256770572e-06, + "loss": 0.4655, + "step": 2577 + }, + { + "epoch": 0.47233418834738, + "grad_norm": 0.3928530117989006, + "learning_rate": 6.342422380373094e-06, + "loss": 0.4761, + "step": 2578 + }, + { + "epoch": 0.4725174056430927, + "grad_norm": 0.35869713797775465, + "learning_rate": 6.3393416439142255e-06, + "loss": 0.4857, + "step": 2579 + }, + { + "epoch": 0.47270062293880544, + "grad_norm": 0.4264670441646961, + "learning_rate": 6.33626035958931e-06, + "loss": 0.481, + "step": 2580 + }, + { + "epoch": 0.4728838402345181, + "grad_norm": 0.38614293779297487, + "learning_rate": 6.333178528658765e-06, + "loss": 0.5089, + "step": 2581 + }, + { + "epoch": 0.47306705753023087, + "grad_norm": 0.3738289668758061, + "learning_rate": 6.330096152383235e-06, + "loss": 0.4339, + "step": 2582 + }, + { + "epoch": 0.47325027482594356, + "grad_norm": 0.39459933436257694, + "learning_rate": 6.3270132320235825e-06, + "loss": 0.4517, + "step": 2583 + }, + { + "epoch": 0.4734334921216563, + "grad_norm": 0.38726088667071384, + "learning_rate": 6.323929768840897e-06, + "loss": 0.4851, + "step": 2584 + }, + { + "epoch": 0.473616709417369, + "grad_norm": 0.3749470969003082, + "learning_rate": 6.320845764096488e-06, + "loss": 0.493, + "step": 2585 + }, + { + "epoch": 0.4737999267130817, + "grad_norm": 0.3995794128262505, + "learning_rate": 6.3177612190518875e-06, + "loss": 0.5007, + "step": 2586 + }, + { + "epoch": 0.47398314400879443, + "grad_norm": 0.4112252636217348, + "learning_rate": 6.314676134968845e-06, + "loss": 0.4999, + "step": 2587 + }, + { + "epoch": 0.4741663613045071, + "grad_norm": 0.38940508373613664, + "learning_rate": 6.311590513109338e-06, + "loss": 0.5096, + "step": 2588 + }, + { + "epoch": 0.47434957860021987, + "grad_norm": 0.39420357023862646, + "learning_rate": 6.308504354735555e-06, + "loss": 0.4763, + "step": 2589 + }, + { + "epoch": 0.47453279589593256, + "grad_norm": 0.3580291110881397, + "learning_rate": 6.305417661109911e-06, + "loss": 0.4649, + "step": 2590 + }, + { + "epoch": 0.4747160131916453, + "grad_norm": 0.36588057715753436, + "learning_rate": 6.302330433495036e-06, + "loss": 0.4479, + "step": 2591 + }, + { + "epoch": 0.474899230487358, + "grad_norm": 0.3756890469287734, + "learning_rate": 6.29924267315378e-06, + "loss": 0.4677, + "step": 2592 + }, + { + "epoch": 0.47508244778307074, + "grad_norm": 0.4291184184547413, + "learning_rate": 6.296154381349211e-06, + "loss": 0.4974, + "step": 2593 + }, + { + "epoch": 0.47526566507878343, + "grad_norm": 0.3790045341481231, + "learning_rate": 6.2930655593446146e-06, + "loss": 0.4858, + "step": 2594 + }, + { + "epoch": 0.4754488823744962, + "grad_norm": 0.4056703217830834, + "learning_rate": 6.28997620840349e-06, + "loss": 0.4791, + "step": 2595 + }, + { + "epoch": 0.47563209967020886, + "grad_norm": 0.41119889493820067, + "learning_rate": 6.286886329789559e-06, + "loss": 0.4626, + "step": 2596 + }, + { + "epoch": 0.4758153169659216, + "grad_norm": 0.33100862594769004, + "learning_rate": 6.283795924766754e-06, + "loss": 0.4558, + "step": 2597 + }, + { + "epoch": 0.4759985342616343, + "grad_norm": 0.34082284456459994, + "learning_rate": 6.280704994599222e-06, + "loss": 0.4519, + "step": 2598 + }, + { + "epoch": 0.476181751557347, + "grad_norm": 0.376348452651224, + "learning_rate": 6.2776135405513314e-06, + "loss": 0.4844, + "step": 2599 + }, + { + "epoch": 0.47636496885305973, + "grad_norm": 0.531461160305974, + "learning_rate": 6.2745215638876586e-06, + "loss": 0.4813, + "step": 2600 + }, + { + "epoch": 0.4765481861487724, + "grad_norm": 0.37430964762455465, + "learning_rate": 6.271429065872998e-06, + "loss": 0.5149, + "step": 2601 + }, + { + "epoch": 0.47673140344448517, + "grad_norm": 0.5186163001556444, + "learning_rate": 6.268336047772352e-06, + "loss": 0.4933, + "step": 2602 + }, + { + "epoch": 0.47691462074019786, + "grad_norm": 0.3802160340640398, + "learning_rate": 6.26524251085094e-06, + "loss": 0.5113, + "step": 2603 + }, + { + "epoch": 0.4770978380359106, + "grad_norm": 0.4356133163853863, + "learning_rate": 6.262148456374196e-06, + "loss": 0.5008, + "step": 2604 + }, + { + "epoch": 0.4772810553316233, + "grad_norm": 0.3703370173501804, + "learning_rate": 6.259053885607756e-06, + "loss": 0.4745, + "step": 2605 + }, + { + "epoch": 0.47746427262733604, + "grad_norm": 0.3781590454920534, + "learning_rate": 6.255958799817478e-06, + "loss": 0.4526, + "step": 2606 + }, + { + "epoch": 0.47764748992304873, + "grad_norm": 0.3289646822012959, + "learning_rate": 6.2528632002694245e-06, + "loss": 0.4869, + "step": 2607 + }, + { + "epoch": 0.4778307072187615, + "grad_norm": 0.43497538600625046, + "learning_rate": 6.249767088229871e-06, + "loss": 0.5055, + "step": 2608 + }, + { + "epoch": 0.47801392451447416, + "grad_norm": 0.43505507795166526, + "learning_rate": 6.2466704649653e-06, + "loss": 0.4736, + "step": 2609 + }, + { + "epoch": 0.47819714181018685, + "grad_norm": 0.4045106146055078, + "learning_rate": 6.2435733317424055e-06, + "loss": 0.4801, + "step": 2610 + }, + { + "epoch": 0.4783803591058996, + "grad_norm": 0.3638264847166178, + "learning_rate": 6.240475689828087e-06, + "loss": 0.5099, + "step": 2611 + }, + { + "epoch": 0.4785635764016123, + "grad_norm": 0.42192732346793493, + "learning_rate": 6.23737754048946e-06, + "loss": 0.5108, + "step": 2612 + }, + { + "epoch": 0.47874679369732503, + "grad_norm": 0.4849536725194334, + "learning_rate": 6.234278884993834e-06, + "loss": 0.493, + "step": 2613 + }, + { + "epoch": 0.4789300109930377, + "grad_norm": 0.377217363879068, + "learning_rate": 6.231179724608739e-06, + "loss": 0.4802, + "step": 2614 + }, + { + "epoch": 0.47911322828875047, + "grad_norm": 0.3929022282057568, + "learning_rate": 6.228080060601904e-06, + "loss": 0.4894, + "step": 2615 + }, + { + "epoch": 0.47929644558446316, + "grad_norm": 0.3568711088138425, + "learning_rate": 6.224979894241265e-06, + "loss": 0.4601, + "step": 2616 + }, + { + "epoch": 0.4794796628801759, + "grad_norm": 0.3452250600579113, + "learning_rate": 6.221879226794967e-06, + "loss": 0.461, + "step": 2617 + }, + { + "epoch": 0.4796628801758886, + "grad_norm": 0.36690268084008215, + "learning_rate": 6.218778059531354e-06, + "loss": 0.4868, + "step": 2618 + }, + { + "epoch": 0.47984609747160134, + "grad_norm": 0.4173051314210318, + "learning_rate": 6.21567639371898e-06, + "loss": 0.5254, + "step": 2619 + }, + { + "epoch": 0.48002931476731403, + "grad_norm": 0.4507443076619217, + "learning_rate": 6.2125742306266e-06, + "loss": 0.4984, + "step": 2620 + }, + { + "epoch": 0.4802125320630268, + "grad_norm": 0.38436958659890863, + "learning_rate": 6.209471571523173e-06, + "loss": 0.4511, + "step": 2621 + }, + { + "epoch": 0.48039574935873947, + "grad_norm": 0.440982519444724, + "learning_rate": 6.206368417677862e-06, + "loss": 0.4786, + "step": 2622 + }, + { + "epoch": 0.48057896665445216, + "grad_norm": 0.35547041740574964, + "learning_rate": 6.203264770360032e-06, + "loss": 0.4666, + "step": 2623 + }, + { + "epoch": 0.4807621839501649, + "grad_norm": 0.37249142511781896, + "learning_rate": 6.200160630839246e-06, + "loss": 0.491, + "step": 2624 + }, + { + "epoch": 0.4809454012458776, + "grad_norm": 0.4584014148979973, + "learning_rate": 6.197056000385275e-06, + "loss": 0.5109, + "step": 2625 + }, + { + "epoch": 0.48112861854159034, + "grad_norm": 0.38097731903257886, + "learning_rate": 6.193950880268085e-06, + "loss": 0.4847, + "step": 2626 + }, + { + "epoch": 0.481311835837303, + "grad_norm": 0.33710104804065255, + "learning_rate": 6.190845271757846e-06, + "loss": 0.4636, + "step": 2627 + }, + { + "epoch": 0.48149505313301577, + "grad_norm": 0.38832073039942233, + "learning_rate": 6.187739176124927e-06, + "loss": 0.4819, + "step": 2628 + }, + { + "epoch": 0.48167827042872846, + "grad_norm": 0.35354719449279465, + "learning_rate": 6.184632594639895e-06, + "loss": 0.4758, + "step": 2629 + }, + { + "epoch": 0.4818614877244412, + "grad_norm": 0.3480101599009128, + "learning_rate": 6.181525528573518e-06, + "loss": 0.4995, + "step": 2630 + }, + { + "epoch": 0.4820447050201539, + "grad_norm": 0.380444579223931, + "learning_rate": 6.1784179791967565e-06, + "loss": 0.465, + "step": 2631 + }, + { + "epoch": 0.48222792231586664, + "grad_norm": 0.42875717985837525, + "learning_rate": 6.175309947780779e-06, + "loss": 0.5161, + "step": 2632 + }, + { + "epoch": 0.48241113961157933, + "grad_norm": 0.3486687159796772, + "learning_rate": 6.172201435596938e-06, + "loss": 0.4595, + "step": 2633 + }, + { + "epoch": 0.482594356907292, + "grad_norm": 0.35070994769653885, + "learning_rate": 6.169092443916798e-06, + "loss": 0.4725, + "step": 2634 + }, + { + "epoch": 0.48277757420300477, + "grad_norm": 0.47826461162392464, + "learning_rate": 6.165982974012104e-06, + "loss": 0.4922, + "step": 2635 + }, + { + "epoch": 0.48296079149871746, + "grad_norm": 0.4097799093815455, + "learning_rate": 6.16287302715481e-06, + "loss": 0.4435, + "step": 2636 + }, + { + "epoch": 0.4831440087944302, + "grad_norm": 0.49409501240471265, + "learning_rate": 6.159762604617054e-06, + "loss": 0.4511, + "step": 2637 + }, + { + "epoch": 0.4833272260901429, + "grad_norm": 0.32922185698090817, + "learning_rate": 6.156651707671179e-06, + "loss": 0.4431, + "step": 2638 + }, + { + "epoch": 0.48351044338585564, + "grad_norm": 0.43589998693776294, + "learning_rate": 6.15354033758971e-06, + "loss": 0.4913, + "step": 2639 + }, + { + "epoch": 0.4836936606815683, + "grad_norm": 0.46935630185597366, + "learning_rate": 6.15042849564538e-06, + "loss": 0.4438, + "step": 2640 + }, + { + "epoch": 0.4838768779772811, + "grad_norm": 0.35187408753902083, + "learning_rate": 6.147316183111101e-06, + "loss": 0.4757, + "step": 2641 + }, + { + "epoch": 0.48406009527299376, + "grad_norm": 0.3551224007772423, + "learning_rate": 6.144203401259987e-06, + "loss": 0.4857, + "step": 2642 + }, + { + "epoch": 0.4842433125687065, + "grad_norm": 0.4067611468464816, + "learning_rate": 6.141090151365341e-06, + "loss": 0.4594, + "step": 2643 + }, + { + "epoch": 0.4844265298644192, + "grad_norm": 0.38220030063601856, + "learning_rate": 6.137976434700656e-06, + "loss": 0.464, + "step": 2644 + }, + { + "epoch": 0.48460974716013194, + "grad_norm": 0.39667408717834457, + "learning_rate": 6.134862252539617e-06, + "loss": 0.4776, + "step": 2645 + }, + { + "epoch": 0.48479296445584463, + "grad_norm": 0.3980341085738827, + "learning_rate": 6.131747606156101e-06, + "loss": 0.495, + "step": 2646 + }, + { + "epoch": 0.4849761817515573, + "grad_norm": 0.3527503512146692, + "learning_rate": 6.128632496824173e-06, + "loss": 0.4817, + "step": 2647 + }, + { + "epoch": 0.48515939904727007, + "grad_norm": 0.5523393433703138, + "learning_rate": 6.125516925818087e-06, + "loss": 0.4776, + "step": 2648 + }, + { + "epoch": 0.48534261634298276, + "grad_norm": 0.45560350902981445, + "learning_rate": 6.122400894412288e-06, + "loss": 0.5108, + "step": 2649 + }, + { + "epoch": 0.4855258336386955, + "grad_norm": 0.3883613637399946, + "learning_rate": 6.119284403881407e-06, + "loss": 0.505, + "step": 2650 + }, + { + "epoch": 0.4857090509344082, + "grad_norm": 0.43054839957129043, + "learning_rate": 6.116167455500265e-06, + "loss": 0.4613, + "step": 2651 + }, + { + "epoch": 0.48589226823012094, + "grad_norm": 0.44699283379261234, + "learning_rate": 6.113050050543869e-06, + "loss": 0.5095, + "step": 2652 + }, + { + "epoch": 0.48607548552583363, + "grad_norm": 0.4253525400492399, + "learning_rate": 6.109932190287411e-06, + "loss": 0.4355, + "step": 2653 + }, + { + "epoch": 0.4862587028215464, + "grad_norm": 0.3789592185972384, + "learning_rate": 6.106813876006274e-06, + "loss": 0.5028, + "step": 2654 + }, + { + "epoch": 0.48644192011725906, + "grad_norm": 0.593473175364269, + "learning_rate": 6.103695108976022e-06, + "loss": 0.4746, + "step": 2655 + }, + { + "epoch": 0.4866251374129718, + "grad_norm": 0.3722676120625255, + "learning_rate": 6.1005758904724065e-06, + "loss": 0.4417, + "step": 2656 + }, + { + "epoch": 0.4868083547086845, + "grad_norm": 0.4217594770549542, + "learning_rate": 6.097456221771364e-06, + "loss": 0.4748, + "step": 2657 + }, + { + "epoch": 0.4869915720043972, + "grad_norm": 0.3941559337345006, + "learning_rate": 6.094336104149014e-06, + "loss": 0.4832, + "step": 2658 + }, + { + "epoch": 0.48717478930010993, + "grad_norm": 0.3517468727202208, + "learning_rate": 6.0912155388816584e-06, + "loss": 0.4724, + "step": 2659 + }, + { + "epoch": 0.4873580065958226, + "grad_norm": 0.432479131802833, + "learning_rate": 6.088094527245788e-06, + "loss": 0.4693, + "step": 2660 + }, + { + "epoch": 0.48754122389153537, + "grad_norm": 0.3642359211792819, + "learning_rate": 6.084973070518067e-06, + "loss": 0.4601, + "step": 2661 + }, + { + "epoch": 0.48772444118724806, + "grad_norm": 0.376126812545194, + "learning_rate": 6.081851169975351e-06, + "loss": 0.4758, + "step": 2662 + }, + { + "epoch": 0.4879076584829608, + "grad_norm": 0.41480118650940473, + "learning_rate": 6.0787288268946685e-06, + "loss": 0.4562, + "step": 2663 + }, + { + "epoch": 0.4880908757786735, + "grad_norm": 0.3677659442377228, + "learning_rate": 6.075606042553239e-06, + "loss": 0.4693, + "step": 2664 + }, + { + "epoch": 0.48827409307438624, + "grad_norm": 0.3983674103639988, + "learning_rate": 6.072482818228452e-06, + "loss": 0.4792, + "step": 2665 + }, + { + "epoch": 0.48845731037009893, + "grad_norm": 0.3768784933428776, + "learning_rate": 6.069359155197885e-06, + "loss": 0.478, + "step": 2666 + }, + { + "epoch": 0.4886405276658117, + "grad_norm": 0.3817763408127909, + "learning_rate": 6.066235054739289e-06, + "loss": 0.4798, + "step": 2667 + }, + { + "epoch": 0.48882374496152436, + "grad_norm": 0.4596770357312781, + "learning_rate": 6.0631105181306e-06, + "loss": 0.4813, + "step": 2668 + }, + { + "epoch": 0.4890069622572371, + "grad_norm": 0.39989028831569035, + "learning_rate": 6.059985546649926e-06, + "loss": 0.474, + "step": 2669 + }, + { + "epoch": 0.4891901795529498, + "grad_norm": 0.38400457775519437, + "learning_rate": 6.056860141575558e-06, + "loss": 0.4357, + "step": 2670 + }, + { + "epoch": 0.4893733968486625, + "grad_norm": 0.3559471525714825, + "learning_rate": 6.053734304185962e-06, + "loss": 0.4912, + "step": 2671 + }, + { + "epoch": 0.48955661414437524, + "grad_norm": 0.3817984688569835, + "learning_rate": 6.050608035759781e-06, + "loss": 0.4769, + "step": 2672 + }, + { + "epoch": 0.4897398314400879, + "grad_norm": 0.39341219022919127, + "learning_rate": 6.047481337575835e-06, + "loss": 0.4528, + "step": 2673 + }, + { + "epoch": 0.48992304873580067, + "grad_norm": 0.40031286428345547, + "learning_rate": 6.044354210913119e-06, + "loss": 0.446, + "step": 2674 + }, + { + "epoch": 0.49010626603151336, + "grad_norm": 0.4805269758521716, + "learning_rate": 6.041226657050804e-06, + "loss": 0.483, + "step": 2675 + }, + { + "epoch": 0.4902894833272261, + "grad_norm": 0.35241515587778094, + "learning_rate": 6.038098677268236e-06, + "loss": 0.4826, + "step": 2676 + }, + { + "epoch": 0.4904727006229388, + "grad_norm": 0.40902285639832336, + "learning_rate": 6.0349702728449336e-06, + "loss": 0.4825, + "step": 2677 + }, + { + "epoch": 0.49065591791865154, + "grad_norm": 0.3855180959903487, + "learning_rate": 6.031841445060589e-06, + "loss": 0.4796, + "step": 2678 + }, + { + "epoch": 0.49083913521436423, + "grad_norm": 0.41876945969464213, + "learning_rate": 6.028712195195071e-06, + "loss": 0.4484, + "step": 2679 + }, + { + "epoch": 0.491022352510077, + "grad_norm": 0.34618598423601504, + "learning_rate": 6.025582524528418e-06, + "loss": 0.44, + "step": 2680 + }, + { + "epoch": 0.49120556980578967, + "grad_norm": 0.3917064661677189, + "learning_rate": 6.022452434340842e-06, + "loss": 0.4867, + "step": 2681 + }, + { + "epoch": 0.49138878710150236, + "grad_norm": 0.3901163153298345, + "learning_rate": 6.0193219259127245e-06, + "loss": 0.484, + "step": 2682 + }, + { + "epoch": 0.4915720043972151, + "grad_norm": 0.4074770547401102, + "learning_rate": 6.01619100052462e-06, + "loss": 0.4887, + "step": 2683 + }, + { + "epoch": 0.4917552216929278, + "grad_norm": 0.35725319337842953, + "learning_rate": 6.013059659457253e-06, + "loss": 0.4302, + "step": 2684 + }, + { + "epoch": 0.49193843898864054, + "grad_norm": 0.40406672516136255, + "learning_rate": 6.009927903991518e-06, + "loss": 0.5049, + "step": 2685 + }, + { + "epoch": 0.4921216562843532, + "grad_norm": 0.3765894390477666, + "learning_rate": 6.00679573540848e-06, + "loss": 0.484, + "step": 2686 + }, + { + "epoch": 0.49230487358006597, + "grad_norm": 0.3327835502977535, + "learning_rate": 6.00366315498937e-06, + "loss": 0.4628, + "step": 2687 + }, + { + "epoch": 0.49248809087577866, + "grad_norm": 0.4431385688448901, + "learning_rate": 6.000530164015592e-06, + "loss": 0.5068, + "step": 2688 + }, + { + "epoch": 0.4926713081714914, + "grad_norm": 0.38114263629098655, + "learning_rate": 5.9973967637687135e-06, + "loss": 0.4672, + "step": 2689 + }, + { + "epoch": 0.4928545254672041, + "grad_norm": 0.43292994075216573, + "learning_rate": 5.994262955530475e-06, + "loss": 0.4972, + "step": 2690 + }, + { + "epoch": 0.49303774276291684, + "grad_norm": 0.3818174127734471, + "learning_rate": 5.991128740582774e-06, + "loss": 0.4694, + "step": 2691 + }, + { + "epoch": 0.49322096005862953, + "grad_norm": 0.41472996823846253, + "learning_rate": 5.9879941202076865e-06, + "loss": 0.456, + "step": 2692 + }, + { + "epoch": 0.4934041773543423, + "grad_norm": 0.32357571760477655, + "learning_rate": 5.984859095687444e-06, + "loss": 0.4484, + "step": 2693 + }, + { + "epoch": 0.49358739465005497, + "grad_norm": 0.4180795286451674, + "learning_rate": 5.981723668304453e-06, + "loss": 0.462, + "step": 2694 + }, + { + "epoch": 0.49377061194576766, + "grad_norm": 0.3312791966667695, + "learning_rate": 5.978587839341274e-06, + "loss": 0.482, + "step": 2695 + }, + { + "epoch": 0.4939538292414804, + "grad_norm": 0.4174870043559363, + "learning_rate": 5.975451610080643e-06, + "loss": 0.4796, + "step": 2696 + }, + { + "epoch": 0.4941370465371931, + "grad_norm": 0.405864500614883, + "learning_rate": 5.972314981805448e-06, + "loss": 0.4962, + "step": 2697 + }, + { + "epoch": 0.49432026383290584, + "grad_norm": 0.3787964556422184, + "learning_rate": 5.969177955798751e-06, + "loss": 0.5045, + "step": 2698 + }, + { + "epoch": 0.4945034811286185, + "grad_norm": 0.374387983539771, + "learning_rate": 5.966040533343772e-06, + "loss": 0.4357, + "step": 2699 + }, + { + "epoch": 0.4946866984243313, + "grad_norm": 0.42310189437618834, + "learning_rate": 5.9629027157238895e-06, + "loss": 0.4531, + "step": 2700 + }, + { + "epoch": 0.49486991572004396, + "grad_norm": 0.39268642068645015, + "learning_rate": 5.959764504222651e-06, + "loss": 0.484, + "step": 2701 + }, + { + "epoch": 0.4950531330157567, + "grad_norm": 0.3940895472838365, + "learning_rate": 5.956625900123758e-06, + "loss": 0.4855, + "step": 2702 + }, + { + "epoch": 0.4952363503114694, + "grad_norm": 0.3342437830142438, + "learning_rate": 5.953486904711081e-06, + "loss": 0.4803, + "step": 2703 + }, + { + "epoch": 0.49541956760718214, + "grad_norm": 0.4299079154017634, + "learning_rate": 5.950347519268641e-06, + "loss": 0.4817, + "step": 2704 + }, + { + "epoch": 0.49560278490289483, + "grad_norm": 0.42871763235680505, + "learning_rate": 5.9472077450806266e-06, + "loss": 0.514, + "step": 2705 + }, + { + "epoch": 0.4957860021986075, + "grad_norm": 0.38952290407151696, + "learning_rate": 5.944067583431378e-06, + "loss": 0.4924, + "step": 2706 + }, + { + "epoch": 0.49596921949432027, + "grad_norm": 0.3444838737772104, + "learning_rate": 5.9409270356054025e-06, + "loss": 0.4594, + "step": 2707 + }, + { + "epoch": 0.49615243679003296, + "grad_norm": 0.36870720783408434, + "learning_rate": 5.937786102887356e-06, + "loss": 0.4656, + "step": 2708 + }, + { + "epoch": 0.4963356540857457, + "grad_norm": 0.3875788613728403, + "learning_rate": 5.934644786562063e-06, + "loss": 0.494, + "step": 2709 + }, + { + "epoch": 0.4965188713814584, + "grad_norm": 0.4067947100771364, + "learning_rate": 5.931503087914492e-06, + "loss": 0.5116, + "step": 2710 + }, + { + "epoch": 0.49670208867717114, + "grad_norm": 0.4164396398124012, + "learning_rate": 5.928361008229777e-06, + "loss": 0.4681, + "step": 2711 + }, + { + "epoch": 0.49688530597288383, + "grad_norm": 0.348557867806117, + "learning_rate": 5.925218548793206e-06, + "loss": 0.4647, + "step": 2712 + }, + { + "epoch": 0.4970685232685966, + "grad_norm": 0.32609070782662797, + "learning_rate": 5.922075710890221e-06, + "loss": 0.4728, + "step": 2713 + }, + { + "epoch": 0.49725174056430926, + "grad_norm": 0.43929032188456524, + "learning_rate": 5.9189324958064194e-06, + "loss": 0.5011, + "step": 2714 + }, + { + "epoch": 0.497434957860022, + "grad_norm": 0.3701748243514993, + "learning_rate": 5.915788904827553e-06, + "loss": 0.4458, + "step": 2715 + }, + { + "epoch": 0.4976181751557347, + "grad_norm": 0.3577823672245004, + "learning_rate": 5.912644939239527e-06, + "loss": 0.4772, + "step": 2716 + }, + { + "epoch": 0.49780139245144744, + "grad_norm": 0.35640413685913674, + "learning_rate": 5.9095006003284e-06, + "loss": 0.4668, + "step": 2717 + }, + { + "epoch": 0.49798460974716013, + "grad_norm": 0.3678128609463679, + "learning_rate": 5.906355889380384e-06, + "loss": 0.466, + "step": 2718 + }, + { + "epoch": 0.4981678270428728, + "grad_norm": 0.37342052972273326, + "learning_rate": 5.9032108076818425e-06, + "loss": 0.4489, + "step": 2719 + }, + { + "epoch": 0.49835104433858557, + "grad_norm": 0.37004531482695346, + "learning_rate": 5.900065356519291e-06, + "loss": 0.4589, + "step": 2720 + }, + { + "epoch": 0.49853426163429826, + "grad_norm": 0.40637846150182877, + "learning_rate": 5.896919537179395e-06, + "loss": 0.4614, + "step": 2721 + }, + { + "epoch": 0.498717478930011, + "grad_norm": 0.4255331589630329, + "learning_rate": 5.893773350948974e-06, + "loss": 0.4797, + "step": 2722 + }, + { + "epoch": 0.4989006962257237, + "grad_norm": 0.36989682428254933, + "learning_rate": 5.890626799114991e-06, + "loss": 0.4785, + "step": 2723 + }, + { + "epoch": 0.49908391352143644, + "grad_norm": 0.37360175533264606, + "learning_rate": 5.887479882964565e-06, + "loss": 0.5131, + "step": 2724 + }, + { + "epoch": 0.49926713081714913, + "grad_norm": 0.44374802226129506, + "learning_rate": 5.884332603784961e-06, + "loss": 0.5174, + "step": 2725 + }, + { + "epoch": 0.4994503481128619, + "grad_norm": 0.3596798859511765, + "learning_rate": 5.881184962863595e-06, + "loss": 0.5339, + "step": 2726 + }, + { + "epoch": 0.49963356540857456, + "grad_norm": 0.3848471416284064, + "learning_rate": 5.878036961488028e-06, + "loss": 0.4748, + "step": 2727 + }, + { + "epoch": 0.4998167827042873, + "grad_norm": 0.3393343027009829, + "learning_rate": 5.8748886009459685e-06, + "loss": 0.4824, + "step": 2728 + }, + { + "epoch": 0.5, + "grad_norm": 0.3855331962057024, + "learning_rate": 5.8717398825252744e-06, + "loss": 0.519, + "step": 2729 + }, + { + "epoch": 0.5001832172957127, + "grad_norm": 0.37456539622575846, + "learning_rate": 5.868590807513947e-06, + "loss": 0.4553, + "step": 2730 + }, + { + "epoch": 0.5003664345914254, + "grad_norm": 0.37935813236647553, + "learning_rate": 5.865441377200137e-06, + "loss": 0.4766, + "step": 2731 + }, + { + "epoch": 0.5005496518871382, + "grad_norm": 0.4449058873445844, + "learning_rate": 5.862291592872135e-06, + "loss": 0.5073, + "step": 2732 + }, + { + "epoch": 0.5007328691828509, + "grad_norm": 0.4083669791420912, + "learning_rate": 5.859141455818387e-06, + "loss": 0.4685, + "step": 2733 + }, + { + "epoch": 0.5009160864785636, + "grad_norm": 0.39542564333243985, + "learning_rate": 5.855990967327468e-06, + "loss": 0.48, + "step": 2734 + }, + { + "epoch": 0.5010993037742763, + "grad_norm": 0.3647775918005535, + "learning_rate": 5.852840128688112e-06, + "loss": 0.4814, + "step": 2735 + }, + { + "epoch": 0.501282521069989, + "grad_norm": 0.3621924996992885, + "learning_rate": 5.849688941189184e-06, + "loss": 0.4884, + "step": 2736 + }, + { + "epoch": 0.5014657383657017, + "grad_norm": 0.43759234375668876, + "learning_rate": 5.846537406119699e-06, + "loss": 0.4929, + "step": 2737 + }, + { + "epoch": 0.5016489556614144, + "grad_norm": 0.3793853555550419, + "learning_rate": 5.8433855247688106e-06, + "loss": 0.4841, + "step": 2738 + }, + { + "epoch": 0.5018321729571271, + "grad_norm": 0.37854631411105866, + "learning_rate": 5.840233298425818e-06, + "loss": 0.4875, + "step": 2739 + }, + { + "epoch": 0.5020153902528399, + "grad_norm": 0.3674096820378392, + "learning_rate": 5.837080728380157e-06, + "loss": 0.4662, + "step": 2740 + }, + { + "epoch": 0.5021986075485526, + "grad_norm": 0.387465729509014, + "learning_rate": 5.8339278159214075e-06, + "loss": 0.4774, + "step": 2741 + }, + { + "epoch": 0.5023818248442653, + "grad_norm": 0.36939648216432863, + "learning_rate": 5.830774562339287e-06, + "loss": 0.4896, + "step": 2742 + }, + { + "epoch": 0.502565042139978, + "grad_norm": 0.38599335557911074, + "learning_rate": 5.827620968923653e-06, + "loss": 0.4846, + "step": 2743 + }, + { + "epoch": 0.5027482594356907, + "grad_norm": 0.34835327441068153, + "learning_rate": 5.824467036964504e-06, + "loss": 0.496, + "step": 2744 + }, + { + "epoch": 0.5029314767314035, + "grad_norm": 0.3975274865194216, + "learning_rate": 5.821312767751975e-06, + "loss": 0.4867, + "step": 2745 + }, + { + "epoch": 0.5031146940271162, + "grad_norm": 0.3780583868796423, + "learning_rate": 5.818158162576339e-06, + "loss": 0.4785, + "step": 2746 + }, + { + "epoch": 0.5032979113228289, + "grad_norm": 0.3249091810860243, + "learning_rate": 5.815003222728007e-06, + "loss": 0.4668, + "step": 2747 + }, + { + "epoch": 0.5034811286185416, + "grad_norm": 0.37469778914009816, + "learning_rate": 5.811847949497529e-06, + "loss": 0.4511, + "step": 2748 + }, + { + "epoch": 0.5036643459142544, + "grad_norm": 0.4129177088016805, + "learning_rate": 5.808692344175585e-06, + "loss": 0.4563, + "step": 2749 + }, + { + "epoch": 0.503847563209967, + "grad_norm": 0.3925198855785553, + "learning_rate": 5.805536408052998e-06, + "loss": 0.4727, + "step": 2750 + }, + { + "epoch": 0.5040307805056797, + "grad_norm": 0.38023027041798463, + "learning_rate": 5.802380142420723e-06, + "loss": 0.5039, + "step": 2751 + }, + { + "epoch": 0.5042139978013924, + "grad_norm": 0.40626816211410266, + "learning_rate": 5.799223548569849e-06, + "loss": 0.4854, + "step": 2752 + }, + { + "epoch": 0.5043972150971051, + "grad_norm": 0.4288318784024032, + "learning_rate": 5.796066627791602e-06, + "loss": 0.4691, + "step": 2753 + }, + { + "epoch": 0.5045804323928179, + "grad_norm": 0.4088544079991914, + "learning_rate": 5.792909381377339e-06, + "loss": 0.5136, + "step": 2754 + }, + { + "epoch": 0.5047636496885306, + "grad_norm": 0.4262461935824578, + "learning_rate": 5.789751810618551e-06, + "loss": 0.4631, + "step": 2755 + }, + { + "epoch": 0.5049468669842433, + "grad_norm": 0.36202131336232085, + "learning_rate": 5.786593916806862e-06, + "loss": 0.447, + "step": 2756 + }, + { + "epoch": 0.505130084279956, + "grad_norm": 0.4396611051494124, + "learning_rate": 5.78343570123403e-06, + "loss": 0.4834, + "step": 2757 + }, + { + "epoch": 0.5053133015756688, + "grad_norm": 0.37939224767002205, + "learning_rate": 5.780277165191941e-06, + "loss": 0.4639, + "step": 2758 + }, + { + "epoch": 0.5054965188713815, + "grad_norm": 1.1177794213599237, + "learning_rate": 5.7771183099726144e-06, + "loss": 0.4833, + "step": 2759 + }, + { + "epoch": 0.5056797361670942, + "grad_norm": 0.40826440312948875, + "learning_rate": 5.773959136868199e-06, + "loss": 0.4681, + "step": 2760 + }, + { + "epoch": 0.5058629534628069, + "grad_norm": 0.39682165370729394, + "learning_rate": 5.770799647170976e-06, + "loss": 0.4572, + "step": 2761 + }, + { + "epoch": 0.5060461707585197, + "grad_norm": 0.3856723957716629, + "learning_rate": 5.767639842173351e-06, + "loss": 0.4424, + "step": 2762 + }, + { + "epoch": 0.5062293880542323, + "grad_norm": 0.35149307259572515, + "learning_rate": 5.764479723167867e-06, + "loss": 0.484, + "step": 2763 + }, + { + "epoch": 0.506412605349945, + "grad_norm": 0.36966173779312345, + "learning_rate": 5.761319291447186e-06, + "loss": 0.513, + "step": 2764 + }, + { + "epoch": 0.5065958226456577, + "grad_norm": 0.3622945326783817, + "learning_rate": 5.758158548304104e-06, + "loss": 0.4545, + "step": 2765 + }, + { + "epoch": 0.5067790399413704, + "grad_norm": 0.3789511857632013, + "learning_rate": 5.7549974950315425e-06, + "loss": 0.4433, + "step": 2766 + }, + { + "epoch": 0.5069622572370832, + "grad_norm": 0.3961364474839944, + "learning_rate": 5.751836132922549e-06, + "loss": 0.4668, + "step": 2767 + }, + { + "epoch": 0.5071454745327959, + "grad_norm": 0.3620008337212967, + "learning_rate": 5.748674463270301e-06, + "loss": 0.4244, + "step": 2768 + }, + { + "epoch": 0.5073286918285086, + "grad_norm": 0.36550069181670775, + "learning_rate": 5.7455124873680955e-06, + "loss": 0.4714, + "step": 2769 + }, + { + "epoch": 0.5075119091242213, + "grad_norm": 0.3960451745822164, + "learning_rate": 5.742350206509362e-06, + "loss": 0.4792, + "step": 2770 + }, + { + "epoch": 0.5076951264199341, + "grad_norm": 0.3990763480918848, + "learning_rate": 5.739187621987649e-06, + "loss": 0.4758, + "step": 2771 + }, + { + "epoch": 0.5078783437156468, + "grad_norm": 0.38440435932890443, + "learning_rate": 5.7360247350966315e-06, + "loss": 0.4831, + "step": 2772 + }, + { + "epoch": 0.5080615610113595, + "grad_norm": 0.36287352658219973, + "learning_rate": 5.7328615471301095e-06, + "loss": 0.4541, + "step": 2773 + }, + { + "epoch": 0.5082447783070722, + "grad_norm": 0.44219907061604324, + "learning_rate": 5.729698059382004e-06, + "loss": 0.4813, + "step": 2774 + }, + { + "epoch": 0.508427995602785, + "grad_norm": 0.3698989955692163, + "learning_rate": 5.726534273146359e-06, + "loss": 0.4965, + "step": 2775 + }, + { + "epoch": 0.5086112128984976, + "grad_norm": 0.3712565895036428, + "learning_rate": 5.723370189717341e-06, + "loss": 0.4601, + "step": 2776 + }, + { + "epoch": 0.5087944301942103, + "grad_norm": 0.3870945674162369, + "learning_rate": 5.720205810389239e-06, + "loss": 0.4529, + "step": 2777 + }, + { + "epoch": 0.508977647489923, + "grad_norm": 0.40711078350416036, + "learning_rate": 5.717041136456462e-06, + "loss": 0.4861, + "step": 2778 + }, + { + "epoch": 0.5091608647856357, + "grad_norm": 0.44393662476336315, + "learning_rate": 5.713876169213538e-06, + "loss": 0.4799, + "step": 2779 + }, + { + "epoch": 0.5093440820813485, + "grad_norm": 0.3649087342670491, + "learning_rate": 5.710710909955118e-06, + "loss": 0.4638, + "step": 2780 + }, + { + "epoch": 0.5095272993770612, + "grad_norm": 0.44305408078324465, + "learning_rate": 5.70754535997597e-06, + "loss": 0.4905, + "step": 2781 + }, + { + "epoch": 0.5097105166727739, + "grad_norm": 0.44710511602061487, + "learning_rate": 5.704379520570982e-06, + "loss": 0.4656, + "step": 2782 + }, + { + "epoch": 0.5098937339684866, + "grad_norm": 0.3768898212593808, + "learning_rate": 5.701213393035163e-06, + "loss": 0.483, + "step": 2783 + }, + { + "epoch": 0.5100769512641994, + "grad_norm": 0.3956197731638185, + "learning_rate": 5.698046978663633e-06, + "loss": 0.5064, + "step": 2784 + }, + { + "epoch": 0.5102601685599121, + "grad_norm": 0.38359602066204457, + "learning_rate": 5.694880278751638e-06, + "loss": 0.4723, + "step": 2785 + }, + { + "epoch": 0.5104433858556248, + "grad_norm": 0.36821179651875974, + "learning_rate": 5.691713294594531e-06, + "loss": 0.4531, + "step": 2786 + }, + { + "epoch": 0.5106266031513375, + "grad_norm": 0.33986653473636935, + "learning_rate": 5.688546027487793e-06, + "loss": 0.4965, + "step": 2787 + }, + { + "epoch": 0.5108098204470503, + "grad_norm": 0.3829022120852112, + "learning_rate": 5.685378478727007e-06, + "loss": 0.4797, + "step": 2788 + }, + { + "epoch": 0.510993037742763, + "grad_norm": 0.3251353454694897, + "learning_rate": 5.682210649607886e-06, + "loss": 0.479, + "step": 2789 + }, + { + "epoch": 0.5111762550384756, + "grad_norm": 0.3903981320240099, + "learning_rate": 5.679042541426245e-06, + "loss": 0.4591, + "step": 2790 + }, + { + "epoch": 0.5113594723341883, + "grad_norm": 0.38367188390506624, + "learning_rate": 5.67587415547802e-06, + "loss": 0.5027, + "step": 2791 + }, + { + "epoch": 0.511542689629901, + "grad_norm": 0.43830841848822, + "learning_rate": 5.67270549305926e-06, + "loss": 0.5046, + "step": 2792 + }, + { + "epoch": 0.5117259069256138, + "grad_norm": 0.3309237014818497, + "learning_rate": 5.669536555466125e-06, + "loss": 0.482, + "step": 2793 + }, + { + "epoch": 0.5119091242213265, + "grad_norm": 0.35608783511110903, + "learning_rate": 5.66636734399489e-06, + "loss": 0.4805, + "step": 2794 + }, + { + "epoch": 0.5120923415170392, + "grad_norm": 0.381478963173924, + "learning_rate": 5.663197859941938e-06, + "loss": 0.4544, + "step": 2795 + }, + { + "epoch": 0.5122755588127519, + "grad_norm": 0.39595976553551976, + "learning_rate": 5.660028104603769e-06, + "loss": 0.4743, + "step": 2796 + }, + { + "epoch": 0.5124587761084647, + "grad_norm": 0.40867905676443567, + "learning_rate": 5.656858079276989e-06, + "loss": 0.5073, + "step": 2797 + }, + { + "epoch": 0.5126419934041774, + "grad_norm": 0.3527369861682892, + "learning_rate": 5.653687785258319e-06, + "loss": 0.4539, + "step": 2798 + }, + { + "epoch": 0.5128252106998901, + "grad_norm": 0.38219313964397705, + "learning_rate": 5.650517223844585e-06, + "loss": 0.4541, + "step": 2799 + }, + { + "epoch": 0.5130084279956028, + "grad_norm": 0.38386615301297883, + "learning_rate": 5.647346396332726e-06, + "loss": 0.4874, + "step": 2800 + }, + { + "epoch": 0.5131916452913154, + "grad_norm": 0.3726417029798126, + "learning_rate": 5.644175304019789e-06, + "loss": 0.4785, + "step": 2801 + }, + { + "epoch": 0.5133748625870282, + "grad_norm": 0.3628897592027456, + "learning_rate": 5.641003948202929e-06, + "loss": 0.4773, + "step": 2802 + }, + { + "epoch": 0.5135580798827409, + "grad_norm": 0.3335307671309961, + "learning_rate": 5.637832330179409e-06, + "loss": 0.4529, + "step": 2803 + }, + { + "epoch": 0.5137412971784536, + "grad_norm": 0.44601247780937237, + "learning_rate": 5.6346604512465995e-06, + "loss": 0.4816, + "step": 2804 + }, + { + "epoch": 0.5139245144741663, + "grad_norm": 0.34863991785960624, + "learning_rate": 5.631488312701977e-06, + "loss": 0.4951, + "step": 2805 + }, + { + "epoch": 0.5141077317698791, + "grad_norm": 0.3820766793490931, + "learning_rate": 5.628315915843122e-06, + "loss": 0.4467, + "step": 2806 + }, + { + "epoch": 0.5142909490655918, + "grad_norm": 0.35635264125047506, + "learning_rate": 5.625143261967728e-06, + "loss": 0.4566, + "step": 2807 + }, + { + "epoch": 0.5144741663613045, + "grad_norm": 0.4711996641872704, + "learning_rate": 5.6219703523735845e-06, + "loss": 0.4713, + "step": 2808 + }, + { + "epoch": 0.5146573836570172, + "grad_norm": 0.41042137308045834, + "learning_rate": 5.618797188358594e-06, + "loss": 0.491, + "step": 2809 + }, + { + "epoch": 0.51484060095273, + "grad_norm": 0.41018240050619975, + "learning_rate": 5.615623771220754e-06, + "loss": 0.4512, + "step": 2810 + }, + { + "epoch": 0.5150238182484427, + "grad_norm": 0.4072188457659445, + "learning_rate": 5.612450102258175e-06, + "loss": 0.458, + "step": 2811 + }, + { + "epoch": 0.5152070355441554, + "grad_norm": 0.3717199084066965, + "learning_rate": 5.6092761827690644e-06, + "loss": 0.4542, + "step": 2812 + }, + { + "epoch": 0.5153902528398681, + "grad_norm": 0.4112503419617134, + "learning_rate": 5.6061020140517345e-06, + "loss": 0.4796, + "step": 2813 + }, + { + "epoch": 0.5155734701355807, + "grad_norm": 0.34348081168540084, + "learning_rate": 5.602927597404596e-06, + "loss": 0.4676, + "step": 2814 + }, + { + "epoch": 0.5157566874312935, + "grad_norm": 0.4147336886926842, + "learning_rate": 5.599752934126168e-06, + "loss": 0.4841, + "step": 2815 + }, + { + "epoch": 0.5159399047270062, + "grad_norm": 0.3847769496581143, + "learning_rate": 5.596578025515064e-06, + "loss": 0.4909, + "step": 2816 + }, + { + "epoch": 0.5161231220227189, + "grad_norm": 0.38263632939989156, + "learning_rate": 5.593402872869999e-06, + "loss": 0.4978, + "step": 2817 + }, + { + "epoch": 0.5163063393184316, + "grad_norm": 0.3675389064146458, + "learning_rate": 5.590227477489792e-06, + "loss": 0.4819, + "step": 2818 + }, + { + "epoch": 0.5164895566141444, + "grad_norm": 0.36445274827340074, + "learning_rate": 5.587051840673356e-06, + "loss": 0.4664, + "step": 2819 + }, + { + "epoch": 0.5166727739098571, + "grad_norm": 0.35215155428207906, + "learning_rate": 5.583875963719704e-06, + "loss": 0.4566, + "step": 2820 + }, + { + "epoch": 0.5168559912055698, + "grad_norm": 0.3655726933533943, + "learning_rate": 5.5806998479279515e-06, + "loss": 0.4371, + "step": 2821 + }, + { + "epoch": 0.5170392085012825, + "grad_norm": 0.34665549819074337, + "learning_rate": 5.577523494597307e-06, + "loss": 0.4625, + "step": 2822 + }, + { + "epoch": 0.5172224257969953, + "grad_norm": 0.397657981358375, + "learning_rate": 5.574346905027076e-06, + "loss": 0.5019, + "step": 2823 + }, + { + "epoch": 0.517405643092708, + "grad_norm": 0.3741013564162191, + "learning_rate": 5.5711700805166655e-06, + "loss": 0.488, + "step": 2824 + }, + { + "epoch": 0.5175888603884207, + "grad_norm": 0.34813475856931037, + "learning_rate": 5.567993022365572e-06, + "loss": 0.4628, + "step": 2825 + }, + { + "epoch": 0.5177720776841334, + "grad_norm": 0.3581306069169864, + "learning_rate": 5.564815731873394e-06, + "loss": 0.4716, + "step": 2826 + }, + { + "epoch": 0.517955294979846, + "grad_norm": 0.3521139375610018, + "learning_rate": 5.56163821033982e-06, + "loss": 0.4514, + "step": 2827 + }, + { + "epoch": 0.5181385122755588, + "grad_norm": 0.4246593635230546, + "learning_rate": 5.5584604590646354e-06, + "loss": 0.4728, + "step": 2828 + }, + { + "epoch": 0.5183217295712715, + "grad_norm": 0.5188493778299671, + "learning_rate": 5.55528247934772e-06, + "loss": 0.5149, + "step": 2829 + }, + { + "epoch": 0.5185049468669842, + "grad_norm": 0.4619175556512451, + "learning_rate": 5.552104272489045e-06, + "loss": 0.4794, + "step": 2830 + }, + { + "epoch": 0.5186881641626969, + "grad_norm": 0.3772249039736225, + "learning_rate": 5.5489258397886755e-06, + "loss": 0.4896, + "step": 2831 + }, + { + "epoch": 0.5188713814584097, + "grad_norm": 0.356953463396354, + "learning_rate": 5.545747182546771e-06, + "loss": 0.476, + "step": 2832 + }, + { + "epoch": 0.5190545987541224, + "grad_norm": 0.4327529150590389, + "learning_rate": 5.5425683020635805e-06, + "loss": 0.4616, + "step": 2833 + }, + { + "epoch": 0.5192378160498351, + "grad_norm": 0.34952783052364766, + "learning_rate": 5.539389199639442e-06, + "loss": 0.4503, + "step": 2834 + }, + { + "epoch": 0.5194210333455478, + "grad_norm": 0.429602823175531, + "learning_rate": 5.536209876574793e-06, + "loss": 0.4762, + "step": 2835 + }, + { + "epoch": 0.5196042506412606, + "grad_norm": 0.34183231295304567, + "learning_rate": 5.533030334170149e-06, + "loss": 0.4573, + "step": 2836 + }, + { + "epoch": 0.5197874679369733, + "grad_norm": 0.3486485931520749, + "learning_rate": 5.529850573726125e-06, + "loss": 0.4627, + "step": 2837 + }, + { + "epoch": 0.519970685232686, + "grad_norm": 0.3806206915460022, + "learning_rate": 5.526670596543422e-06, + "loss": 0.4701, + "step": 2838 + }, + { + "epoch": 0.5201539025283987, + "grad_norm": 0.31792940125070873, + "learning_rate": 5.52349040392283e-06, + "loss": 0.4374, + "step": 2839 + }, + { + "epoch": 0.5203371198241113, + "grad_norm": 0.4371454649249866, + "learning_rate": 5.520309997165223e-06, + "loss": 0.4929, + "step": 2840 + }, + { + "epoch": 0.5205203371198242, + "grad_norm": 0.371616242783777, + "learning_rate": 5.5171293775715704e-06, + "loss": 0.5076, + "step": 2841 + }, + { + "epoch": 0.5207035544155368, + "grad_norm": 0.5396610760886152, + "learning_rate": 5.513948546442923e-06, + "loss": 0.4951, + "step": 2842 + }, + { + "epoch": 0.5208867717112495, + "grad_norm": 0.4120896853660915, + "learning_rate": 5.510767505080419e-06, + "loss": 0.4747, + "step": 2843 + }, + { + "epoch": 0.5210699890069622, + "grad_norm": 0.37452071665561426, + "learning_rate": 5.507586254785283e-06, + "loss": 0.4866, + "step": 2844 + }, + { + "epoch": 0.521253206302675, + "grad_norm": 0.37912360014591456, + "learning_rate": 5.504404796858826e-06, + "loss": 0.4611, + "step": 2845 + }, + { + "epoch": 0.5214364235983877, + "grad_norm": 0.3682103820427267, + "learning_rate": 5.501223132602443e-06, + "loss": 0.4855, + "step": 2846 + }, + { + "epoch": 0.5216196408941004, + "grad_norm": 0.361832494682015, + "learning_rate": 5.498041263317614e-06, + "loss": 0.4657, + "step": 2847 + }, + { + "epoch": 0.5218028581898131, + "grad_norm": 0.3838512388460457, + "learning_rate": 5.494859190305903e-06, + "loss": 0.4858, + "step": 2848 + }, + { + "epoch": 0.5219860754855258, + "grad_norm": 0.41392426005477884, + "learning_rate": 5.491676914868954e-06, + "loss": 0.4775, + "step": 2849 + }, + { + "epoch": 0.5221692927812386, + "grad_norm": 0.34136945446695127, + "learning_rate": 5.488494438308499e-06, + "loss": 0.4683, + "step": 2850 + }, + { + "epoch": 0.5223525100769513, + "grad_norm": 0.3724337718499766, + "learning_rate": 5.4853117619263496e-06, + "loss": 0.481, + "step": 2851 + }, + { + "epoch": 0.522535727372664, + "grad_norm": 0.36631446381516874, + "learning_rate": 5.4821288870243995e-06, + "loss": 0.4359, + "step": 2852 + }, + { + "epoch": 0.5227189446683767, + "grad_norm": 0.3860339163956676, + "learning_rate": 5.478945814904624e-06, + "loss": 0.4422, + "step": 2853 + }, + { + "epoch": 0.5229021619640895, + "grad_norm": 0.41470723211174776, + "learning_rate": 5.475762546869076e-06, + "loss": 0.4804, + "step": 2854 + }, + { + "epoch": 0.5230853792598021, + "grad_norm": 0.35952376817351184, + "learning_rate": 5.472579084219893e-06, + "loss": 0.4843, + "step": 2855 + }, + { + "epoch": 0.5232685965555148, + "grad_norm": 0.4284055124573125, + "learning_rate": 5.469395428259291e-06, + "loss": 0.4669, + "step": 2856 + }, + { + "epoch": 0.5234518138512275, + "grad_norm": 0.3950368583083783, + "learning_rate": 5.466211580289563e-06, + "loss": 0.4801, + "step": 2857 + }, + { + "epoch": 0.5236350311469403, + "grad_norm": 0.37109516802841763, + "learning_rate": 5.463027541613083e-06, + "loss": 0.477, + "step": 2858 + }, + { + "epoch": 0.523818248442653, + "grad_norm": 0.3304081968687978, + "learning_rate": 5.4598433135323015e-06, + "loss": 0.4698, + "step": 2859 + }, + { + "epoch": 0.5240014657383657, + "grad_norm": 0.3974618643677324, + "learning_rate": 5.456658897349745e-06, + "loss": 0.4936, + "step": 2860 + }, + { + "epoch": 0.5241846830340784, + "grad_norm": 0.3676858892604378, + "learning_rate": 5.453474294368022e-06, + "loss": 0.4753, + "step": 2861 + }, + { + "epoch": 0.5243679003297911, + "grad_norm": 0.4001836412967568, + "learning_rate": 5.450289505889811e-06, + "loss": 0.454, + "step": 2862 + }, + { + "epoch": 0.5245511176255039, + "grad_norm": 0.3948440862595113, + "learning_rate": 5.447104533217873e-06, + "loss": 0.4673, + "step": 2863 + }, + { + "epoch": 0.5247343349212166, + "grad_norm": 0.3621669622882833, + "learning_rate": 5.443919377655036e-06, + "loss": 0.5125, + "step": 2864 + }, + { + "epoch": 0.5249175522169293, + "grad_norm": 0.4124120444352978, + "learning_rate": 5.440734040504214e-06, + "loss": 0.4939, + "step": 2865 + }, + { + "epoch": 0.525100769512642, + "grad_norm": 0.35808462501030025, + "learning_rate": 5.437548523068382e-06, + "loss": 0.4932, + "step": 2866 + }, + { + "epoch": 0.5252839868083548, + "grad_norm": 0.39518159152536025, + "learning_rate": 5.434362826650603e-06, + "loss": 0.5003, + "step": 2867 + }, + { + "epoch": 0.5254672041040674, + "grad_norm": 0.35555153905366266, + "learning_rate": 5.431176952554001e-06, + "loss": 0.4908, + "step": 2868 + }, + { + "epoch": 0.5256504213997801, + "grad_norm": 0.3618384123537011, + "learning_rate": 5.42799090208178e-06, + "loss": 0.4279, + "step": 2869 + }, + { + "epoch": 0.5258336386954928, + "grad_norm": 0.3527157657304505, + "learning_rate": 5.424804676537212e-06, + "loss": 0.4921, + "step": 2870 + }, + { + "epoch": 0.5260168559912056, + "grad_norm": 0.4360747680680866, + "learning_rate": 5.421618277223644e-06, + "loss": 0.4681, + "step": 2871 + }, + { + "epoch": 0.5262000732869183, + "grad_norm": 0.3979795855422701, + "learning_rate": 5.418431705444491e-06, + "loss": 0.4913, + "step": 2872 + }, + { + "epoch": 0.526383290582631, + "grad_norm": 0.39940943008161817, + "learning_rate": 5.415244962503242e-06, + "loss": 0.4915, + "step": 2873 + }, + { + "epoch": 0.5265665078783437, + "grad_norm": 0.35211949742430754, + "learning_rate": 5.412058049703452e-06, + "loss": 0.4936, + "step": 2874 + }, + { + "epoch": 0.5267497251740564, + "grad_norm": 0.35634358762706225, + "learning_rate": 5.4088709683487494e-06, + "loss": 0.464, + "step": 2875 + }, + { + "epoch": 0.5269329424697692, + "grad_norm": 0.36997575334712507, + "learning_rate": 5.405683719742829e-06, + "loss": 0.4703, + "step": 2876 + }, + { + "epoch": 0.5271161597654819, + "grad_norm": 0.4157476241421092, + "learning_rate": 5.402496305189453e-06, + "loss": 0.504, + "step": 2877 + }, + { + "epoch": 0.5272993770611946, + "grad_norm": 0.38685559414741966, + "learning_rate": 5.399308725992456e-06, + "loss": 0.4814, + "step": 2878 + }, + { + "epoch": 0.5274825943569073, + "grad_norm": 0.3483494243303595, + "learning_rate": 5.396120983455736e-06, + "loss": 0.481, + "step": 2879 + }, + { + "epoch": 0.52766581165262, + "grad_norm": 0.36555168940379923, + "learning_rate": 5.3929330788832576e-06, + "loss": 0.4729, + "step": 2880 + }, + { + "epoch": 0.5278490289483327, + "grad_norm": 0.34899931353718594, + "learning_rate": 5.389745013579055e-06, + "loss": 0.4532, + "step": 2881 + }, + { + "epoch": 0.5280322462440454, + "grad_norm": 0.3639764900880931, + "learning_rate": 5.386556788847226e-06, + "loss": 0.4692, + "step": 2882 + }, + { + "epoch": 0.5282154635397581, + "grad_norm": 0.4661304297937794, + "learning_rate": 5.383368405991932e-06, + "loss": 0.5113, + "step": 2883 + }, + { + "epoch": 0.5283986808354708, + "grad_norm": 0.3601443720964451, + "learning_rate": 5.380179866317403e-06, + "loss": 0.4799, + "step": 2884 + }, + { + "epoch": 0.5285818981311836, + "grad_norm": 0.46626517384955435, + "learning_rate": 5.376991171127931e-06, + "loss": 0.469, + "step": 2885 + }, + { + "epoch": 0.5287651154268963, + "grad_norm": 0.37278046156442574, + "learning_rate": 5.373802321727868e-06, + "loss": 0.4873, + "step": 2886 + }, + { + "epoch": 0.528948332722609, + "grad_norm": 0.34859282489803667, + "learning_rate": 5.370613319421638e-06, + "loss": 0.4849, + "step": 2887 + }, + { + "epoch": 0.5291315500183217, + "grad_norm": 0.3517860655690429, + "learning_rate": 5.367424165513719e-06, + "loss": 0.4681, + "step": 2888 + }, + { + "epoch": 0.5293147673140345, + "grad_norm": 0.3795011301502316, + "learning_rate": 5.364234861308655e-06, + "loss": 0.4647, + "step": 2889 + }, + { + "epoch": 0.5294979846097472, + "grad_norm": 0.3422447727216203, + "learning_rate": 5.36104540811105e-06, + "loss": 0.4555, + "step": 2890 + }, + { + "epoch": 0.5296812019054599, + "grad_norm": 0.40073589742835186, + "learning_rate": 5.357855807225573e-06, + "loss": 0.4857, + "step": 2891 + }, + { + "epoch": 0.5298644192011726, + "grad_norm": 0.7193860431185534, + "learning_rate": 5.354666059956944e-06, + "loss": 0.4445, + "step": 2892 + }, + { + "epoch": 0.5300476364968854, + "grad_norm": 0.4186517280793463, + "learning_rate": 5.351476167609956e-06, + "loss": 0.482, + "step": 2893 + }, + { + "epoch": 0.530230853792598, + "grad_norm": 0.39548554303715294, + "learning_rate": 5.3482861314894485e-06, + "loss": 0.5013, + "step": 2894 + }, + { + "epoch": 0.5304140710883107, + "grad_norm": 0.35594961523045976, + "learning_rate": 5.3450959529003295e-06, + "loss": 0.4769, + "step": 2895 + }, + { + "epoch": 0.5305972883840234, + "grad_norm": 0.37538162589557583, + "learning_rate": 5.341905633147561e-06, + "loss": 0.5052, + "step": 2896 + }, + { + "epoch": 0.5307805056797361, + "grad_norm": 0.4247691508030218, + "learning_rate": 5.33871517353616e-06, + "loss": 0.5116, + "step": 2897 + }, + { + "epoch": 0.5309637229754489, + "grad_norm": 0.4114926709621181, + "learning_rate": 5.335524575371207e-06, + "loss": 0.4968, + "step": 2898 + }, + { + "epoch": 0.5311469402711616, + "grad_norm": 0.3974125326221091, + "learning_rate": 5.332333839957835e-06, + "loss": 0.4903, + "step": 2899 + }, + { + "epoch": 0.5313301575668743, + "grad_norm": 0.45799400702984955, + "learning_rate": 5.329142968601235e-06, + "loss": 0.5141, + "step": 2900 + }, + { + "epoch": 0.531513374862587, + "grad_norm": 0.3449525725893118, + "learning_rate": 5.325951962606652e-06, + "loss": 0.4758, + "step": 2901 + }, + { + "epoch": 0.5316965921582998, + "grad_norm": 0.3804584433180558, + "learning_rate": 5.322760823279387e-06, + "loss": 0.4512, + "step": 2902 + }, + { + "epoch": 0.5318798094540125, + "grad_norm": 0.4494280023735995, + "learning_rate": 5.319569551924796e-06, + "loss": 0.5056, + "step": 2903 + }, + { + "epoch": 0.5320630267497252, + "grad_norm": 0.3955072889662097, + "learning_rate": 5.316378149848288e-06, + "loss": 0.4927, + "step": 2904 + }, + { + "epoch": 0.5322462440454379, + "grad_norm": 0.42919220375892686, + "learning_rate": 5.313186618355325e-06, + "loss": 0.4916, + "step": 2905 + }, + { + "epoch": 0.5324294613411507, + "grad_norm": 0.40505314747590065, + "learning_rate": 5.309994958751425e-06, + "loss": 0.4908, + "step": 2906 + }, + { + "epoch": 0.5326126786368633, + "grad_norm": 0.4456449571578341, + "learning_rate": 5.3068031723421545e-06, + "loss": 0.5081, + "step": 2907 + }, + { + "epoch": 0.532795895932576, + "grad_norm": 0.3721241522017317, + "learning_rate": 5.303611260433134e-06, + "loss": 0.4749, + "step": 2908 + }, + { + "epoch": 0.5329791132282887, + "grad_norm": 0.38250684425244236, + "learning_rate": 5.300419224330035e-06, + "loss": 0.4485, + "step": 2909 + }, + { + "epoch": 0.5331623305240014, + "grad_norm": 0.4066445854359944, + "learning_rate": 5.2972270653385795e-06, + "loss": 0.4784, + "step": 2910 + }, + { + "epoch": 0.5333455478197142, + "grad_norm": 0.3851717388861089, + "learning_rate": 5.2940347847645425e-06, + "loss": 0.4653, + "step": 2911 + }, + { + "epoch": 0.5335287651154269, + "grad_norm": 0.3847978382682243, + "learning_rate": 5.290842383913741e-06, + "loss": 0.4976, + "step": 2912 + }, + { + "epoch": 0.5337119824111396, + "grad_norm": 0.4099519165225841, + "learning_rate": 5.287649864092053e-06, + "loss": 0.4694, + "step": 2913 + }, + { + "epoch": 0.5338951997068523, + "grad_norm": 0.3608571281303376, + "learning_rate": 5.284457226605392e-06, + "loss": 0.4994, + "step": 2914 + }, + { + "epoch": 0.5340784170025651, + "grad_norm": 0.38645401938215057, + "learning_rate": 5.281264472759731e-06, + "loss": 0.4578, + "step": 2915 + }, + { + "epoch": 0.5342616342982778, + "grad_norm": 0.3587512691919146, + "learning_rate": 5.2780716038610835e-06, + "loss": 0.4677, + "step": 2916 + }, + { + "epoch": 0.5344448515939905, + "grad_norm": 0.3915772277479349, + "learning_rate": 5.274878621215516e-06, + "loss": 0.4621, + "step": 2917 + }, + { + "epoch": 0.5346280688897032, + "grad_norm": 0.3489866287293818, + "learning_rate": 5.271685526129131e-06, + "loss": 0.4849, + "step": 2918 + }, + { + "epoch": 0.534811286185416, + "grad_norm": 0.34140728871959686, + "learning_rate": 5.2684923199080916e-06, + "loss": 0.489, + "step": 2919 + }, + { + "epoch": 0.5349945034811286, + "grad_norm": 0.38452077235446463, + "learning_rate": 5.265299003858594e-06, + "loss": 0.4698, + "step": 2920 + }, + { + "epoch": 0.5351777207768413, + "grad_norm": 0.36188439432129715, + "learning_rate": 5.262105579286886e-06, + "loss": 0.467, + "step": 2921 + }, + { + "epoch": 0.535360938072554, + "grad_norm": 0.40136003826650657, + "learning_rate": 5.258912047499257e-06, + "loss": 0.4897, + "step": 2922 + }, + { + "epoch": 0.5355441553682667, + "grad_norm": 0.38301156903817407, + "learning_rate": 5.255718409802041e-06, + "loss": 0.4846, + "step": 2923 + }, + { + "epoch": 0.5357273726639795, + "grad_norm": 0.4062414309735644, + "learning_rate": 5.252524667501615e-06, + "loss": 0.424, + "step": 2924 + }, + { + "epoch": 0.5359105899596922, + "grad_norm": 0.5669309183397974, + "learning_rate": 5.2493308219044006e-06, + "loss": 0.5234, + "step": 2925 + }, + { + "epoch": 0.5360938072554049, + "grad_norm": 0.35353928602965695, + "learning_rate": 5.246136874316859e-06, + "loss": 0.4733, + "step": 2926 + }, + { + "epoch": 0.5362770245511176, + "grad_norm": 0.3714597276288667, + "learning_rate": 5.242942826045496e-06, + "loss": 0.4764, + "step": 2927 + }, + { + "epoch": 0.5364602418468304, + "grad_norm": 0.37661479847969237, + "learning_rate": 5.239748678396854e-06, + "loss": 0.4537, + "step": 2928 + }, + { + "epoch": 0.5366434591425431, + "grad_norm": 0.36875780969579863, + "learning_rate": 5.236554432677523e-06, + "loss": 0.4502, + "step": 2929 + }, + { + "epoch": 0.5368266764382558, + "grad_norm": 0.37308350275176266, + "learning_rate": 5.233360090194125e-06, + "loss": 0.4513, + "step": 2930 + }, + { + "epoch": 0.5370098937339685, + "grad_norm": 0.3949888510977111, + "learning_rate": 5.230165652253329e-06, + "loss": 0.4719, + "step": 2931 + }, + { + "epoch": 0.5371931110296811, + "grad_norm": 0.4017042328027038, + "learning_rate": 5.226971120161838e-06, + "loss": 0.4628, + "step": 2932 + }, + { + "epoch": 0.537376328325394, + "grad_norm": 0.3509668407036333, + "learning_rate": 5.223776495226395e-06, + "loss": 0.4847, + "step": 2933 + }, + { + "epoch": 0.5375595456211066, + "grad_norm": 0.4245190133910541, + "learning_rate": 5.2205817787537825e-06, + "loss": 0.465, + "step": 2934 + }, + { + "epoch": 0.5377427629168193, + "grad_norm": 0.3579730150567853, + "learning_rate": 5.2173869720508186e-06, + "loss": 0.4795, + "step": 2935 + }, + { + "epoch": 0.537925980212532, + "grad_norm": 0.37265014399835306, + "learning_rate": 5.214192076424358e-06, + "loss": 0.4672, + "step": 2936 + }, + { + "epoch": 0.5381091975082448, + "grad_norm": 0.4179269118836094, + "learning_rate": 5.210997093181296e-06, + "loss": 0.5018, + "step": 2937 + }, + { + "epoch": 0.5382924148039575, + "grad_norm": 0.3254090650905135, + "learning_rate": 5.207802023628554e-06, + "loss": 0.5234, + "step": 2938 + }, + { + "epoch": 0.5384756320996702, + "grad_norm": 0.38656661518569324, + "learning_rate": 5.2046068690731035e-06, + "loss": 0.4514, + "step": 2939 + }, + { + "epoch": 0.5386588493953829, + "grad_norm": 0.392235416857842, + "learning_rate": 5.2014116308219344e-06, + "loss": 0.467, + "step": 2940 + }, + { + "epoch": 0.5388420666910957, + "grad_norm": 0.3783619599342323, + "learning_rate": 5.198216310182085e-06, + "loss": 0.4837, + "step": 2941 + }, + { + "epoch": 0.5390252839868084, + "grad_norm": 0.38311083355310077, + "learning_rate": 5.195020908460615e-06, + "loss": 0.4569, + "step": 2942 + }, + { + "epoch": 0.5392085012825211, + "grad_norm": 0.36048480834825547, + "learning_rate": 5.1918254269646305e-06, + "loss": 0.4693, + "step": 2943 + }, + { + "epoch": 0.5393917185782338, + "grad_norm": 0.5096803473624115, + "learning_rate": 5.188629867001257e-06, + "loss": 0.4906, + "step": 2944 + }, + { + "epoch": 0.5395749358739464, + "grad_norm": 0.39305512464764275, + "learning_rate": 5.185434229877662e-06, + "loss": 0.4496, + "step": 2945 + }, + { + "epoch": 0.5397581531696592, + "grad_norm": 0.38441166520861675, + "learning_rate": 5.182238516901039e-06, + "loss": 0.4853, + "step": 2946 + }, + { + "epoch": 0.5399413704653719, + "grad_norm": 0.3647631577483791, + "learning_rate": 5.179042729378616e-06, + "loss": 0.4796, + "step": 2947 + }, + { + "epoch": 0.5401245877610846, + "grad_norm": 0.33704086659708365, + "learning_rate": 5.175846868617648e-06, + "loss": 0.4328, + "step": 2948 + }, + { + "epoch": 0.5403078050567973, + "grad_norm": 0.4172774407255861, + "learning_rate": 5.172650935925422e-06, + "loss": 0.468, + "step": 2949 + }, + { + "epoch": 0.5404910223525101, + "grad_norm": 0.39166173820838274, + "learning_rate": 5.1694549326092545e-06, + "loss": 0.4781, + "step": 2950 + }, + { + "epoch": 0.5406742396482228, + "grad_norm": 0.35023848048516243, + "learning_rate": 5.166258859976491e-06, + "loss": 0.4696, + "step": 2951 + }, + { + "epoch": 0.5408574569439355, + "grad_norm": 0.40463812223445605, + "learning_rate": 5.163062719334504e-06, + "loss": 0.4519, + "step": 2952 + }, + { + "epoch": 0.5410406742396482, + "grad_norm": 0.3295763274091267, + "learning_rate": 5.159866511990696e-06, + "loss": 0.4404, + "step": 2953 + }, + { + "epoch": 0.541223891535361, + "grad_norm": 0.4006302305354375, + "learning_rate": 5.156670239252495e-06, + "loss": 0.4981, + "step": 2954 + }, + { + "epoch": 0.5414071088310737, + "grad_norm": 0.3278656070953833, + "learning_rate": 5.153473902427355e-06, + "loss": 0.4338, + "step": 2955 + }, + { + "epoch": 0.5415903261267864, + "grad_norm": 0.3802064956693402, + "learning_rate": 5.15027750282276e-06, + "loss": 0.4707, + "step": 2956 + }, + { + "epoch": 0.5417735434224991, + "grad_norm": 0.36364509241673776, + "learning_rate": 5.147081041746214e-06, + "loss": 0.4571, + "step": 2957 + }, + { + "epoch": 0.5419567607182117, + "grad_norm": 0.4222701241101004, + "learning_rate": 5.143884520505253e-06, + "loss": 0.4912, + "step": 2958 + }, + { + "epoch": 0.5421399780139246, + "grad_norm": 0.3690593576892358, + "learning_rate": 5.140687940407432e-06, + "loss": 0.4641, + "step": 2959 + }, + { + "epoch": 0.5423231953096372, + "grad_norm": 0.3642258437028377, + "learning_rate": 5.137491302760331e-06, + "loss": 0.5207, + "step": 2960 + }, + { + "epoch": 0.5425064126053499, + "grad_norm": 0.36258777119524377, + "learning_rate": 5.134294608871557e-06, + "loss": 0.4745, + "step": 2961 + }, + { + "epoch": 0.5426896299010626, + "grad_norm": 0.3592437874632688, + "learning_rate": 5.131097860048736e-06, + "loss": 0.457, + "step": 2962 + }, + { + "epoch": 0.5428728471967754, + "grad_norm": 0.34657579706751257, + "learning_rate": 5.127901057599517e-06, + "loss": 0.4811, + "step": 2963 + }, + { + "epoch": 0.5430560644924881, + "grad_norm": 0.38986574250324335, + "learning_rate": 5.124704202831574e-06, + "loss": 0.4761, + "step": 2964 + }, + { + "epoch": 0.5432392817882008, + "grad_norm": 0.354631208567437, + "learning_rate": 5.1215072970526004e-06, + "loss": 0.4727, + "step": 2965 + }, + { + "epoch": 0.5434224990839135, + "grad_norm": 0.39620267097722783, + "learning_rate": 5.118310341570308e-06, + "loss": 0.461, + "step": 2966 + }, + { + "epoch": 0.5436057163796263, + "grad_norm": 0.3809700887755271, + "learning_rate": 5.115113337692434e-06, + "loss": 0.4501, + "step": 2967 + }, + { + "epoch": 0.543788933675339, + "grad_norm": 0.6226141609011149, + "learning_rate": 5.1119162867267315e-06, + "loss": 0.4683, + "step": 2968 + }, + { + "epoch": 0.5439721509710517, + "grad_norm": 0.34478781370360134, + "learning_rate": 5.108719189980975e-06, + "loss": 0.4703, + "step": 2969 + }, + { + "epoch": 0.5441553682667644, + "grad_norm": 0.4188173225476515, + "learning_rate": 5.105522048762954e-06, + "loss": 0.4885, + "step": 2970 + }, + { + "epoch": 0.544338585562477, + "grad_norm": 0.5429472014892446, + "learning_rate": 5.1023248643804845e-06, + "loss": 0.4716, + "step": 2971 + }, + { + "epoch": 0.5445218028581899, + "grad_norm": 0.37228534397712537, + "learning_rate": 5.099127638141389e-06, + "loss": 0.4393, + "step": 2972 + }, + { + "epoch": 0.5447050201539025, + "grad_norm": 0.3745435037338173, + "learning_rate": 5.095930371353517e-06, + "loss": 0.4993, + "step": 2973 + }, + { + "epoch": 0.5448882374496152, + "grad_norm": 0.41412810270459044, + "learning_rate": 5.092733065324727e-06, + "loss": 0.4732, + "step": 2974 + }, + { + "epoch": 0.5450714547453279, + "grad_norm": 0.39667040566940515, + "learning_rate": 5.089535721362901e-06, + "loss": 0.5159, + "step": 2975 + }, + { + "epoch": 0.5452546720410407, + "grad_norm": 0.3506641882154736, + "learning_rate": 5.086338340775928e-06, + "loss": 0.4786, + "step": 2976 + }, + { + "epoch": 0.5454378893367534, + "grad_norm": 0.37045713084352444, + "learning_rate": 5.083140924871719e-06, + "loss": 0.4947, + "step": 2977 + }, + { + "epoch": 0.5456211066324661, + "grad_norm": 0.36590535185443057, + "learning_rate": 5.079943474958197e-06, + "loss": 0.4867, + "step": 2978 + }, + { + "epoch": 0.5458043239281788, + "grad_norm": 0.3809584128397892, + "learning_rate": 5.076745992343297e-06, + "loss": 0.4843, + "step": 2979 + }, + { + "epoch": 0.5459875412238915, + "grad_norm": 0.41125163480054133, + "learning_rate": 5.073548478334972e-06, + "loss": 0.4905, + "step": 2980 + }, + { + "epoch": 0.5461707585196043, + "grad_norm": 0.41888585924992366, + "learning_rate": 5.070350934241182e-06, + "loss": 0.5035, + "step": 2981 + }, + { + "epoch": 0.546353975815317, + "grad_norm": 0.3726285728522986, + "learning_rate": 5.067153361369905e-06, + "loss": 0.486, + "step": 2982 + }, + { + "epoch": 0.5465371931110297, + "grad_norm": 0.36792399066944415, + "learning_rate": 5.063955761029128e-06, + "loss": 0.4925, + "step": 2983 + }, + { + "epoch": 0.5467204104067424, + "grad_norm": 0.3977818481259419, + "learning_rate": 5.060758134526847e-06, + "loss": 0.4521, + "step": 2984 + }, + { + "epoch": 0.5469036277024552, + "grad_norm": 0.36585915155651066, + "learning_rate": 5.057560483171072e-06, + "loss": 0.4848, + "step": 2985 + }, + { + "epoch": 0.5470868449981678, + "grad_norm": 0.3345347835265389, + "learning_rate": 5.054362808269824e-06, + "loss": 0.4383, + "step": 2986 + }, + { + "epoch": 0.5472700622938805, + "grad_norm": 0.4530801780101565, + "learning_rate": 5.0511651111311285e-06, + "loss": 0.4737, + "step": 2987 + }, + { + "epoch": 0.5474532795895932, + "grad_norm": 0.3731202638551678, + "learning_rate": 5.047967393063027e-06, + "loss": 0.4965, + "step": 2988 + }, + { + "epoch": 0.547636496885306, + "grad_norm": 0.3400953185078803, + "learning_rate": 5.044769655373564e-06, + "loss": 0.4701, + "step": 2989 + }, + { + "epoch": 0.5478197141810187, + "grad_norm": 0.36743728384406593, + "learning_rate": 5.041571899370794e-06, + "loss": 0.4787, + "step": 2990 + }, + { + "epoch": 0.5480029314767314, + "grad_norm": 0.4103703822856523, + "learning_rate": 5.03837412636278e-06, + "loss": 0.457, + "step": 2991 + }, + { + "epoch": 0.5481861487724441, + "grad_norm": 0.3710505913853357, + "learning_rate": 5.03517633765759e-06, + "loss": 0.5121, + "step": 2992 + }, + { + "epoch": 0.5483693660681568, + "grad_norm": 0.35290976311155, + "learning_rate": 5.031978534563301e-06, + "loss": 0.4958, + "step": 2993 + }, + { + "epoch": 0.5485525833638696, + "grad_norm": 0.40567005135937617, + "learning_rate": 5.028780718387991e-06, + "loss": 0.4982, + "step": 2994 + }, + { + "epoch": 0.5487358006595823, + "grad_norm": 0.39949656554763924, + "learning_rate": 5.025582890439752e-06, + "loss": 0.4538, + "step": 2995 + }, + { + "epoch": 0.548919017955295, + "grad_norm": 0.35941919521494875, + "learning_rate": 5.022385052026668e-06, + "loss": 0.4728, + "step": 2996 + }, + { + "epoch": 0.5491022352510077, + "grad_norm": 0.38895581209117513, + "learning_rate": 5.019187204456842e-06, + "loss": 0.4882, + "step": 2997 + }, + { + "epoch": 0.5492854525467205, + "grad_norm": 0.36464831757502797, + "learning_rate": 5.015989349038369e-06, + "loss": 0.4603, + "step": 2998 + }, + { + "epoch": 0.5494686698424331, + "grad_norm": 0.4077652274710313, + "learning_rate": 5.012791487079355e-06, + "loss": 0.5101, + "step": 2999 + }, + { + "epoch": 0.5496518871381458, + "grad_norm": 0.36034817480747944, + "learning_rate": 5.0095936198879e-06, + "loss": 0.4624, + "step": 3000 + }, + { + "epoch": 0.5498351044338585, + "grad_norm": 0.4285338207306717, + "learning_rate": 5.006395748772119e-06, + "loss": 0.5193, + "step": 3001 + }, + { + "epoch": 0.5500183217295713, + "grad_norm": 0.3746957176964203, + "learning_rate": 5.003197875040115e-06, + "loss": 0.4696, + "step": 3002 + }, + { + "epoch": 0.550201539025284, + "grad_norm": 0.45507949686267096, + "learning_rate": 5e-06, + "loss": 0.4921, + "step": 3003 + }, + { + "epoch": 0.5503847563209967, + "grad_norm": 0.3609229066338744, + "learning_rate": 4.996802124959886e-06, + "loss": 0.5219, + "step": 3004 + }, + { + "epoch": 0.5505679736167094, + "grad_norm": 0.44642590441418006, + "learning_rate": 4.993604251227883e-06, + "loss": 0.4661, + "step": 3005 + }, + { + "epoch": 0.5507511909124221, + "grad_norm": 0.38751554388621373, + "learning_rate": 4.990406380112101e-06, + "loss": 0.4934, + "step": 3006 + }, + { + "epoch": 0.5509344082081349, + "grad_norm": 0.44084090414468313, + "learning_rate": 4.987208512920647e-06, + "loss": 0.4609, + "step": 3007 + }, + { + "epoch": 0.5511176255038476, + "grad_norm": 0.3826735702459895, + "learning_rate": 4.984010650961632e-06, + "loss": 0.4668, + "step": 3008 + }, + { + "epoch": 0.5513008427995603, + "grad_norm": 0.39052208118436643, + "learning_rate": 4.9808127955431605e-06, + "loss": 0.4765, + "step": 3009 + }, + { + "epoch": 0.551484060095273, + "grad_norm": 0.3538214074954766, + "learning_rate": 4.977614947973333e-06, + "loss": 0.4278, + "step": 3010 + }, + { + "epoch": 0.5516672773909858, + "grad_norm": 0.3525653240434517, + "learning_rate": 4.97441710956025e-06, + "loss": 0.4528, + "step": 3011 + }, + { + "epoch": 0.5518504946866984, + "grad_norm": 0.41849241943747617, + "learning_rate": 4.971219281612009e-06, + "loss": 0.5089, + "step": 3012 + }, + { + "epoch": 0.5520337119824111, + "grad_norm": 0.36896131804116705, + "learning_rate": 4.968021465436701e-06, + "loss": 0.4628, + "step": 3013 + }, + { + "epoch": 0.5522169292781238, + "grad_norm": 1.8962860706469424, + "learning_rate": 4.964823662342411e-06, + "loss": 0.4411, + "step": 3014 + }, + { + "epoch": 0.5524001465738366, + "grad_norm": 0.39520310244465107, + "learning_rate": 4.9616258736372206e-06, + "loss": 0.4573, + "step": 3015 + }, + { + "epoch": 0.5525833638695493, + "grad_norm": 0.3884957729352293, + "learning_rate": 4.958428100629206e-06, + "loss": 0.4787, + "step": 3016 + }, + { + "epoch": 0.552766581165262, + "grad_norm": 0.3739154780713499, + "learning_rate": 4.955230344626438e-06, + "loss": 0.4529, + "step": 3017 + }, + { + "epoch": 0.5529497984609747, + "grad_norm": 0.4645397422934775, + "learning_rate": 4.952032606936974e-06, + "loss": 0.5004, + "step": 3018 + }, + { + "epoch": 0.5531330157566874, + "grad_norm": 0.376257628657457, + "learning_rate": 4.948834888868872e-06, + "loss": 0.4695, + "step": 3019 + }, + { + "epoch": 0.5533162330524002, + "grad_norm": 0.7509468817825836, + "learning_rate": 4.945637191730177e-06, + "loss": 0.461, + "step": 3020 + }, + { + "epoch": 0.5534994503481129, + "grad_norm": 0.33093606690741983, + "learning_rate": 4.942439516828929e-06, + "loss": 0.461, + "step": 3021 + }, + { + "epoch": 0.5536826676438256, + "grad_norm": 0.342594923652272, + "learning_rate": 4.939241865473154e-06, + "loss": 0.483, + "step": 3022 + }, + { + "epoch": 0.5538658849395383, + "grad_norm": 0.42461934574028787, + "learning_rate": 4.936044238970874e-06, + "loss": 0.5221, + "step": 3023 + }, + { + "epoch": 0.5540491022352511, + "grad_norm": 0.3843975972052995, + "learning_rate": 4.932846638630095e-06, + "loss": 0.4718, + "step": 3024 + }, + { + "epoch": 0.5542323195309637, + "grad_norm": 0.35214701703123336, + "learning_rate": 4.929649065758819e-06, + "loss": 0.4936, + "step": 3025 + }, + { + "epoch": 0.5544155368266764, + "grad_norm": 0.39385494079941263, + "learning_rate": 4.926451521665029e-06, + "loss": 0.4569, + "step": 3026 + }, + { + "epoch": 0.5545987541223891, + "grad_norm": 0.3471261653623269, + "learning_rate": 4.9232540076567034e-06, + "loss": 0.4598, + "step": 3027 + }, + { + "epoch": 0.5547819714181018, + "grad_norm": 0.4254387065784489, + "learning_rate": 4.920056525041805e-06, + "loss": 0.4598, + "step": 3028 + }, + { + "epoch": 0.5549651887138146, + "grad_norm": 0.3477326042007442, + "learning_rate": 4.916859075128283e-06, + "loss": 0.4737, + "step": 3029 + }, + { + "epoch": 0.5551484060095273, + "grad_norm": 0.3583300309759384, + "learning_rate": 4.913661659224074e-06, + "loss": 0.4692, + "step": 3030 + }, + { + "epoch": 0.55533162330524, + "grad_norm": 0.34071521398866017, + "learning_rate": 4.910464278637101e-06, + "loss": 0.4583, + "step": 3031 + }, + { + "epoch": 0.5555148406009527, + "grad_norm": 0.3491675887668868, + "learning_rate": 4.9072669346752734e-06, + "loss": 0.4597, + "step": 3032 + }, + { + "epoch": 0.5556980578966655, + "grad_norm": 0.38898178790274124, + "learning_rate": 4.904069628646486e-06, + "loss": 0.4659, + "step": 3033 + }, + { + "epoch": 0.5558812751923782, + "grad_norm": 0.4310466722487581, + "learning_rate": 4.900872361858613e-06, + "loss": 0.4926, + "step": 3034 + }, + { + "epoch": 0.5560644924880909, + "grad_norm": 0.38569440984474046, + "learning_rate": 4.897675135619517e-06, + "loss": 0.4975, + "step": 3035 + }, + { + "epoch": 0.5562477097838036, + "grad_norm": 0.4430394243979431, + "learning_rate": 4.894477951237045e-06, + "loss": 0.4876, + "step": 3036 + }, + { + "epoch": 0.5564309270795164, + "grad_norm": 0.40782021540263214, + "learning_rate": 4.8912808100190275e-06, + "loss": 0.4353, + "step": 3037 + }, + { + "epoch": 0.556614144375229, + "grad_norm": 0.3706579167777146, + "learning_rate": 4.88808371327327e-06, + "loss": 0.452, + "step": 3038 + }, + { + "epoch": 0.5567973616709417, + "grad_norm": 0.3342714030205408, + "learning_rate": 4.884886662307567e-06, + "loss": 0.4424, + "step": 3039 + }, + { + "epoch": 0.5569805789666544, + "grad_norm": 0.3400620563273464, + "learning_rate": 4.881689658429693e-06, + "loss": 0.474, + "step": 3040 + }, + { + "epoch": 0.5571637962623671, + "grad_norm": 0.37477780306979874, + "learning_rate": 4.878492702947401e-06, + "loss": 0.4719, + "step": 3041 + }, + { + "epoch": 0.5573470135580799, + "grad_norm": 0.4023244017066431, + "learning_rate": 4.875295797168427e-06, + "loss": 0.4657, + "step": 3042 + }, + { + "epoch": 0.5575302308537926, + "grad_norm": 0.35993096960545706, + "learning_rate": 4.872098942400484e-06, + "loss": 0.4881, + "step": 3043 + }, + { + "epoch": 0.5577134481495053, + "grad_norm": 0.3408372914295605, + "learning_rate": 4.868902139951265e-06, + "loss": 0.4587, + "step": 3044 + }, + { + "epoch": 0.557896665445218, + "grad_norm": 0.41192915257702006, + "learning_rate": 4.8657053911284445e-06, + "loss": 0.4591, + "step": 3045 + }, + { + "epoch": 0.5580798827409308, + "grad_norm": 0.3584497304578561, + "learning_rate": 4.86250869723967e-06, + "loss": 0.4182, + "step": 3046 + }, + { + "epoch": 0.5582631000366435, + "grad_norm": 0.3607355707176063, + "learning_rate": 4.859312059592569e-06, + "loss": 0.4549, + "step": 3047 + }, + { + "epoch": 0.5584463173323562, + "grad_norm": 0.37328529033014385, + "learning_rate": 4.856115479494747e-06, + "loss": 0.4732, + "step": 3048 + }, + { + "epoch": 0.5586295346280689, + "grad_norm": 0.32431243654751757, + "learning_rate": 4.852918958253787e-06, + "loss": 0.4659, + "step": 3049 + }, + { + "epoch": 0.5588127519237817, + "grad_norm": 0.3971694740019051, + "learning_rate": 4.849722497177242e-06, + "loss": 0.4874, + "step": 3050 + }, + { + "epoch": 0.5589959692194943, + "grad_norm": 0.42362917357060553, + "learning_rate": 4.846526097572646e-06, + "loss": 0.459, + "step": 3051 + }, + { + "epoch": 0.559179186515207, + "grad_norm": 0.42189231248016035, + "learning_rate": 4.843329760747505e-06, + "loss": 0.5082, + "step": 3052 + }, + { + "epoch": 0.5593624038109197, + "grad_norm": 0.3346084353694287, + "learning_rate": 4.840133488009306e-06, + "loss": 0.464, + "step": 3053 + }, + { + "epoch": 0.5595456211066324, + "grad_norm": 0.4003637317955546, + "learning_rate": 4.836937280665497e-06, + "loss": 0.4593, + "step": 3054 + }, + { + "epoch": 0.5597288384023452, + "grad_norm": 0.4006410735017939, + "learning_rate": 4.83374114002351e-06, + "loss": 0.4703, + "step": 3055 + }, + { + "epoch": 0.5599120556980579, + "grad_norm": 0.38582218290002546, + "learning_rate": 4.830545067390746e-06, + "loss": 0.4931, + "step": 3056 + }, + { + "epoch": 0.5600952729937706, + "grad_norm": 0.39858644345264665, + "learning_rate": 4.827349064074581e-06, + "loss": 0.4778, + "step": 3057 + }, + { + "epoch": 0.5602784902894833, + "grad_norm": 0.36011543461086254, + "learning_rate": 4.824153131382354e-06, + "loss": 0.4631, + "step": 3058 + }, + { + "epoch": 0.5604617075851961, + "grad_norm": 0.35477133777824665, + "learning_rate": 4.820957270621385e-06, + "loss": 0.4634, + "step": 3059 + }, + { + "epoch": 0.5606449248809088, + "grad_norm": 0.39235049222066515, + "learning_rate": 4.817761483098961e-06, + "loss": 0.4834, + "step": 3060 + }, + { + "epoch": 0.5608281421766215, + "grad_norm": 0.356708467469616, + "learning_rate": 4.81456577012234e-06, + "loss": 0.4614, + "step": 3061 + }, + { + "epoch": 0.5610113594723342, + "grad_norm": 0.4023601789791014, + "learning_rate": 4.811370132998744e-06, + "loss": 0.423, + "step": 3062 + }, + { + "epoch": 0.5611945767680468, + "grad_norm": 0.3694080756344917, + "learning_rate": 4.80817457303537e-06, + "loss": 0.4902, + "step": 3063 + }, + { + "epoch": 0.5613777940637596, + "grad_norm": 0.4075503536678815, + "learning_rate": 4.804979091539385e-06, + "loss": 0.5068, + "step": 3064 + }, + { + "epoch": 0.5615610113594723, + "grad_norm": 0.3611503385733387, + "learning_rate": 4.801783689817918e-06, + "loss": 0.459, + "step": 3065 + }, + { + "epoch": 0.561744228655185, + "grad_norm": 0.3695413269567028, + "learning_rate": 4.798588369178067e-06, + "loss": 0.4502, + "step": 3066 + }, + { + "epoch": 0.5619274459508977, + "grad_norm": 0.39836545763359166, + "learning_rate": 4.795393130926899e-06, + "loss": 0.4969, + "step": 3067 + }, + { + "epoch": 0.5621106632466105, + "grad_norm": 0.3365585814117656, + "learning_rate": 4.792197976371446e-06, + "loss": 0.4583, + "step": 3068 + }, + { + "epoch": 0.5622938805423232, + "grad_norm": 0.38906298512693704, + "learning_rate": 4.789002906818707e-06, + "loss": 0.4495, + "step": 3069 + }, + { + "epoch": 0.5624770978380359, + "grad_norm": 0.4059950331091548, + "learning_rate": 4.785807923575644e-06, + "loss": 0.4706, + "step": 3070 + }, + { + "epoch": 0.5626603151337486, + "grad_norm": 0.38072106279290563, + "learning_rate": 4.782613027949183e-06, + "loss": 0.4687, + "step": 3071 + }, + { + "epoch": 0.5628435324294614, + "grad_norm": 0.38457397221908995, + "learning_rate": 4.779418221246218e-06, + "loss": 0.4894, + "step": 3072 + }, + { + "epoch": 0.5630267497251741, + "grad_norm": 0.3213042113105466, + "learning_rate": 4.776223504773607e-06, + "loss": 0.4704, + "step": 3073 + }, + { + "epoch": 0.5632099670208868, + "grad_norm": 0.4086690149764143, + "learning_rate": 4.773028879838164e-06, + "loss": 0.5033, + "step": 3074 + }, + { + "epoch": 0.5633931843165995, + "grad_norm": 0.41229899652618895, + "learning_rate": 4.769834347746672e-06, + "loss": 0.4562, + "step": 3075 + }, + { + "epoch": 0.5635764016123122, + "grad_norm": 0.48704419040281205, + "learning_rate": 4.766639909805875e-06, + "loss": 0.4967, + "step": 3076 + }, + { + "epoch": 0.563759618908025, + "grad_norm": 0.4505129252023178, + "learning_rate": 4.76344556732248e-06, + "loss": 0.477, + "step": 3077 + }, + { + "epoch": 0.5639428362037376, + "grad_norm": 0.41675282994752194, + "learning_rate": 4.760251321603147e-06, + "loss": 0.4933, + "step": 3078 + }, + { + "epoch": 0.5641260534994503, + "grad_norm": 0.37915614418127414, + "learning_rate": 4.7570571739545055e-06, + "loss": 0.4833, + "step": 3079 + }, + { + "epoch": 0.564309270795163, + "grad_norm": 0.3816164215347716, + "learning_rate": 4.753863125683142e-06, + "loss": 0.4498, + "step": 3080 + }, + { + "epoch": 0.5644924880908758, + "grad_norm": 0.3516334999800291, + "learning_rate": 4.750669178095602e-06, + "loss": 0.4746, + "step": 3081 + }, + { + "epoch": 0.5646757053865885, + "grad_norm": 0.40711908363080224, + "learning_rate": 4.7474753324983855e-06, + "loss": 0.5001, + "step": 3082 + }, + { + "epoch": 0.5648589226823012, + "grad_norm": 0.405519832056979, + "learning_rate": 4.744281590197961e-06, + "loss": 0.499, + "step": 3083 + }, + { + "epoch": 0.5650421399780139, + "grad_norm": 0.3955182456556916, + "learning_rate": 4.7410879525007444e-06, + "loss": 0.5042, + "step": 3084 + }, + { + "epoch": 0.5652253572737267, + "grad_norm": 0.39023164331760823, + "learning_rate": 4.737894420713116e-06, + "loss": 0.4842, + "step": 3085 + }, + { + "epoch": 0.5654085745694394, + "grad_norm": 0.3611894731220293, + "learning_rate": 4.734700996141408e-06, + "loss": 0.4638, + "step": 3086 + }, + { + "epoch": 0.5655917918651521, + "grad_norm": 0.3663949637677696, + "learning_rate": 4.731507680091909e-06, + "loss": 0.482, + "step": 3087 + }, + { + "epoch": 0.5657750091608648, + "grad_norm": 0.39243254907649366, + "learning_rate": 4.728314473870869e-06, + "loss": 0.4755, + "step": 3088 + }, + { + "epoch": 0.5659582264565775, + "grad_norm": 0.37931351609842306, + "learning_rate": 4.725121378784487e-06, + "loss": 0.499, + "step": 3089 + }, + { + "epoch": 0.5661414437522903, + "grad_norm": 0.3679955276603955, + "learning_rate": 4.721928396138918e-06, + "loss": 0.4679, + "step": 3090 + }, + { + "epoch": 0.5663246610480029, + "grad_norm": 0.37286524070443805, + "learning_rate": 4.71873552724027e-06, + "loss": 0.4934, + "step": 3091 + }, + { + "epoch": 0.5665078783437156, + "grad_norm": 0.3581401060788165, + "learning_rate": 4.715542773394609e-06, + "loss": 0.4639, + "step": 3092 + }, + { + "epoch": 0.5666910956394283, + "grad_norm": 0.4944839522717179, + "learning_rate": 4.712350135907951e-06, + "loss": 0.4555, + "step": 3093 + }, + { + "epoch": 0.5668743129351411, + "grad_norm": 0.3773580461873827, + "learning_rate": 4.70915761608626e-06, + "loss": 0.4723, + "step": 3094 + }, + { + "epoch": 0.5670575302308538, + "grad_norm": 0.3883641380661997, + "learning_rate": 4.70596521523546e-06, + "loss": 0.5029, + "step": 3095 + }, + { + "epoch": 0.5672407475265665, + "grad_norm": 0.34701280633245957, + "learning_rate": 4.702772934661421e-06, + "loss": 0.4789, + "step": 3096 + }, + { + "epoch": 0.5674239648222792, + "grad_norm": 0.39863313930275435, + "learning_rate": 4.699580775669967e-06, + "loss": 0.4558, + "step": 3097 + }, + { + "epoch": 0.567607182117992, + "grad_norm": 0.37465241701443214, + "learning_rate": 4.696388739566868e-06, + "loss": 0.4598, + "step": 3098 + }, + { + "epoch": 0.5677903994137047, + "grad_norm": 0.3729442445536715, + "learning_rate": 4.693196827657848e-06, + "loss": 0.4561, + "step": 3099 + }, + { + "epoch": 0.5679736167094174, + "grad_norm": 0.38556828564813417, + "learning_rate": 4.690005041248576e-06, + "loss": 0.4638, + "step": 3100 + }, + { + "epoch": 0.5681568340051301, + "grad_norm": 0.3305296393303103, + "learning_rate": 4.686813381644676e-06, + "loss": 0.4615, + "step": 3101 + }, + { + "epoch": 0.5683400513008428, + "grad_norm": 0.35015841336773623, + "learning_rate": 4.6836218501517135e-06, + "loss": 0.4601, + "step": 3102 + }, + { + "epoch": 0.5685232685965556, + "grad_norm": 0.36657344001092046, + "learning_rate": 4.680430448075205e-06, + "loss": 0.4714, + "step": 3103 + }, + { + "epoch": 0.5687064858922682, + "grad_norm": 0.3680000579785531, + "learning_rate": 4.677239176720613e-06, + "loss": 0.486, + "step": 3104 + }, + { + "epoch": 0.5688897031879809, + "grad_norm": 0.42383810586471016, + "learning_rate": 4.67404803739335e-06, + "loss": 0.46, + "step": 3105 + }, + { + "epoch": 0.5690729204836936, + "grad_norm": 0.3819192317502884, + "learning_rate": 4.670857031398766e-06, + "loss": 0.4799, + "step": 3106 + }, + { + "epoch": 0.5692561377794064, + "grad_norm": 0.35409508939859935, + "learning_rate": 4.667666160042166e-06, + "loss": 0.4557, + "step": 3107 + }, + { + "epoch": 0.5694393550751191, + "grad_norm": 0.3768258549206939, + "learning_rate": 4.664475424628795e-06, + "loss": 0.4655, + "step": 3108 + }, + { + "epoch": 0.5696225723708318, + "grad_norm": 0.3356425680201244, + "learning_rate": 4.661284826463842e-06, + "loss": 0.4493, + "step": 3109 + }, + { + "epoch": 0.5698057896665445, + "grad_norm": 0.37127689112134266, + "learning_rate": 4.658094366852442e-06, + "loss": 0.4647, + "step": 3110 + }, + { + "epoch": 0.5699890069622572, + "grad_norm": 1.59864474967665, + "learning_rate": 4.654904047099671e-06, + "loss": 0.4753, + "step": 3111 + }, + { + "epoch": 0.57017222425797, + "grad_norm": 0.4002133341134483, + "learning_rate": 4.651713868510552e-06, + "loss": 0.4942, + "step": 3112 + }, + { + "epoch": 0.5703554415536827, + "grad_norm": 0.40377734407222904, + "learning_rate": 4.6485238323900465e-06, + "loss": 0.4899, + "step": 3113 + }, + { + "epoch": 0.5705386588493954, + "grad_norm": 0.38467837522690307, + "learning_rate": 4.645333940043057e-06, + "loss": 0.4914, + "step": 3114 + }, + { + "epoch": 0.570721876145108, + "grad_norm": 0.3269570834287816, + "learning_rate": 4.642144192774429e-06, + "loss": 0.4614, + "step": 3115 + }, + { + "epoch": 0.5709050934408209, + "grad_norm": 0.38229824539147894, + "learning_rate": 4.63895459188895e-06, + "loss": 0.4886, + "step": 3116 + }, + { + "epoch": 0.5710883107365335, + "grad_norm": 0.4094637725031508, + "learning_rate": 4.6357651386913476e-06, + "loss": 0.4809, + "step": 3117 + }, + { + "epoch": 0.5712715280322462, + "grad_norm": 0.5056478139206053, + "learning_rate": 4.632575834486283e-06, + "loss": 0.4804, + "step": 3118 + }, + { + "epoch": 0.5714547453279589, + "grad_norm": 0.4139762052558197, + "learning_rate": 4.629386680578363e-06, + "loss": 0.4667, + "step": 3119 + }, + { + "epoch": 0.5716379626236717, + "grad_norm": 0.3681026176222857, + "learning_rate": 4.6261976782721315e-06, + "loss": 0.4678, + "step": 3120 + }, + { + "epoch": 0.5718211799193844, + "grad_norm": 0.35368475179304354, + "learning_rate": 4.623008828872071e-06, + "loss": 0.4907, + "step": 3121 + }, + { + "epoch": 0.5720043972150971, + "grad_norm": 0.374693398306428, + "learning_rate": 4.619820133682599e-06, + "loss": 0.4761, + "step": 3122 + }, + { + "epoch": 0.5721876145108098, + "grad_norm": 0.3901247116349376, + "learning_rate": 4.616631594008069e-06, + "loss": 0.4752, + "step": 3123 + }, + { + "epoch": 0.5723708318065225, + "grad_norm": 0.39016782297099584, + "learning_rate": 4.613443211152775e-06, + "loss": 0.4938, + "step": 3124 + }, + { + "epoch": 0.5725540491022353, + "grad_norm": 0.3617749323847835, + "learning_rate": 4.610254986420946e-06, + "loss": 0.4956, + "step": 3125 + }, + { + "epoch": 0.572737266397948, + "grad_norm": 0.36235207867255703, + "learning_rate": 4.607066921116743e-06, + "loss": 0.4813, + "step": 3126 + }, + { + "epoch": 0.5729204836936607, + "grad_norm": 0.3563629407497168, + "learning_rate": 4.603879016544266e-06, + "loss": 0.436, + "step": 3127 + }, + { + "epoch": 0.5731037009893734, + "grad_norm": 0.36620277042657723, + "learning_rate": 4.600691274007545e-06, + "loss": 0.4284, + "step": 3128 + }, + { + "epoch": 0.5732869182850862, + "grad_norm": 0.34641262438650583, + "learning_rate": 4.597503694810549e-06, + "loss": 0.4402, + "step": 3129 + }, + { + "epoch": 0.5734701355807988, + "grad_norm": 0.340387712659804, + "learning_rate": 4.594316280257173e-06, + "loss": 0.4905, + "step": 3130 + }, + { + "epoch": 0.5736533528765115, + "grad_norm": 0.3879953214677975, + "learning_rate": 4.591129031651252e-06, + "loss": 0.4867, + "step": 3131 + }, + { + "epoch": 0.5738365701722242, + "grad_norm": 0.3526720990486133, + "learning_rate": 4.587941950296548e-06, + "loss": 0.4814, + "step": 3132 + }, + { + "epoch": 0.574019787467937, + "grad_norm": 0.34906848400450996, + "learning_rate": 4.58475503749676e-06, + "loss": 0.4799, + "step": 3133 + }, + { + "epoch": 0.5742030047636497, + "grad_norm": 0.42232801199190145, + "learning_rate": 4.5815682945555095e-06, + "loss": 0.4846, + "step": 3134 + }, + { + "epoch": 0.5743862220593624, + "grad_norm": 0.39307394232320786, + "learning_rate": 4.578381722776357e-06, + "loss": 0.4744, + "step": 3135 + }, + { + "epoch": 0.5745694393550751, + "grad_norm": 0.4383560910040067, + "learning_rate": 4.575195323462789e-06, + "loss": 0.4229, + "step": 3136 + }, + { + "epoch": 0.5747526566507878, + "grad_norm": 0.3790862915479159, + "learning_rate": 4.5720090979182226e-06, + "loss": 0.4922, + "step": 3137 + }, + { + "epoch": 0.5749358739465006, + "grad_norm": 0.35851805823478117, + "learning_rate": 4.568823047446e-06, + "loss": 0.4419, + "step": 3138 + }, + { + "epoch": 0.5751190912422133, + "grad_norm": 0.37553059637594327, + "learning_rate": 4.5656371733493986e-06, + "loss": 0.4955, + "step": 3139 + }, + { + "epoch": 0.575302308537926, + "grad_norm": 0.34239091839678776, + "learning_rate": 4.562451476931617e-06, + "loss": 0.4562, + "step": 3140 + }, + { + "epoch": 0.5754855258336387, + "grad_norm": 0.36853187637460383, + "learning_rate": 4.559265959495789e-06, + "loss": 0.4559, + "step": 3141 + }, + { + "epoch": 0.5756687431293515, + "grad_norm": 0.4055837207551395, + "learning_rate": 4.556080622344966e-06, + "loss": 0.4585, + "step": 3142 + }, + { + "epoch": 0.5758519604250641, + "grad_norm": 0.4035077472358186, + "learning_rate": 4.552895466782129e-06, + "loss": 0.4837, + "step": 3143 + }, + { + "epoch": 0.5760351777207768, + "grad_norm": 0.37763399504887496, + "learning_rate": 4.54971049411019e-06, + "loss": 0.4858, + "step": 3144 + }, + { + "epoch": 0.5762183950164895, + "grad_norm": 0.3272918408288259, + "learning_rate": 4.54652570563198e-06, + "loss": 0.4534, + "step": 3145 + }, + { + "epoch": 0.5764016123122023, + "grad_norm": 0.44100852938903623, + "learning_rate": 4.543341102650256e-06, + "loss": 0.4901, + "step": 3146 + }, + { + "epoch": 0.576584829607915, + "grad_norm": 0.3546997965576367, + "learning_rate": 4.540156686467699e-06, + "loss": 0.4522, + "step": 3147 + }, + { + "epoch": 0.5767680469036277, + "grad_norm": 0.34375602357895785, + "learning_rate": 4.536972458386918e-06, + "loss": 0.4508, + "step": 3148 + }, + { + "epoch": 0.5769512641993404, + "grad_norm": 0.4279960918981899, + "learning_rate": 4.533788419710438e-06, + "loss": 0.449, + "step": 3149 + }, + { + "epoch": 0.5771344814950531, + "grad_norm": 0.385757250522885, + "learning_rate": 4.53060457174071e-06, + "loss": 0.4925, + "step": 3150 + }, + { + "epoch": 0.5773176987907659, + "grad_norm": 0.4627447133110814, + "learning_rate": 4.527420915780108e-06, + "loss": 0.5147, + "step": 3151 + }, + { + "epoch": 0.5775009160864786, + "grad_norm": 0.43834441097181, + "learning_rate": 4.524237453130925e-06, + "loss": 0.52, + "step": 3152 + }, + { + "epoch": 0.5776841333821913, + "grad_norm": 0.35016800344720833, + "learning_rate": 4.521054185095378e-06, + "loss": 0.4821, + "step": 3153 + }, + { + "epoch": 0.577867350677904, + "grad_norm": 0.3673462093000476, + "learning_rate": 4.517871112975601e-06, + "loss": 0.4728, + "step": 3154 + }, + { + "epoch": 0.5780505679736168, + "grad_norm": 0.35645458481596165, + "learning_rate": 4.514688238073651e-06, + "loss": 0.4947, + "step": 3155 + }, + { + "epoch": 0.5782337852693294, + "grad_norm": 0.3780429945224492, + "learning_rate": 4.511505561691501e-06, + "loss": 0.4832, + "step": 3156 + }, + { + "epoch": 0.5784170025650421, + "grad_norm": 0.43738027163322923, + "learning_rate": 4.508323085131048e-06, + "loss": 0.4838, + "step": 3157 + }, + { + "epoch": 0.5786002198607548, + "grad_norm": 0.4203463374367096, + "learning_rate": 4.505140809694099e-06, + "loss": 0.4678, + "step": 3158 + }, + { + "epoch": 0.5787834371564675, + "grad_norm": 0.4052582569339036, + "learning_rate": 4.501958736682387e-06, + "loss": 0.4832, + "step": 3159 + }, + { + "epoch": 0.5789666544521803, + "grad_norm": 0.3556187768954678, + "learning_rate": 4.4987768673975575e-06, + "loss": 0.4518, + "step": 3160 + }, + { + "epoch": 0.579149871747893, + "grad_norm": 0.4128521200164016, + "learning_rate": 4.495595203141176e-06, + "loss": 0.4688, + "step": 3161 + }, + { + "epoch": 0.5793330890436057, + "grad_norm": 0.3750611186666859, + "learning_rate": 4.492413745214718e-06, + "loss": 0.4813, + "step": 3162 + }, + { + "epoch": 0.5795163063393184, + "grad_norm": 0.37611356381414457, + "learning_rate": 4.489232494919583e-06, + "loss": 0.4637, + "step": 3163 + }, + { + "epoch": 0.5796995236350312, + "grad_norm": 0.5804757739317091, + "learning_rate": 4.4860514535570784e-06, + "loss": 0.4905, + "step": 3164 + }, + { + "epoch": 0.5798827409307439, + "grad_norm": 0.3724995632377092, + "learning_rate": 4.482870622428432e-06, + "loss": 0.4708, + "step": 3165 + }, + { + "epoch": 0.5800659582264566, + "grad_norm": 0.39702225281046527, + "learning_rate": 4.479690002834779e-06, + "loss": 0.4897, + "step": 3166 + }, + { + "epoch": 0.5802491755221693, + "grad_norm": 0.35264468266197835, + "learning_rate": 4.476509596077172e-06, + "loss": 0.4919, + "step": 3167 + }, + { + "epoch": 0.5804323928178821, + "grad_norm": 0.3884263103910947, + "learning_rate": 4.473329403456579e-06, + "loss": 0.4753, + "step": 3168 + }, + { + "epoch": 0.5806156101135947, + "grad_norm": 0.3601987136244253, + "learning_rate": 4.470149426273876e-06, + "loss": 0.4827, + "step": 3169 + }, + { + "epoch": 0.5807988274093074, + "grad_norm": 0.3880912808253205, + "learning_rate": 4.4669696658298536e-06, + "loss": 0.4848, + "step": 3170 + }, + { + "epoch": 0.5809820447050201, + "grad_norm": 0.42927775971470394, + "learning_rate": 4.463790123425209e-06, + "loss": 0.5008, + "step": 3171 + }, + { + "epoch": 0.5811652620007328, + "grad_norm": 0.3563899760173673, + "learning_rate": 4.460610800360558e-06, + "loss": 0.4656, + "step": 3172 + }, + { + "epoch": 0.5813484792964456, + "grad_norm": 0.36070762128328254, + "learning_rate": 4.457431697936422e-06, + "loss": 0.4563, + "step": 3173 + }, + { + "epoch": 0.5815316965921583, + "grad_norm": 0.69688988972388, + "learning_rate": 4.4542528174532305e-06, + "loss": 0.4418, + "step": 3174 + }, + { + "epoch": 0.581714913887871, + "grad_norm": 0.38109816794128887, + "learning_rate": 4.451074160211325e-06, + "loss": 0.4682, + "step": 3175 + }, + { + "epoch": 0.5818981311835837, + "grad_norm": 0.36905169783384256, + "learning_rate": 4.447895727510956e-06, + "loss": 0.4606, + "step": 3176 + }, + { + "epoch": 0.5820813484792965, + "grad_norm": 0.3503813337020013, + "learning_rate": 4.444717520652282e-06, + "loss": 0.4421, + "step": 3177 + }, + { + "epoch": 0.5822645657750092, + "grad_norm": 0.39072420310402833, + "learning_rate": 4.441539540935365e-06, + "loss": 0.4609, + "step": 3178 + }, + { + "epoch": 0.5824477830707219, + "grad_norm": 0.39662369870131087, + "learning_rate": 4.438361789660182e-06, + "loss": 0.4848, + "step": 3179 + }, + { + "epoch": 0.5826310003664346, + "grad_norm": 0.3734550437349449, + "learning_rate": 4.4351842681266064e-06, + "loss": 0.4693, + "step": 3180 + }, + { + "epoch": 0.5828142176621474, + "grad_norm": 0.4148265013660583, + "learning_rate": 4.43200697763443e-06, + "loss": 0.4934, + "step": 3181 + }, + { + "epoch": 0.58299743495786, + "grad_norm": 0.3820577449886745, + "learning_rate": 4.428829919483336e-06, + "loss": 0.4876, + "step": 3182 + }, + { + "epoch": 0.5831806522535727, + "grad_norm": 0.34649138055012, + "learning_rate": 4.425653094972925e-06, + "loss": 0.4801, + "step": 3183 + }, + { + "epoch": 0.5833638695492854, + "grad_norm": 0.3329444173954086, + "learning_rate": 4.422476505402694e-06, + "loss": 0.442, + "step": 3184 + }, + { + "epoch": 0.5835470868449981, + "grad_norm": 0.3583613198396569, + "learning_rate": 4.419300152072051e-06, + "loss": 0.5012, + "step": 3185 + }, + { + "epoch": 0.5837303041407109, + "grad_norm": 0.34491323872204827, + "learning_rate": 4.416124036280297e-06, + "loss": 0.4802, + "step": 3186 + }, + { + "epoch": 0.5839135214364236, + "grad_norm": 0.39803368312194737, + "learning_rate": 4.412948159326647e-06, + "loss": 0.4946, + "step": 3187 + }, + { + "epoch": 0.5840967387321363, + "grad_norm": 0.3295580108587731, + "learning_rate": 4.40977252251021e-06, + "loss": 0.4429, + "step": 3188 + }, + { + "epoch": 0.584279956027849, + "grad_norm": 0.3863844668783294, + "learning_rate": 4.406597127130003e-06, + "loss": 0.4297, + "step": 3189 + }, + { + "epoch": 0.5844631733235618, + "grad_norm": 0.3861977656105318, + "learning_rate": 4.403421974484938e-06, + "loss": 0.4649, + "step": 3190 + }, + { + "epoch": 0.5846463906192745, + "grad_norm": 0.37991096428514454, + "learning_rate": 4.400247065873834e-06, + "loss": 0.4732, + "step": 3191 + }, + { + "epoch": 0.5848296079149872, + "grad_norm": 0.390441967147969, + "learning_rate": 4.397072402595404e-06, + "loss": 0.5095, + "step": 3192 + }, + { + "epoch": 0.5850128252106999, + "grad_norm": 0.3482150955028248, + "learning_rate": 4.393897985948268e-06, + "loss": 0.4474, + "step": 3193 + }, + { + "epoch": 0.5851960425064126, + "grad_norm": 0.3730706636252824, + "learning_rate": 4.390723817230937e-06, + "loss": 0.4788, + "step": 3194 + }, + { + "epoch": 0.5853792598021254, + "grad_norm": 0.4268108094738352, + "learning_rate": 4.3875498977418255e-06, + "loss": 0.457, + "step": 3195 + }, + { + "epoch": 0.585562477097838, + "grad_norm": 0.3801499865703363, + "learning_rate": 4.384376228779246e-06, + "loss": 0.4651, + "step": 3196 + }, + { + "epoch": 0.5857456943935507, + "grad_norm": 0.3965825730216222, + "learning_rate": 4.381202811641409e-06, + "loss": 0.4732, + "step": 3197 + }, + { + "epoch": 0.5859289116892634, + "grad_norm": 0.6549080598602415, + "learning_rate": 4.378029647626416e-06, + "loss": 0.4493, + "step": 3198 + }, + { + "epoch": 0.5861121289849762, + "grad_norm": 0.4075193306605999, + "learning_rate": 4.3748567380322735e-06, + "loss": 0.5111, + "step": 3199 + }, + { + "epoch": 0.5862953462806889, + "grad_norm": 0.3909668781020062, + "learning_rate": 4.371684084156878e-06, + "loss": 0.4661, + "step": 3200 + }, + { + "epoch": 0.5864785635764016, + "grad_norm": 0.37701594797423776, + "learning_rate": 4.3685116872980255e-06, + "loss": 0.4639, + "step": 3201 + }, + { + "epoch": 0.5866617808721143, + "grad_norm": 0.34604006575622553, + "learning_rate": 4.365339548753402e-06, + "loss": 0.4785, + "step": 3202 + }, + { + "epoch": 0.5868449981678271, + "grad_norm": 0.46183749257336376, + "learning_rate": 4.362167669820593e-06, + "loss": 0.4701, + "step": 3203 + }, + { + "epoch": 0.5870282154635398, + "grad_norm": 0.3959468024878008, + "learning_rate": 4.358996051797071e-06, + "loss": 0.4902, + "step": 3204 + }, + { + "epoch": 0.5872114327592525, + "grad_norm": 0.4432973129626992, + "learning_rate": 4.355824695980212e-06, + "loss": 0.4856, + "step": 3205 + }, + { + "epoch": 0.5873946500549652, + "grad_norm": 0.4007869604471848, + "learning_rate": 4.352653603667276e-06, + "loss": 0.5027, + "step": 3206 + }, + { + "epoch": 0.5875778673506779, + "grad_norm": 0.3560212206932624, + "learning_rate": 4.349482776155417e-06, + "loss": 0.4749, + "step": 3207 + }, + { + "epoch": 0.5877610846463907, + "grad_norm": 0.3456678595669956, + "learning_rate": 4.346312214741682e-06, + "loss": 0.468, + "step": 3208 + }, + { + "epoch": 0.5879443019421033, + "grad_norm": 0.399718091153767, + "learning_rate": 4.343141920723013e-06, + "loss": 0.4673, + "step": 3209 + }, + { + "epoch": 0.588127519237816, + "grad_norm": 0.4219557594975826, + "learning_rate": 4.339971895396233e-06, + "loss": 0.4863, + "step": 3210 + }, + { + "epoch": 0.5883107365335287, + "grad_norm": 0.3621115344008167, + "learning_rate": 4.3368021400580635e-06, + "loss": 0.4374, + "step": 3211 + }, + { + "epoch": 0.5884939538292415, + "grad_norm": 0.3367939423815905, + "learning_rate": 4.333632656005112e-06, + "loss": 0.4748, + "step": 3212 + }, + { + "epoch": 0.5886771711249542, + "grad_norm": 0.3944240297595139, + "learning_rate": 4.330463444533877e-06, + "loss": 0.5025, + "step": 3213 + }, + { + "epoch": 0.5888603884206669, + "grad_norm": 0.32649317025540725, + "learning_rate": 4.3272945069407404e-06, + "loss": 0.4621, + "step": 3214 + }, + { + "epoch": 0.5890436057163796, + "grad_norm": 0.4244592566351569, + "learning_rate": 4.324125844521981e-06, + "loss": 0.4751, + "step": 3215 + }, + { + "epoch": 0.5892268230120924, + "grad_norm": 0.35927325305263147, + "learning_rate": 4.320957458573756e-06, + "loss": 0.473, + "step": 3216 + }, + { + "epoch": 0.5894100403078051, + "grad_norm": 0.34115890725125336, + "learning_rate": 4.317789350392116e-06, + "loss": 0.4964, + "step": 3217 + }, + { + "epoch": 0.5895932576035178, + "grad_norm": 0.42509519235085785, + "learning_rate": 4.314621521272994e-06, + "loss": 0.4895, + "step": 3218 + }, + { + "epoch": 0.5897764748992305, + "grad_norm": 0.3790800280473008, + "learning_rate": 4.311453972512209e-06, + "loss": 0.4921, + "step": 3219 + }, + { + "epoch": 0.5899596921949432, + "grad_norm": 0.3739335711998157, + "learning_rate": 4.308286705405469e-06, + "loss": 0.4734, + "step": 3220 + }, + { + "epoch": 0.590142909490656, + "grad_norm": 0.38837408654299566, + "learning_rate": 4.305119721248365e-06, + "loss": 0.4731, + "step": 3221 + }, + { + "epoch": 0.5903261267863686, + "grad_norm": 0.43945550067493544, + "learning_rate": 4.301953021336369e-06, + "loss": 0.4878, + "step": 3222 + }, + { + "epoch": 0.5905093440820813, + "grad_norm": 0.36484498856247866, + "learning_rate": 4.2987866069648384e-06, + "loss": 0.4363, + "step": 3223 + }, + { + "epoch": 0.590692561377794, + "grad_norm": 0.3724567230092767, + "learning_rate": 4.295620479429018e-06, + "loss": 0.4765, + "step": 3224 + }, + { + "epoch": 0.5908757786735068, + "grad_norm": 0.3942857083671079, + "learning_rate": 4.292454640024032e-06, + "loss": 0.4802, + "step": 3225 + }, + { + "epoch": 0.5910589959692195, + "grad_norm": 0.43761642400875916, + "learning_rate": 4.289289090044885e-06, + "loss": 0.4411, + "step": 3226 + }, + { + "epoch": 0.5912422132649322, + "grad_norm": 0.38579320660059957, + "learning_rate": 4.286123830786463e-06, + "loss": 0.4848, + "step": 3227 + }, + { + "epoch": 0.5914254305606449, + "grad_norm": 0.3754217946546912, + "learning_rate": 4.282958863543539e-06, + "loss": 0.4689, + "step": 3228 + }, + { + "epoch": 0.5916086478563577, + "grad_norm": 0.3935079146365459, + "learning_rate": 4.2797941896107625e-06, + "loss": 0.4926, + "step": 3229 + }, + { + "epoch": 0.5917918651520704, + "grad_norm": 0.3997898856736963, + "learning_rate": 4.27662981028266e-06, + "loss": 0.4964, + "step": 3230 + }, + { + "epoch": 0.5919750824477831, + "grad_norm": 0.46428919487523207, + "learning_rate": 4.2734657268536426e-06, + "loss": 0.4874, + "step": 3231 + }, + { + "epoch": 0.5921582997434958, + "grad_norm": 0.44062197093175026, + "learning_rate": 4.270301940617996e-06, + "loss": 0.4787, + "step": 3232 + }, + { + "epoch": 0.5923415170392085, + "grad_norm": 0.3352016063868999, + "learning_rate": 4.267138452869891e-06, + "loss": 0.4439, + "step": 3233 + }, + { + "epoch": 0.5925247343349213, + "grad_norm": 0.3675365473394679, + "learning_rate": 4.263975264903369e-06, + "loss": 0.4903, + "step": 3234 + }, + { + "epoch": 0.592707951630634, + "grad_norm": 0.3903938615069962, + "learning_rate": 4.260812378012353e-06, + "loss": 0.4834, + "step": 3235 + }, + { + "epoch": 0.5928911689263466, + "grad_norm": 0.3949595588434657, + "learning_rate": 4.257649793490639e-06, + "loss": 0.4782, + "step": 3236 + }, + { + "epoch": 0.5930743862220593, + "grad_norm": 0.36096674207763313, + "learning_rate": 4.254487512631907e-06, + "loss": 0.4713, + "step": 3237 + }, + { + "epoch": 0.5932576035177721, + "grad_norm": 0.40382524565681127, + "learning_rate": 4.251325536729702e-06, + "loss": 0.4791, + "step": 3238 + }, + { + "epoch": 0.5934408208134848, + "grad_norm": 0.3664711478737434, + "learning_rate": 4.2481638670774524e-06, + "loss": 0.426, + "step": 3239 + }, + { + "epoch": 0.5936240381091975, + "grad_norm": 0.3573410696419821, + "learning_rate": 4.245002504968459e-06, + "loss": 0.4755, + "step": 3240 + }, + { + "epoch": 0.5938072554049102, + "grad_norm": 0.36845317959067464, + "learning_rate": 4.241841451695898e-06, + "loss": 0.4373, + "step": 3241 + }, + { + "epoch": 0.5939904727006229, + "grad_norm": 0.3388176413214471, + "learning_rate": 4.238680708552816e-06, + "loss": 0.4336, + "step": 3242 + }, + { + "epoch": 0.5941736899963357, + "grad_norm": 0.3750781122068856, + "learning_rate": 4.235520276832134e-06, + "loss": 0.501, + "step": 3243 + }, + { + "epoch": 0.5943569072920484, + "grad_norm": 0.36440141301179285, + "learning_rate": 4.232360157826649e-06, + "loss": 0.4535, + "step": 3244 + }, + { + "epoch": 0.5945401245877611, + "grad_norm": 0.34522673261995696, + "learning_rate": 4.2292003528290256e-06, + "loss": 0.4814, + "step": 3245 + }, + { + "epoch": 0.5947233418834738, + "grad_norm": 0.3644210121918719, + "learning_rate": 4.226040863131802e-06, + "loss": 0.4478, + "step": 3246 + }, + { + "epoch": 0.5949065591791866, + "grad_norm": 0.36097756739126274, + "learning_rate": 4.222881690027386e-06, + "loss": 0.4604, + "step": 3247 + }, + { + "epoch": 0.5950897764748992, + "grad_norm": 0.3786748369783585, + "learning_rate": 4.21972283480806e-06, + "loss": 0.4844, + "step": 3248 + }, + { + "epoch": 0.5952729937706119, + "grad_norm": 0.4120827104886208, + "learning_rate": 4.216564298765971e-06, + "loss": 0.4528, + "step": 3249 + }, + { + "epoch": 0.5954562110663246, + "grad_norm": 0.3845984302303326, + "learning_rate": 4.213406083193139e-06, + "loss": 0.4891, + "step": 3250 + }, + { + "epoch": 0.5956394283620374, + "grad_norm": 0.42362371162602624, + "learning_rate": 4.2102481893814504e-06, + "loss": 0.4698, + "step": 3251 + }, + { + "epoch": 0.5958226456577501, + "grad_norm": 0.41905297506549305, + "learning_rate": 4.207090618622662e-06, + "loss": 0.4853, + "step": 3252 + }, + { + "epoch": 0.5960058629534628, + "grad_norm": 0.41851853864283367, + "learning_rate": 4.203933372208401e-06, + "loss": 0.4692, + "step": 3253 + }, + { + "epoch": 0.5961890802491755, + "grad_norm": 0.3627530325514232, + "learning_rate": 4.200776451430153e-06, + "loss": 0.4286, + "step": 3254 + }, + { + "epoch": 0.5963722975448882, + "grad_norm": 0.3498692983463641, + "learning_rate": 4.197619857579279e-06, + "loss": 0.4762, + "step": 3255 + }, + { + "epoch": 0.596555514840601, + "grad_norm": 0.4361187165313998, + "learning_rate": 4.194463591947002e-06, + "loss": 0.463, + "step": 3256 + }, + { + "epoch": 0.5967387321363137, + "grad_norm": 0.3865616410975838, + "learning_rate": 4.191307655824417e-06, + "loss": 0.4798, + "step": 3257 + }, + { + "epoch": 0.5969219494320264, + "grad_norm": 0.4312205474860647, + "learning_rate": 4.188152050502473e-06, + "loss": 0.5005, + "step": 3258 + }, + { + "epoch": 0.5971051667277391, + "grad_norm": 0.34766602024367504, + "learning_rate": 4.1849967772719935e-06, + "loss": 0.4566, + "step": 3259 + }, + { + "epoch": 0.5972883840234519, + "grad_norm": 0.3656894129494746, + "learning_rate": 4.181841837423662e-06, + "loss": 0.4902, + "step": 3260 + }, + { + "epoch": 0.5974716013191645, + "grad_norm": 0.3553982114963205, + "learning_rate": 4.178687232248028e-06, + "loss": 0.4594, + "step": 3261 + }, + { + "epoch": 0.5976548186148772, + "grad_norm": 0.3904959414801244, + "learning_rate": 4.175532963035497e-06, + "loss": 0.4693, + "step": 3262 + }, + { + "epoch": 0.5978380359105899, + "grad_norm": 0.4745543850446089, + "learning_rate": 4.172379031076348e-06, + "loss": 0.5156, + "step": 3263 + }, + { + "epoch": 0.5980212532063027, + "grad_norm": 0.41061453961539274, + "learning_rate": 4.169225437660714e-06, + "loss": 0.4697, + "step": 3264 + }, + { + "epoch": 0.5982044705020154, + "grad_norm": 0.42817726133500134, + "learning_rate": 4.166072184078596e-06, + "loss": 0.4891, + "step": 3265 + }, + { + "epoch": 0.5983876877977281, + "grad_norm": 0.46522544817243594, + "learning_rate": 4.162919271619844e-06, + "loss": 0.4789, + "step": 3266 + }, + { + "epoch": 0.5985709050934408, + "grad_norm": 0.41534295805740984, + "learning_rate": 4.159766701574184e-06, + "loss": 0.4716, + "step": 3267 + }, + { + "epoch": 0.5987541223891535, + "grad_norm": 0.38879974059023537, + "learning_rate": 4.15661447523119e-06, + "loss": 0.4699, + "step": 3268 + }, + { + "epoch": 0.5989373396848663, + "grad_norm": 0.3593410567415435, + "learning_rate": 4.153462593880304e-06, + "loss": 0.4626, + "step": 3269 + }, + { + "epoch": 0.599120556980579, + "grad_norm": 0.3460898119869833, + "learning_rate": 4.150311058810817e-06, + "loss": 0.4704, + "step": 3270 + }, + { + "epoch": 0.5993037742762917, + "grad_norm": 0.3615767581361931, + "learning_rate": 4.14715987131189e-06, + "loss": 0.4841, + "step": 3271 + }, + { + "epoch": 0.5994869915720044, + "grad_norm": 0.40285229112057547, + "learning_rate": 4.144009032672532e-06, + "loss": 0.4661, + "step": 3272 + }, + { + "epoch": 0.5996702088677172, + "grad_norm": 0.4005746642403547, + "learning_rate": 4.140858544181615e-06, + "loss": 0.4451, + "step": 3273 + }, + { + "epoch": 0.5998534261634298, + "grad_norm": 0.4072178303184198, + "learning_rate": 4.1377084071278654e-06, + "loss": 0.4874, + "step": 3274 + }, + { + "epoch": 0.6000366434591425, + "grad_norm": 0.3833025575736666, + "learning_rate": 4.1345586227998645e-06, + "loss": 0.4792, + "step": 3275 + }, + { + "epoch": 0.6002198607548552, + "grad_norm": 0.4233411723435603, + "learning_rate": 4.131409192486054e-06, + "loss": 0.4841, + "step": 3276 + }, + { + "epoch": 0.600403078050568, + "grad_norm": 0.3286893361028882, + "learning_rate": 4.128260117474728e-06, + "loss": 0.4204, + "step": 3277 + }, + { + "epoch": 0.6005862953462807, + "grad_norm": 0.36508385202796356, + "learning_rate": 4.125111399054034e-06, + "loss": 0.4561, + "step": 3278 + }, + { + "epoch": 0.6007695126419934, + "grad_norm": 0.38432905027467973, + "learning_rate": 4.121963038511973e-06, + "loss": 0.4933, + "step": 3279 + }, + { + "epoch": 0.6009527299377061, + "grad_norm": 0.5728067186258301, + "learning_rate": 4.118815037136405e-06, + "loss": 0.4873, + "step": 3280 + }, + { + "epoch": 0.6011359472334188, + "grad_norm": 0.32439808513378116, + "learning_rate": 4.11566739621504e-06, + "loss": 0.4304, + "step": 3281 + }, + { + "epoch": 0.6013191645291316, + "grad_norm": 0.35074318590462544, + "learning_rate": 4.1125201170354365e-06, + "loss": 0.4776, + "step": 3282 + }, + { + "epoch": 0.6015023818248443, + "grad_norm": 0.39504665360850666, + "learning_rate": 4.109373200885011e-06, + "loss": 0.4981, + "step": 3283 + }, + { + "epoch": 0.601685599120557, + "grad_norm": 0.3523157202895909, + "learning_rate": 4.106226649051027e-06, + "loss": 0.4675, + "step": 3284 + }, + { + "epoch": 0.6018688164162697, + "grad_norm": 0.38688955767161504, + "learning_rate": 4.103080462820606e-06, + "loss": 0.4469, + "step": 3285 + }, + { + "epoch": 0.6020520337119825, + "grad_norm": 0.4037212619297821, + "learning_rate": 4.09993464348071e-06, + "loss": 0.4966, + "step": 3286 + }, + { + "epoch": 0.6022352510076951, + "grad_norm": 0.338664130586936, + "learning_rate": 4.096789192318158e-06, + "loss": 0.4566, + "step": 3287 + }, + { + "epoch": 0.6024184683034078, + "grad_norm": 0.3730377521839032, + "learning_rate": 4.093644110619616e-06, + "loss": 0.4673, + "step": 3288 + }, + { + "epoch": 0.6026016855991205, + "grad_norm": 0.4410154194921889, + "learning_rate": 4.0904993996716015e-06, + "loss": 0.497, + "step": 3289 + }, + { + "epoch": 0.6027849028948332, + "grad_norm": 0.3474152298406819, + "learning_rate": 4.087355060760475e-06, + "loss": 0.4842, + "step": 3290 + }, + { + "epoch": 0.602968120190546, + "grad_norm": 0.39338377028821103, + "learning_rate": 4.084211095172448e-06, + "loss": 0.5153, + "step": 3291 + }, + { + "epoch": 0.6031513374862587, + "grad_norm": 0.46388513687309446, + "learning_rate": 4.081067504193582e-06, + "loss": 0.4754, + "step": 3292 + }, + { + "epoch": 0.6033345547819714, + "grad_norm": 0.35439786091759473, + "learning_rate": 4.077924289109781e-06, + "loss": 0.4641, + "step": 3293 + }, + { + "epoch": 0.6035177720776841, + "grad_norm": 0.3400717901610813, + "learning_rate": 4.074781451206795e-06, + "loss": 0.4907, + "step": 3294 + }, + { + "epoch": 0.6037009893733969, + "grad_norm": 0.5968859035737533, + "learning_rate": 4.071638991770225e-06, + "loss": 0.4534, + "step": 3295 + }, + { + "epoch": 0.6038842066691096, + "grad_norm": 0.35300902391021466, + "learning_rate": 4.0684969120855096e-06, + "loss": 0.4794, + "step": 3296 + }, + { + "epoch": 0.6040674239648223, + "grad_norm": 0.3663899793384046, + "learning_rate": 4.065355213437941e-06, + "loss": 0.4655, + "step": 3297 + }, + { + "epoch": 0.604250641260535, + "grad_norm": 0.3731638746108795, + "learning_rate": 4.062213897112645e-06, + "loss": 0.4995, + "step": 3298 + }, + { + "epoch": 0.6044338585562478, + "grad_norm": 0.34089212198020424, + "learning_rate": 4.059072964394599e-06, + "loss": 0.4546, + "step": 3299 + }, + { + "epoch": 0.6046170758519605, + "grad_norm": 0.3386444315353538, + "learning_rate": 4.055932416568623e-06, + "loss": 0.4741, + "step": 3300 + }, + { + "epoch": 0.6048002931476731, + "grad_norm": 0.3362900873564531, + "learning_rate": 4.052792254919377e-06, + "loss": 0.4516, + "step": 3301 + }, + { + "epoch": 0.6049835104433858, + "grad_norm": 0.534322389460024, + "learning_rate": 4.049652480731361e-06, + "loss": 0.4785, + "step": 3302 + }, + { + "epoch": 0.6051667277390985, + "grad_norm": 0.3718626080518362, + "learning_rate": 4.0465130952889194e-06, + "loss": 0.5092, + "step": 3303 + }, + { + "epoch": 0.6053499450348113, + "grad_norm": 0.3859006167009798, + "learning_rate": 4.0433740998762415e-06, + "loss": 0.4857, + "step": 3304 + }, + { + "epoch": 0.605533162330524, + "grad_norm": 0.39066559203718143, + "learning_rate": 4.040235495777351e-06, + "loss": 0.4314, + "step": 3305 + }, + { + "epoch": 0.6057163796262367, + "grad_norm": 0.37644682265086415, + "learning_rate": 4.037097284276112e-06, + "loss": 0.4751, + "step": 3306 + }, + { + "epoch": 0.6058995969219494, + "grad_norm": 0.43903459855723626, + "learning_rate": 4.0339594666562294e-06, + "loss": 0.4582, + "step": 3307 + }, + { + "epoch": 0.6060828142176622, + "grad_norm": 0.3699736577445379, + "learning_rate": 4.0308220442012495e-06, + "loss": 0.4609, + "step": 3308 + }, + { + "epoch": 0.6062660315133749, + "grad_norm": 0.42756536851629257, + "learning_rate": 4.0276850181945525e-06, + "loss": 0.4882, + "step": 3309 + }, + { + "epoch": 0.6064492488090876, + "grad_norm": 0.4080749512763187, + "learning_rate": 4.02454838991936e-06, + "loss": 0.4791, + "step": 3310 + }, + { + "epoch": 0.6066324661048003, + "grad_norm": 0.5944905217342814, + "learning_rate": 4.021412160658727e-06, + "loss": 0.4764, + "step": 3311 + }, + { + "epoch": 0.6068156834005131, + "grad_norm": 0.369199522390826, + "learning_rate": 4.0182763316955495e-06, + "loss": 0.4896, + "step": 3312 + }, + { + "epoch": 0.6069989006962258, + "grad_norm": 0.3896741072997416, + "learning_rate": 4.015140904312557e-06, + "loss": 0.4494, + "step": 3313 + }, + { + "epoch": 0.6071821179919384, + "grad_norm": 0.34065968494020876, + "learning_rate": 4.012005879792315e-06, + "loss": 0.4866, + "step": 3314 + }, + { + "epoch": 0.6073653352876511, + "grad_norm": 0.3589086007656234, + "learning_rate": 4.008871259417227e-06, + "loss": 0.4754, + "step": 3315 + }, + { + "epoch": 0.6075485525833638, + "grad_norm": 0.3734726010751245, + "learning_rate": 4.005737044469528e-06, + "loss": 0.4853, + "step": 3316 + }, + { + "epoch": 0.6077317698790766, + "grad_norm": 0.36392838196244376, + "learning_rate": 4.002603236231288e-06, + "loss": 0.4625, + "step": 3317 + }, + { + "epoch": 0.6079149871747893, + "grad_norm": 0.4033765708212323, + "learning_rate": 3.9994698359844085e-06, + "loss": 0.4829, + "step": 3318 + }, + { + "epoch": 0.608098204470502, + "grad_norm": 0.35439050998367794, + "learning_rate": 3.99633684501063e-06, + "loss": 0.4326, + "step": 3319 + }, + { + "epoch": 0.6082814217662147, + "grad_norm": 0.3439026295030526, + "learning_rate": 3.9932042645915224e-06, + "loss": 0.4982, + "step": 3320 + }, + { + "epoch": 0.6084646390619275, + "grad_norm": 0.3323169715795941, + "learning_rate": 3.990072096008484e-06, + "loss": 0.4636, + "step": 3321 + }, + { + "epoch": 0.6086478563576402, + "grad_norm": 0.3723440488888345, + "learning_rate": 3.9869403405427475e-06, + "loss": 0.4815, + "step": 3322 + }, + { + "epoch": 0.6088310736533529, + "grad_norm": 0.35460440636858903, + "learning_rate": 3.983808999475381e-06, + "loss": 0.436, + "step": 3323 + }, + { + "epoch": 0.6090142909490656, + "grad_norm": 0.4121519641073631, + "learning_rate": 3.980678074087277e-06, + "loss": 0.5019, + "step": 3324 + }, + { + "epoch": 0.6091975082447783, + "grad_norm": 0.38659222731348764, + "learning_rate": 3.9775475656591594e-06, + "loss": 0.4659, + "step": 3325 + }, + { + "epoch": 0.609380725540491, + "grad_norm": 0.3603288477385579, + "learning_rate": 3.974417475471583e-06, + "loss": 0.485, + "step": 3326 + }, + { + "epoch": 0.6095639428362037, + "grad_norm": 0.46183540556732916, + "learning_rate": 3.9712878048049284e-06, + "loss": 0.5004, + "step": 3327 + }, + { + "epoch": 0.6097471601319164, + "grad_norm": 0.3612919378513964, + "learning_rate": 3.968158554939412e-06, + "loss": 0.5009, + "step": 3328 + }, + { + "epoch": 0.6099303774276291, + "grad_norm": 0.3910922714795563, + "learning_rate": 3.965029727155069e-06, + "loss": 0.4867, + "step": 3329 + }, + { + "epoch": 0.6101135947233419, + "grad_norm": 0.39633836659584154, + "learning_rate": 3.961901322731766e-06, + "loss": 0.449, + "step": 3330 + }, + { + "epoch": 0.6102968120190546, + "grad_norm": 0.38115961908245777, + "learning_rate": 3.958773342949196e-06, + "loss": 0.4747, + "step": 3331 + }, + { + "epoch": 0.6104800293147673, + "grad_norm": 0.33284427746348605, + "learning_rate": 3.955645789086883e-06, + "loss": 0.4575, + "step": 3332 + }, + { + "epoch": 0.61066324661048, + "grad_norm": 0.3821677020949391, + "learning_rate": 3.952518662424167e-06, + "loss": 0.4858, + "step": 3333 + }, + { + "epoch": 0.6108464639061928, + "grad_norm": 0.4009014380114965, + "learning_rate": 3.949391964240221e-06, + "loss": 0.4318, + "step": 3334 + }, + { + "epoch": 0.6110296812019055, + "grad_norm": 0.44000938090181235, + "learning_rate": 3.94626569581404e-06, + "loss": 0.4779, + "step": 3335 + }, + { + "epoch": 0.6112128984976182, + "grad_norm": 0.40730570469567723, + "learning_rate": 3.943139858424445e-06, + "loss": 0.4716, + "step": 3336 + }, + { + "epoch": 0.6113961157933309, + "grad_norm": 0.33655966447673424, + "learning_rate": 3.940014453350076e-06, + "loss": 0.4832, + "step": 3337 + }, + { + "epoch": 0.6115793330890436, + "grad_norm": 0.360061717888351, + "learning_rate": 3.936889481869403e-06, + "loss": 0.4521, + "step": 3338 + }, + { + "epoch": 0.6117625503847564, + "grad_norm": 0.3121566288995717, + "learning_rate": 3.933764945260712e-06, + "loss": 0.4611, + "step": 3339 + }, + { + "epoch": 0.611945767680469, + "grad_norm": 0.3655686160599967, + "learning_rate": 3.930640844802119e-06, + "loss": 0.4792, + "step": 3340 + }, + { + "epoch": 0.6121289849761817, + "grad_norm": 0.3184169765542266, + "learning_rate": 3.92751718177155e-06, + "loss": 0.4598, + "step": 3341 + }, + { + "epoch": 0.6123122022718944, + "grad_norm": 0.3448183481186073, + "learning_rate": 3.9243939574467625e-06, + "loss": 0.4856, + "step": 3342 + }, + { + "epoch": 0.6124954195676072, + "grad_norm": 0.349808866231569, + "learning_rate": 3.9212711731053315e-06, + "loss": 0.448, + "step": 3343 + }, + { + "epoch": 0.6126786368633199, + "grad_norm": 0.3475677996342215, + "learning_rate": 3.918148830024651e-06, + "loss": 0.4667, + "step": 3344 + }, + { + "epoch": 0.6128618541590326, + "grad_norm": 0.3944639731431349, + "learning_rate": 3.915026929481935e-06, + "loss": 0.486, + "step": 3345 + }, + { + "epoch": 0.6130450714547453, + "grad_norm": 0.43288681581663946, + "learning_rate": 3.9119054727542135e-06, + "loss": 0.5017, + "step": 3346 + }, + { + "epoch": 0.6132282887504581, + "grad_norm": 0.3580534881016692, + "learning_rate": 3.9087844611183415e-06, + "loss": 0.4845, + "step": 3347 + }, + { + "epoch": 0.6134115060461708, + "grad_norm": 0.3705288832890555, + "learning_rate": 3.905663895850988e-06, + "loss": 0.4829, + "step": 3348 + }, + { + "epoch": 0.6135947233418835, + "grad_norm": 0.3711417703037963, + "learning_rate": 3.902543778228638e-06, + "loss": 0.4583, + "step": 3349 + }, + { + "epoch": 0.6137779406375962, + "grad_norm": 0.4010220854159373, + "learning_rate": 3.899424109527594e-06, + "loss": 0.4446, + "step": 3350 + }, + { + "epoch": 0.6139611579333089, + "grad_norm": 0.3700588095233422, + "learning_rate": 3.896304891023979e-06, + "loss": 0.4749, + "step": 3351 + }, + { + "epoch": 0.6141443752290217, + "grad_norm": 0.3863735168982707, + "learning_rate": 3.893186123993727e-06, + "loss": 0.4801, + "step": 3352 + }, + { + "epoch": 0.6143275925247343, + "grad_norm": 0.37734735039970213, + "learning_rate": 3.89006780971259e-06, + "loss": 0.4697, + "step": 3353 + }, + { + "epoch": 0.614510809820447, + "grad_norm": 0.43614331655717437, + "learning_rate": 3.886949949456133e-06, + "loss": 0.4826, + "step": 3354 + }, + { + "epoch": 0.6146940271161597, + "grad_norm": 0.41416844558813, + "learning_rate": 3.883832544499735e-06, + "loss": 0.4814, + "step": 3355 + }, + { + "epoch": 0.6148772444118725, + "grad_norm": 0.3840583437325919, + "learning_rate": 3.880715596118596e-06, + "loss": 0.4861, + "step": 3356 + }, + { + "epoch": 0.6150604617075852, + "grad_norm": 1.5827048716139795, + "learning_rate": 3.877599105587714e-06, + "loss": 0.461, + "step": 3357 + }, + { + "epoch": 0.6152436790032979, + "grad_norm": 0.33728519698992276, + "learning_rate": 3.874483074181915e-06, + "loss": 0.4361, + "step": 3358 + }, + { + "epoch": 0.6154268962990106, + "grad_norm": 0.3892940741219592, + "learning_rate": 3.871367503175828e-06, + "loss": 0.4823, + "step": 3359 + }, + { + "epoch": 0.6156101135947234, + "grad_norm": 0.3671721163790301, + "learning_rate": 3.868252393843902e-06, + "loss": 0.4699, + "step": 3360 + }, + { + "epoch": 0.6157933308904361, + "grad_norm": 0.3861159968344066, + "learning_rate": 3.865137747460385e-06, + "loss": 0.4588, + "step": 3361 + }, + { + "epoch": 0.6159765481861488, + "grad_norm": 0.38191850453923193, + "learning_rate": 3.862023565299346e-06, + "loss": 0.4608, + "step": 3362 + }, + { + "epoch": 0.6161597654818615, + "grad_norm": 0.33682639019253413, + "learning_rate": 3.858909848634661e-06, + "loss": 0.464, + "step": 3363 + }, + { + "epoch": 0.6163429827775742, + "grad_norm": 0.37092191775209565, + "learning_rate": 3.8557965987400156e-06, + "loss": 0.4619, + "step": 3364 + }, + { + "epoch": 0.616526200073287, + "grad_norm": 0.3878901717403647, + "learning_rate": 3.8526838168889e-06, + "loss": 0.4748, + "step": 3365 + }, + { + "epoch": 0.6167094173689996, + "grad_norm": 0.39928952055424216, + "learning_rate": 3.8495715043546225e-06, + "loss": 0.4841, + "step": 3366 + }, + { + "epoch": 0.6168926346647123, + "grad_norm": 0.36984083415160346, + "learning_rate": 3.84645966241029e-06, + "loss": 0.4509, + "step": 3367 + }, + { + "epoch": 0.617075851960425, + "grad_norm": 0.43143681897166014, + "learning_rate": 3.8433482923288245e-06, + "loss": 0.4926, + "step": 3368 + }, + { + "epoch": 0.6172590692561378, + "grad_norm": 0.39736331745243536, + "learning_rate": 3.840237395382947e-06, + "loss": 0.4296, + "step": 3369 + }, + { + "epoch": 0.6174422865518505, + "grad_norm": 0.3346071880701682, + "learning_rate": 3.837126972845192e-06, + "loss": 0.4655, + "step": 3370 + }, + { + "epoch": 0.6176255038475632, + "grad_norm": 0.3678217591892569, + "learning_rate": 3.8340170259878965e-06, + "loss": 0.4825, + "step": 3371 + }, + { + "epoch": 0.6178087211432759, + "grad_norm": 0.40832968867190783, + "learning_rate": 3.830907556083204e-06, + "loss": 0.4596, + "step": 3372 + }, + { + "epoch": 0.6179919384389886, + "grad_norm": 0.39028994718588056, + "learning_rate": 3.827798564403063e-06, + "loss": 0.4926, + "step": 3373 + }, + { + "epoch": 0.6181751557347014, + "grad_norm": 0.3727381875263331, + "learning_rate": 3.824690052219223e-06, + "loss": 0.4498, + "step": 3374 + }, + { + "epoch": 0.6183583730304141, + "grad_norm": 0.3770967571827475, + "learning_rate": 3.821582020803244e-06, + "loss": 0.4816, + "step": 3375 + }, + { + "epoch": 0.6185415903261268, + "grad_norm": 0.3677418395766839, + "learning_rate": 3.818474471426485e-06, + "loss": 0.457, + "step": 3376 + }, + { + "epoch": 0.6187248076218395, + "grad_norm": 0.4393531482364124, + "learning_rate": 3.815367405360106e-06, + "loss": 0.4665, + "step": 3377 + }, + { + "epoch": 0.6189080249175523, + "grad_norm": 0.3994840010566743, + "learning_rate": 3.8122608238750746e-06, + "loss": 0.5108, + "step": 3378 + }, + { + "epoch": 0.619091242213265, + "grad_norm": 0.3902946208866115, + "learning_rate": 3.8091547282421544e-06, + "loss": 0.4729, + "step": 3379 + }, + { + "epoch": 0.6192744595089776, + "grad_norm": 0.36047052014306097, + "learning_rate": 3.8060491197319166e-06, + "loss": 0.4782, + "step": 3380 + }, + { + "epoch": 0.6194576768046903, + "grad_norm": 0.3227185779888876, + "learning_rate": 3.802943999614727e-06, + "loss": 0.4816, + "step": 3381 + }, + { + "epoch": 0.6196408941004031, + "grad_norm": 0.38515368184924453, + "learning_rate": 3.799839369160756e-06, + "loss": 0.4968, + "step": 3382 + }, + { + "epoch": 0.6198241113961158, + "grad_norm": 0.40691464191289234, + "learning_rate": 3.7967352296399696e-06, + "loss": 0.4949, + "step": 3383 + }, + { + "epoch": 0.6200073286918285, + "grad_norm": 0.4483467977975031, + "learning_rate": 3.79363158232214e-06, + "loss": 0.4637, + "step": 3384 + }, + { + "epoch": 0.6201905459875412, + "grad_norm": 0.7314258477496398, + "learning_rate": 3.790528428476828e-06, + "loss": 0.4954, + "step": 3385 + }, + { + "epoch": 0.6203737632832539, + "grad_norm": 0.36327185678620344, + "learning_rate": 3.787425769373401e-06, + "loss": 0.4792, + "step": 3386 + }, + { + "epoch": 0.6205569805789667, + "grad_norm": 0.41644011702240735, + "learning_rate": 3.7843236062810216e-06, + "loss": 0.4883, + "step": 3387 + }, + { + "epoch": 0.6207401978746794, + "grad_norm": 0.3412670428933387, + "learning_rate": 3.7812219404686484e-06, + "loss": 0.4575, + "step": 3388 + }, + { + "epoch": 0.6209234151703921, + "grad_norm": 0.4512406484756083, + "learning_rate": 3.778120773205035e-06, + "loss": 0.4885, + "step": 3389 + }, + { + "epoch": 0.6211066324661048, + "grad_norm": 0.33786157328867994, + "learning_rate": 3.775020105758736e-06, + "loss": 0.4402, + "step": 3390 + }, + { + "epoch": 0.6212898497618176, + "grad_norm": 0.4166062568929626, + "learning_rate": 3.7719199393980977e-06, + "loss": 0.4514, + "step": 3391 + }, + { + "epoch": 0.6214730670575302, + "grad_norm": 0.41355321725132443, + "learning_rate": 3.7688202753912634e-06, + "loss": 0.4686, + "step": 3392 + }, + { + "epoch": 0.6216562843532429, + "grad_norm": 0.4047176800778511, + "learning_rate": 3.765721115006168e-06, + "loss": 0.4715, + "step": 3393 + }, + { + "epoch": 0.6218395016489556, + "grad_norm": 0.3799684823135988, + "learning_rate": 3.762622459510542e-06, + "loss": 0.4783, + "step": 3394 + }, + { + "epoch": 0.6220227189446684, + "grad_norm": 0.4117099383017388, + "learning_rate": 3.7595243101719126e-06, + "loss": 0.4791, + "step": 3395 + }, + { + "epoch": 0.6222059362403811, + "grad_norm": 0.32305546509666183, + "learning_rate": 3.7564266682575966e-06, + "loss": 0.4508, + "step": 3396 + }, + { + "epoch": 0.6223891535360938, + "grad_norm": 0.3747661592878273, + "learning_rate": 3.753329535034702e-06, + "loss": 0.4921, + "step": 3397 + }, + { + "epoch": 0.6225723708318065, + "grad_norm": 0.47414657007429, + "learning_rate": 3.7502329117701297e-06, + "loss": 0.4543, + "step": 3398 + }, + { + "epoch": 0.6227555881275192, + "grad_norm": 0.4200518871306099, + "learning_rate": 3.7471367997305764e-06, + "loss": 0.4675, + "step": 3399 + }, + { + "epoch": 0.622938805423232, + "grad_norm": 0.39776912270226433, + "learning_rate": 3.7440412001825238e-06, + "loss": 0.4877, + "step": 3400 + }, + { + "epoch": 0.6231220227189447, + "grad_norm": 0.33669101746510566, + "learning_rate": 3.7409461143922455e-06, + "loss": 0.4581, + "step": 3401 + }, + { + "epoch": 0.6233052400146574, + "grad_norm": 0.33992019895266523, + "learning_rate": 3.7378515436258056e-06, + "loss": 0.4611, + "step": 3402 + }, + { + "epoch": 0.6234884573103701, + "grad_norm": 0.34526963310385533, + "learning_rate": 3.7347574891490597e-06, + "loss": 0.4606, + "step": 3403 + }, + { + "epoch": 0.6236716746060829, + "grad_norm": 0.4031944830734456, + "learning_rate": 3.7316639522276493e-06, + "loss": 0.4615, + "step": 3404 + }, + { + "epoch": 0.6238548919017955, + "grad_norm": 0.37649232647240166, + "learning_rate": 3.7285709341270037e-06, + "loss": 0.4798, + "step": 3405 + }, + { + "epoch": 0.6240381091975082, + "grad_norm": 0.3303406481230134, + "learning_rate": 3.725478436112342e-06, + "loss": 0.4649, + "step": 3406 + }, + { + "epoch": 0.6242213264932209, + "grad_norm": 0.36610882261819505, + "learning_rate": 3.722386459448668e-06, + "loss": 0.456, + "step": 3407 + }, + { + "epoch": 0.6244045437889337, + "grad_norm": 0.33772515081786897, + "learning_rate": 3.719295005400779e-06, + "loss": 0.4896, + "step": 3408 + }, + { + "epoch": 0.6245877610846464, + "grad_norm": 0.3848976211487813, + "learning_rate": 3.7162040752332485e-06, + "loss": 0.4702, + "step": 3409 + }, + { + "epoch": 0.6247709783803591, + "grad_norm": 0.46617888665168655, + "learning_rate": 3.713113670210443e-06, + "loss": 0.4844, + "step": 3410 + }, + { + "epoch": 0.6249541956760718, + "grad_norm": 0.38255069185326285, + "learning_rate": 3.71002379159651e-06, + "loss": 0.4859, + "step": 3411 + }, + { + "epoch": 0.6251374129717845, + "grad_norm": 0.3539583241238941, + "learning_rate": 3.706934440655389e-06, + "loss": 0.4622, + "step": 3412 + }, + { + "epoch": 0.6253206302674973, + "grad_norm": 0.4089075771262958, + "learning_rate": 3.70384561865079e-06, + "loss": 0.4575, + "step": 3413 + }, + { + "epoch": 0.62550384756321, + "grad_norm": 0.3862421415619821, + "learning_rate": 3.700757326846221e-06, + "loss": 0.4833, + "step": 3414 + }, + { + "epoch": 0.6256870648589227, + "grad_norm": 0.3643763570223953, + "learning_rate": 3.6976695665049656e-06, + "loss": 0.4723, + "step": 3415 + }, + { + "epoch": 0.6258702821546354, + "grad_norm": 0.3530022854685333, + "learning_rate": 3.6945823388900913e-06, + "loss": 0.4733, + "step": 3416 + }, + { + "epoch": 0.6260534994503482, + "grad_norm": 0.33522794524320454, + "learning_rate": 3.691495645264446e-06, + "loss": 0.4475, + "step": 3417 + }, + { + "epoch": 0.6262367167460609, + "grad_norm": 0.33496744151540203, + "learning_rate": 3.688409486890664e-06, + "loss": 0.4587, + "step": 3418 + }, + { + "epoch": 0.6264199340417735, + "grad_norm": 0.40647539022082196, + "learning_rate": 3.6853238650311553e-06, + "loss": 0.4946, + "step": 3419 + }, + { + "epoch": 0.6266031513374862, + "grad_norm": 0.39420793377221275, + "learning_rate": 3.6822387809481155e-06, + "loss": 0.4438, + "step": 3420 + }, + { + "epoch": 0.6267863686331989, + "grad_norm": 0.4087705564514289, + "learning_rate": 3.679154235903514e-06, + "loss": 0.4783, + "step": 3421 + }, + { + "epoch": 0.6269695859289117, + "grad_norm": 0.3849230704723878, + "learning_rate": 3.6760702311591032e-06, + "loss": 0.4666, + "step": 3422 + }, + { + "epoch": 0.6271528032246244, + "grad_norm": 0.3743776466128573, + "learning_rate": 3.6729867679764183e-06, + "loss": 0.4568, + "step": 3423 + }, + { + "epoch": 0.6273360205203371, + "grad_norm": 0.3626507057042326, + "learning_rate": 3.669903847616767e-06, + "loss": 0.477, + "step": 3424 + }, + { + "epoch": 0.6275192378160498, + "grad_norm": 0.3768683693381953, + "learning_rate": 3.6668214713412364e-06, + "loss": 0.4752, + "step": 3425 + }, + { + "epoch": 0.6277024551117626, + "grad_norm": 0.372956492863803, + "learning_rate": 3.6637396404106907e-06, + "loss": 0.471, + "step": 3426 + }, + { + "epoch": 0.6278856724074753, + "grad_norm": 0.3623634477467274, + "learning_rate": 3.6606583560857757e-06, + "loss": 0.4691, + "step": 3427 + }, + { + "epoch": 0.628068889703188, + "grad_norm": 0.3654773740807239, + "learning_rate": 3.6575776196269075e-06, + "loss": 0.4589, + "step": 3428 + }, + { + "epoch": 0.6282521069989007, + "grad_norm": 0.3599616685715066, + "learning_rate": 3.654497432294281e-06, + "loss": 0.5018, + "step": 3429 + }, + { + "epoch": 0.6284353242946135, + "grad_norm": 0.3706421176656778, + "learning_rate": 3.6514177953478656e-06, + "loss": 0.4775, + "step": 3430 + }, + { + "epoch": 0.6286185415903262, + "grad_norm": 0.35829124434103354, + "learning_rate": 3.6483387100474054e-06, + "loss": 0.4721, + "step": 3431 + }, + { + "epoch": 0.6288017588860388, + "grad_norm": 0.39984663977682633, + "learning_rate": 3.645260177652422e-06, + "loss": 0.4807, + "step": 3432 + }, + { + "epoch": 0.6289849761817515, + "grad_norm": 0.35025444185117444, + "learning_rate": 3.642182199422206e-06, + "loss": 0.4951, + "step": 3433 + }, + { + "epoch": 0.6291681934774642, + "grad_norm": 0.36671047481915986, + "learning_rate": 3.6391047766158243e-06, + "loss": 0.4839, + "step": 3434 + }, + { + "epoch": 0.629351410773177, + "grad_norm": 0.5064340298169169, + "learning_rate": 3.636027910492115e-06, + "loss": 0.471, + "step": 3435 + }, + { + "epoch": 0.6295346280688897, + "grad_norm": 0.3346809459394731, + "learning_rate": 3.632951602309693e-06, + "loss": 0.4585, + "step": 3436 + }, + { + "epoch": 0.6297178453646024, + "grad_norm": 0.3666266324043109, + "learning_rate": 3.629875853326935e-06, + "loss": 0.4685, + "step": 3437 + }, + { + "epoch": 0.6299010626603151, + "grad_norm": 0.42734384578850443, + "learning_rate": 3.6268006648020005e-06, + "loss": 0.4706, + "step": 3438 + }, + { + "epoch": 0.6300842799560279, + "grad_norm": 0.36277576315958787, + "learning_rate": 3.623726037992811e-06, + "loss": 0.494, + "step": 3439 + }, + { + "epoch": 0.6302674972517406, + "grad_norm": 0.3638105357596318, + "learning_rate": 3.6206519741570674e-06, + "loss": 0.4579, + "step": 3440 + }, + { + "epoch": 0.6304507145474533, + "grad_norm": 0.36202131112274694, + "learning_rate": 3.6175784745522273e-06, + "loss": 0.4369, + "step": 3441 + }, + { + "epoch": 0.630633931843166, + "grad_norm": 0.4353293472651712, + "learning_rate": 3.61450554043553e-06, + "loss": 0.4778, + "step": 3442 + }, + { + "epoch": 0.6308171491388788, + "grad_norm": 0.34366143901775337, + "learning_rate": 3.611433173063976e-06, + "loss": 0.4438, + "step": 3443 + }, + { + "epoch": 0.6310003664345915, + "grad_norm": 0.3758857101998102, + "learning_rate": 3.608361373694339e-06, + "loss": 0.4522, + "step": 3444 + }, + { + "epoch": 0.6311835837303041, + "grad_norm": 0.3524910596291595, + "learning_rate": 3.6052901435831533e-06, + "loss": 0.46, + "step": 3445 + }, + { + "epoch": 0.6313668010260168, + "grad_norm": 0.3326251240922405, + "learning_rate": 3.6022194839867282e-06, + "loss": 0.4423, + "step": 3446 + }, + { + "epoch": 0.6315500183217295, + "grad_norm": 0.4166545903844291, + "learning_rate": 3.5991493961611353e-06, + "loss": 0.4939, + "step": 3447 + }, + { + "epoch": 0.6317332356174423, + "grad_norm": 0.3961400544630863, + "learning_rate": 3.5960798813622137e-06, + "loss": 0.4324, + "step": 3448 + }, + { + "epoch": 0.631916452913155, + "grad_norm": 0.3674974134485457, + "learning_rate": 3.593010940845566e-06, + "loss": 0.4556, + "step": 3449 + }, + { + "epoch": 0.6320996702088677, + "grad_norm": 0.3890522202378581, + "learning_rate": 3.589942575866562e-06, + "loss": 0.4673, + "step": 3450 + }, + { + "epoch": 0.6322828875045804, + "grad_norm": 0.3756290012411645, + "learning_rate": 3.5868747876803356e-06, + "loss": 0.4464, + "step": 3451 + }, + { + "epoch": 0.6324661048002932, + "grad_norm": 0.4087829017495404, + "learning_rate": 3.583807577541787e-06, + "loss": 0.4888, + "step": 3452 + }, + { + "epoch": 0.6326493220960059, + "grad_norm": 2.482123269569229, + "learning_rate": 3.5807409467055737e-06, + "loss": 0.4595, + "step": 3453 + }, + { + "epoch": 0.6328325393917186, + "grad_norm": 0.5554873483062074, + "learning_rate": 3.577674896426121e-06, + "loss": 0.4768, + "step": 3454 + }, + { + "epoch": 0.6330157566874313, + "grad_norm": 0.44365336073816447, + "learning_rate": 3.5746094279576182e-06, + "loss": 0.4748, + "step": 3455 + }, + { + "epoch": 0.633198973983144, + "grad_norm": 0.4146770903766801, + "learning_rate": 3.571544542554014e-06, + "loss": 0.4485, + "step": 3456 + }, + { + "epoch": 0.6333821912788568, + "grad_norm": 0.3745085045702995, + "learning_rate": 3.5684802414690174e-06, + "loss": 0.4661, + "step": 3457 + }, + { + "epoch": 0.6335654085745694, + "grad_norm": 0.3783881705015424, + "learning_rate": 3.565416525956099e-06, + "loss": 0.4588, + "step": 3458 + }, + { + "epoch": 0.6337486258702821, + "grad_norm": 0.3468524594706612, + "learning_rate": 3.562353397268492e-06, + "loss": 0.4306, + "step": 3459 + }, + { + "epoch": 0.6339318431659948, + "grad_norm": 0.34250416542600726, + "learning_rate": 3.55929085665919e-06, + "loss": 0.4957, + "step": 3460 + }, + { + "epoch": 0.6341150604617076, + "grad_norm": 0.38691657059403695, + "learning_rate": 3.5562289053809417e-06, + "loss": 0.4711, + "step": 3461 + }, + { + "epoch": 0.6342982777574203, + "grad_norm": 0.40934344680238033, + "learning_rate": 3.5531675446862586e-06, + "loss": 0.4638, + "step": 3462 + }, + { + "epoch": 0.634481495053133, + "grad_norm": 0.3906074957726188, + "learning_rate": 3.550106775827408e-06, + "loss": 0.4534, + "step": 3463 + }, + { + "epoch": 0.6346647123488457, + "grad_norm": 0.40424103444392073, + "learning_rate": 3.5470466000564197e-06, + "loss": 0.457, + "step": 3464 + }, + { + "epoch": 0.6348479296445585, + "grad_norm": 0.37135572208067924, + "learning_rate": 3.543987018625073e-06, + "loss": 0.4849, + "step": 3465 + }, + { + "epoch": 0.6350311469402712, + "grad_norm": 0.34952897683939, + "learning_rate": 3.5409280327849123e-06, + "loss": 0.469, + "step": 3466 + }, + { + "epoch": 0.6352143642359839, + "grad_norm": 0.3613897105517585, + "learning_rate": 3.5378696437872334e-06, + "loss": 0.4524, + "step": 3467 + }, + { + "epoch": 0.6353975815316966, + "grad_norm": 0.3527825682366731, + "learning_rate": 3.534811852883091e-06, + "loss": 0.4426, + "step": 3468 + }, + { + "epoch": 0.6355807988274093, + "grad_norm": 0.413745130417646, + "learning_rate": 3.5317546613232904e-06, + "loss": 0.4892, + "step": 3469 + }, + { + "epoch": 0.635764016123122, + "grad_norm": 0.36217762970317147, + "learning_rate": 3.528698070358398e-06, + "loss": 0.4385, + "step": 3470 + }, + { + "epoch": 0.6359472334188347, + "grad_norm": 0.3873199425818144, + "learning_rate": 3.5256420812387293e-06, + "loss": 0.487, + "step": 3471 + }, + { + "epoch": 0.6361304507145474, + "grad_norm": 0.36887459986925014, + "learning_rate": 3.5225866952143583e-06, + "loss": 0.4839, + "step": 3472 + }, + { + "epoch": 0.6363136680102601, + "grad_norm": 0.3762685573891245, + "learning_rate": 3.519531913535107e-06, + "loss": 0.4847, + "step": 3473 + }, + { + "epoch": 0.6364968853059729, + "grad_norm": 0.3876597950648676, + "learning_rate": 3.5164777374505522e-06, + "loss": 0.4654, + "step": 3474 + }, + { + "epoch": 0.6366801026016856, + "grad_norm": 0.389499620690064, + "learning_rate": 3.5134241682100266e-06, + "loss": 0.4755, + "step": 3475 + }, + { + "epoch": 0.6368633198973983, + "grad_norm": 0.38419318511889156, + "learning_rate": 3.5103712070626116e-06, + "loss": 0.4173, + "step": 3476 + }, + { + "epoch": 0.637046537193111, + "grad_norm": 0.37344769813877693, + "learning_rate": 3.507318855257137e-06, + "loss": 0.457, + "step": 3477 + }, + { + "epoch": 0.6372297544888238, + "grad_norm": 0.35605422789815827, + "learning_rate": 3.504267114042188e-06, + "loss": 0.4651, + "step": 3478 + }, + { + "epoch": 0.6374129717845365, + "grad_norm": 0.35216333564639746, + "learning_rate": 3.5012159846660987e-06, + "loss": 0.4735, + "step": 3479 + }, + { + "epoch": 0.6375961890802492, + "grad_norm": 0.3684926279502396, + "learning_rate": 3.4981654683769537e-06, + "loss": 0.491, + "step": 3480 + }, + { + "epoch": 0.6377794063759619, + "grad_norm": 0.6416320155444767, + "learning_rate": 3.4951155664225845e-06, + "loss": 0.5109, + "step": 3481 + }, + { + "epoch": 0.6379626236716746, + "grad_norm": 0.3311283706147146, + "learning_rate": 3.4920662800505713e-06, + "loss": 0.4951, + "step": 3482 + }, + { + "epoch": 0.6381458409673874, + "grad_norm": 0.38636678573281646, + "learning_rate": 3.4890176105082436e-06, + "loss": 0.4392, + "step": 3483 + }, + { + "epoch": 0.6383290582631, + "grad_norm": 0.4163698242292531, + "learning_rate": 3.4859695590426824e-06, + "loss": 0.4961, + "step": 3484 + }, + { + "epoch": 0.6385122755588127, + "grad_norm": 0.4359808520953443, + "learning_rate": 3.482922126900708e-06, + "loss": 0.4831, + "step": 3485 + }, + { + "epoch": 0.6386954928545254, + "grad_norm": 0.34637257759709444, + "learning_rate": 3.4798753153288934e-06, + "loss": 0.4699, + "step": 3486 + }, + { + "epoch": 0.6388787101502382, + "grad_norm": 0.37900559292552355, + "learning_rate": 3.4768291255735544e-06, + "loss": 0.4636, + "step": 3487 + }, + { + "epoch": 0.6390619274459509, + "grad_norm": 0.3538772725438127, + "learning_rate": 3.4737835588807584e-06, + "loss": 0.4595, + "step": 3488 + }, + { + "epoch": 0.6392451447416636, + "grad_norm": 0.43757235936032834, + "learning_rate": 3.470738616496307e-06, + "loss": 0.4783, + "step": 3489 + }, + { + "epoch": 0.6394283620373763, + "grad_norm": 0.40956563573499927, + "learning_rate": 3.467694299665757e-06, + "loss": 0.5142, + "step": 3490 + }, + { + "epoch": 0.6396115793330891, + "grad_norm": 0.37251978174694766, + "learning_rate": 3.464650609634403e-06, + "loss": 0.4931, + "step": 3491 + }, + { + "epoch": 0.6397947966288018, + "grad_norm": 0.35548169209825403, + "learning_rate": 3.46160754764729e-06, + "loss": 0.4645, + "step": 3492 + }, + { + "epoch": 0.6399780139245145, + "grad_norm": 0.31268147037920563, + "learning_rate": 3.4585651149491937e-06, + "loss": 0.4639, + "step": 3493 + }, + { + "epoch": 0.6401612312202272, + "grad_norm": 0.37157686849902105, + "learning_rate": 3.4555233127846465e-06, + "loss": 0.4857, + "step": 3494 + }, + { + "epoch": 0.6403444485159399, + "grad_norm": 0.3590054816908552, + "learning_rate": 3.452482142397915e-06, + "loss": 0.4513, + "step": 3495 + }, + { + "epoch": 0.6405276658116527, + "grad_norm": 0.3797523101447906, + "learning_rate": 3.449441605033009e-06, + "loss": 0.4321, + "step": 3496 + }, + { + "epoch": 0.6407108831073653, + "grad_norm": 0.3783050064767098, + "learning_rate": 3.446401701933676e-06, + "loss": 0.4593, + "step": 3497 + }, + { + "epoch": 0.640894100403078, + "grad_norm": 0.3115627069008165, + "learning_rate": 3.4433624343434115e-06, + "loss": 0.4241, + "step": 3498 + }, + { + "epoch": 0.6410773176987907, + "grad_norm": 0.3799429242374115, + "learning_rate": 3.4403238035054453e-06, + "loss": 0.4559, + "step": 3499 + }, + { + "epoch": 0.6412605349945035, + "grad_norm": 0.4240815344192379, + "learning_rate": 3.437285810662749e-06, + "loss": 0.4516, + "step": 3500 + }, + { + "epoch": 0.6414437522902162, + "grad_norm": 0.3711412884805633, + "learning_rate": 3.4342484570580315e-06, + "loss": 0.4429, + "step": 3501 + }, + { + "epoch": 0.6416269695859289, + "grad_norm": 0.4056211302748523, + "learning_rate": 3.43121174393374e-06, + "loss": 0.5073, + "step": 3502 + }, + { + "epoch": 0.6418101868816416, + "grad_norm": 0.43407514191028274, + "learning_rate": 3.4281756725320636e-06, + "loss": 0.4398, + "step": 3503 + }, + { + "epoch": 0.6419934041773543, + "grad_norm": 0.4211158133777933, + "learning_rate": 3.4251402440949256e-06, + "loss": 0.5037, + "step": 3504 + }, + { + "epoch": 0.6421766214730671, + "grad_norm": 0.445147601987858, + "learning_rate": 3.4221054598639857e-06, + "loss": 0.4782, + "step": 3505 + }, + { + "epoch": 0.6423598387687798, + "grad_norm": 0.35797330776687786, + "learning_rate": 3.4190713210806414e-06, + "loss": 0.4571, + "step": 3506 + }, + { + "epoch": 0.6425430560644925, + "grad_norm": 0.37382480133879875, + "learning_rate": 3.4160378289860272e-06, + "loss": 0.4766, + "step": 3507 + }, + { + "epoch": 0.6427262733602052, + "grad_norm": 0.3989667379416838, + "learning_rate": 3.4130049848210134e-06, + "loss": 0.4831, + "step": 3508 + }, + { + "epoch": 0.642909490655918, + "grad_norm": 0.3166260953679194, + "learning_rate": 3.4099727898262004e-06, + "loss": 0.4699, + "step": 3509 + }, + { + "epoch": 0.6430927079516306, + "grad_norm": 0.3385192585501824, + "learning_rate": 3.406941245241927e-06, + "loss": 0.4507, + "step": 3510 + }, + { + "epoch": 0.6432759252473433, + "grad_norm": 0.36839975936953884, + "learning_rate": 3.403910352308266e-06, + "loss": 0.4557, + "step": 3511 + }, + { + "epoch": 0.643459142543056, + "grad_norm": 0.3911811814121315, + "learning_rate": 3.400880112265026e-06, + "loss": 0.4842, + "step": 3512 + }, + { + "epoch": 0.6436423598387688, + "grad_norm": 0.41615345423216277, + "learning_rate": 3.397850526351741e-06, + "loss": 0.4493, + "step": 3513 + }, + { + "epoch": 0.6438255771344815, + "grad_norm": 0.3840788849888087, + "learning_rate": 3.3948215958076842e-06, + "loss": 0.4376, + "step": 3514 + }, + { + "epoch": 0.6440087944301942, + "grad_norm": 0.4041537256726371, + "learning_rate": 3.391793321871857e-06, + "loss": 0.4532, + "step": 3515 + }, + { + "epoch": 0.6441920117259069, + "grad_norm": 0.3970830239384753, + "learning_rate": 3.388765705782998e-06, + "loss": 0.46, + "step": 3516 + }, + { + "epoch": 0.6443752290216196, + "grad_norm": 0.385585925897048, + "learning_rate": 3.3857387487795668e-06, + "loss": 0.5081, + "step": 3517 + }, + { + "epoch": 0.6445584463173324, + "grad_norm": 0.3328117538262262, + "learning_rate": 3.3827124520997628e-06, + "loss": 0.4829, + "step": 3518 + }, + { + "epoch": 0.6447416636130451, + "grad_norm": 0.36362937914832516, + "learning_rate": 3.37968681698151e-06, + "loss": 0.4792, + "step": 3519 + }, + { + "epoch": 0.6449248809087578, + "grad_norm": 0.4213735741090599, + "learning_rate": 3.3766618446624656e-06, + "loss": 0.4976, + "step": 3520 + }, + { + "epoch": 0.6451080982044705, + "grad_norm": 0.3881846699021779, + "learning_rate": 3.3736375363800078e-06, + "loss": 0.5278, + "step": 3521 + }, + { + "epoch": 0.6452913155001833, + "grad_norm": 0.32839914542503323, + "learning_rate": 3.3706138933712544e-06, + "loss": 0.4845, + "step": 3522 + }, + { + "epoch": 0.645474532795896, + "grad_norm": 0.36179370255331816, + "learning_rate": 3.3675909168730435e-06, + "loss": 0.4542, + "step": 3523 + }, + { + "epoch": 0.6456577500916086, + "grad_norm": 0.3944701031233938, + "learning_rate": 3.364568608121943e-06, + "loss": 0.4604, + "step": 3524 + }, + { + "epoch": 0.6458409673873213, + "grad_norm": 0.35637552896164815, + "learning_rate": 3.3615469683542457e-06, + "loss": 0.4697, + "step": 3525 + }, + { + "epoch": 0.6460241846830341, + "grad_norm": 0.47042065723709525, + "learning_rate": 3.358525998805972e-06, + "loss": 0.4716, + "step": 3526 + }, + { + "epoch": 0.6462074019787468, + "grad_norm": 0.40080911870245983, + "learning_rate": 3.3555057007128704e-06, + "loss": 0.461, + "step": 3527 + }, + { + "epoch": 0.6463906192744595, + "grad_norm": 0.41348480024307604, + "learning_rate": 3.3524860753104127e-06, + "loss": 0.4869, + "step": 3528 + }, + { + "epoch": 0.6465738365701722, + "grad_norm": 0.32649611909430737, + "learning_rate": 3.3494671238337935e-06, + "loss": 0.4712, + "step": 3529 + }, + { + "epoch": 0.6467570538658849, + "grad_norm": 0.38662764850068554, + "learning_rate": 3.346448847517933e-06, + "loss": 0.4644, + "step": 3530 + }, + { + "epoch": 0.6469402711615977, + "grad_norm": 0.39143276965356133, + "learning_rate": 3.343431247597479e-06, + "loss": 0.4836, + "step": 3531 + }, + { + "epoch": 0.6471234884573104, + "grad_norm": 0.3509381014530727, + "learning_rate": 3.3404143253067987e-06, + "loss": 0.4735, + "step": 3532 + }, + { + "epoch": 0.6473067057530231, + "grad_norm": 0.3796827025033695, + "learning_rate": 3.3373980818799823e-06, + "loss": 0.4604, + "step": 3533 + }, + { + "epoch": 0.6474899230487358, + "grad_norm": 0.3940374501209101, + "learning_rate": 3.334382518550841e-06, + "loss": 0.4571, + "step": 3534 + }, + { + "epoch": 0.6476731403444486, + "grad_norm": 0.29899833922492824, + "learning_rate": 3.3313676365529134e-06, + "loss": 0.4666, + "step": 3535 + }, + { + "epoch": 0.6478563576401613, + "grad_norm": 0.3973588990281339, + "learning_rate": 3.3283534371194547e-06, + "loss": 0.5021, + "step": 3536 + }, + { + "epoch": 0.6480395749358739, + "grad_norm": 0.39772748029318056, + "learning_rate": 3.3253399214834407e-06, + "loss": 0.4636, + "step": 3537 + }, + { + "epoch": 0.6482227922315866, + "grad_norm": 0.41868905268098267, + "learning_rate": 3.3223270908775695e-06, + "loss": 0.5021, + "step": 3538 + }, + { + "epoch": 0.6484060095272994, + "grad_norm": 0.4279368003691416, + "learning_rate": 3.319314946534258e-06, + "loss": 0.4957, + "step": 3539 + }, + { + "epoch": 0.6485892268230121, + "grad_norm": 0.36477344049855465, + "learning_rate": 3.3163034896856437e-06, + "loss": 0.4229, + "step": 3540 + }, + { + "epoch": 0.6487724441187248, + "grad_norm": 0.36051725460825584, + "learning_rate": 3.31329272156358e-06, + "loss": 0.4534, + "step": 3541 + }, + { + "epoch": 0.6489556614144375, + "grad_norm": 0.6014115157122704, + "learning_rate": 3.3102826433996417e-06, + "loss": 0.4708, + "step": 3542 + }, + { + "epoch": 0.6491388787101502, + "grad_norm": 0.39799589744473024, + "learning_rate": 3.3072732564251187e-06, + "loss": 0.463, + "step": 3543 + }, + { + "epoch": 0.649322096005863, + "grad_norm": 0.38344865304165143, + "learning_rate": 3.3042645618710234e-06, + "loss": 0.4727, + "step": 3544 + }, + { + "epoch": 0.6495053133015757, + "grad_norm": 0.34199663089769966, + "learning_rate": 3.3012565609680752e-06, + "loss": 0.4553, + "step": 3545 + }, + { + "epoch": 0.6496885305972884, + "grad_norm": 0.35105709232651594, + "learning_rate": 3.298249254946719e-06, + "loss": 0.4515, + "step": 3546 + }, + { + "epoch": 0.6498717478930011, + "grad_norm": 0.3734304022681076, + "learning_rate": 3.295242645037112e-06, + "loss": 0.4682, + "step": 3547 + }, + { + "epoch": 0.6500549651887139, + "grad_norm": 0.3771927191390024, + "learning_rate": 3.2922367324691286e-06, + "loss": 0.4694, + "step": 3548 + }, + { + "epoch": 0.6502381824844266, + "grad_norm": 0.4397759262392246, + "learning_rate": 3.28923151847235e-06, + "loss": 0.4586, + "step": 3549 + }, + { + "epoch": 0.6504213997801392, + "grad_norm": 0.3811390491048839, + "learning_rate": 3.286227004276083e-06, + "loss": 0.5015, + "step": 3550 + }, + { + "epoch": 0.6506046170758519, + "grad_norm": 0.4652989704884051, + "learning_rate": 3.2832231911093415e-06, + "loss": 0.4867, + "step": 3551 + }, + { + "epoch": 0.6507878343715646, + "grad_norm": 0.39484311913877573, + "learning_rate": 3.2802200802008548e-06, + "loss": 0.467, + "step": 3552 + }, + { + "epoch": 0.6509710516672774, + "grad_norm": 0.5584032248553762, + "learning_rate": 3.2772176727790622e-06, + "loss": 0.4942, + "step": 3553 + }, + { + "epoch": 0.6511542689629901, + "grad_norm": 0.38048902784266914, + "learning_rate": 3.2742159700721173e-06, + "loss": 0.4543, + "step": 3554 + }, + { + "epoch": 0.6513374862587028, + "grad_norm": 0.37238737216175505, + "learning_rate": 3.271214973307887e-06, + "loss": 0.4878, + "step": 3555 + }, + { + "epoch": 0.6515207035544155, + "grad_norm": 0.4323784650767414, + "learning_rate": 3.268214683713948e-06, + "loss": 0.4637, + "step": 3556 + }, + { + "epoch": 0.6517039208501283, + "grad_norm": 0.3696334672896221, + "learning_rate": 3.265215102517585e-06, + "loss": 0.4797, + "step": 3557 + }, + { + "epoch": 0.651887138145841, + "grad_norm": 0.9014011930146032, + "learning_rate": 3.2622162309457962e-06, + "loss": 0.4865, + "step": 3558 + }, + { + "epoch": 0.6520703554415537, + "grad_norm": 0.34395774534535334, + "learning_rate": 3.2592180702252898e-06, + "loss": 0.4843, + "step": 3559 + }, + { + "epoch": 0.6522535727372664, + "grad_norm": 0.42243003944197993, + "learning_rate": 3.2562206215824834e-06, + "loss": 0.4988, + "step": 3560 + }, + { + "epoch": 0.6524367900329792, + "grad_norm": 0.37577464688011614, + "learning_rate": 3.2532238862434983e-06, + "loss": 0.4577, + "step": 3561 + }, + { + "epoch": 0.6526200073286919, + "grad_norm": 0.39477280886735666, + "learning_rate": 3.25022786543417e-06, + "loss": 0.4949, + "step": 3562 + }, + { + "epoch": 0.6528032246244045, + "grad_norm": 0.4098335180603605, + "learning_rate": 3.2472325603800382e-06, + "loss": 0.4532, + "step": 3563 + }, + { + "epoch": 0.6529864419201172, + "grad_norm": 0.43194634826511535, + "learning_rate": 3.2442379723063543e-06, + "loss": 0.4885, + "step": 3564 + }, + { + "epoch": 0.6531696592158299, + "grad_norm": 0.40157330821237697, + "learning_rate": 3.2412441024380703e-06, + "loss": 0.4859, + "step": 3565 + }, + { + "epoch": 0.6533528765115427, + "grad_norm": 0.3402369096953985, + "learning_rate": 3.2382509519998475e-06, + "loss": 0.4821, + "step": 3566 + }, + { + "epoch": 0.6535360938072554, + "grad_norm": 0.4132618314549237, + "learning_rate": 3.235258522216052e-06, + "loss": 0.4822, + "step": 3567 + }, + { + "epoch": 0.6537193111029681, + "grad_norm": 0.3375238402367875, + "learning_rate": 3.2322668143107604e-06, + "loss": 0.468, + "step": 3568 + }, + { + "epoch": 0.6539025283986808, + "grad_norm": 0.3787445578705219, + "learning_rate": 3.229275829507743e-06, + "loss": 0.4399, + "step": 3569 + }, + { + "epoch": 0.6540857456943936, + "grad_norm": 0.39596344732551964, + "learning_rate": 3.2262855690304855e-06, + "loss": 0.4981, + "step": 3570 + }, + { + "epoch": 0.6542689629901063, + "grad_norm": 0.3513250357882315, + "learning_rate": 3.2232960341021703e-06, + "loss": 0.4817, + "step": 3571 + }, + { + "epoch": 0.654452180285819, + "grad_norm": 0.35567303592424276, + "learning_rate": 3.220307225945688e-06, + "loss": 0.4722, + "step": 3572 + }, + { + "epoch": 0.6546353975815317, + "grad_norm": 0.3809360317701113, + "learning_rate": 3.217319145783625e-06, + "loss": 0.4885, + "step": 3573 + }, + { + "epoch": 0.6548186148772445, + "grad_norm": 0.36407113344124864, + "learning_rate": 3.214331794838277e-06, + "loss": 0.4599, + "step": 3574 + }, + { + "epoch": 0.6550018321729572, + "grad_norm": 0.3845152790203452, + "learning_rate": 3.211345174331638e-06, + "loss": 0.4819, + "step": 3575 + }, + { + "epoch": 0.6551850494686698, + "grad_norm": 0.3768758953198083, + "learning_rate": 3.208359285485405e-06, + "loss": 0.468, + "step": 3576 + }, + { + "epoch": 0.6553682667643825, + "grad_norm": 0.426156243855698, + "learning_rate": 3.2053741295209705e-06, + "loss": 0.4783, + "step": 3577 + }, + { + "epoch": 0.6555514840600952, + "grad_norm": 0.37819437925881616, + "learning_rate": 3.202389707659435e-06, + "loss": 0.4613, + "step": 3578 + }, + { + "epoch": 0.655734701355808, + "grad_norm": 0.36748352745726787, + "learning_rate": 3.1994060211215936e-06, + "loss": 0.514, + "step": 3579 + }, + { + "epoch": 0.6559179186515207, + "grad_norm": 0.3376691005467414, + "learning_rate": 3.196423071127943e-06, + "loss": 0.4438, + "step": 3580 + }, + { + "epoch": 0.6561011359472334, + "grad_norm": 0.34985685550276574, + "learning_rate": 3.1934408588986753e-06, + "loss": 0.4761, + "step": 3581 + }, + { + "epoch": 0.6562843532429461, + "grad_norm": 0.37620198908100855, + "learning_rate": 3.1904593856536826e-06, + "loss": 0.5157, + "step": 3582 + }, + { + "epoch": 0.6564675705386589, + "grad_norm": 0.3612050772053323, + "learning_rate": 3.1874786526125577e-06, + "loss": 0.443, + "step": 3583 + }, + { + "epoch": 0.6566507878343716, + "grad_norm": 0.3536814264575201, + "learning_rate": 3.184498660994588e-06, + "loss": 0.474, + "step": 3584 + }, + { + "epoch": 0.6568340051300843, + "grad_norm": 0.3534674258185699, + "learning_rate": 3.1815194120187555e-06, + "loss": 0.4692, + "step": 3585 + }, + { + "epoch": 0.657017222425797, + "grad_norm": 0.3834255763758721, + "learning_rate": 3.178540906903741e-06, + "loss": 0.4943, + "step": 3586 + }, + { + "epoch": 0.6572004397215098, + "grad_norm": 0.34331721059853854, + "learning_rate": 3.1755631468679217e-06, + "loss": 0.4762, + "step": 3587 + }, + { + "epoch": 0.6573836570172225, + "grad_norm": 0.3450853014221691, + "learning_rate": 3.1725861331293694e-06, + "loss": 0.466, + "step": 3588 + }, + { + "epoch": 0.6575668743129351, + "grad_norm": 0.3894544344884122, + "learning_rate": 3.1696098669058485e-06, + "loss": 0.464, + "step": 3589 + }, + { + "epoch": 0.6577500916086478, + "grad_norm": 0.3259721919096669, + "learning_rate": 3.1666343494148194e-06, + "loss": 0.4508, + "step": 3590 + }, + { + "epoch": 0.6579333089043605, + "grad_norm": 0.3847875875663655, + "learning_rate": 3.1636595818734355e-06, + "loss": 0.4736, + "step": 3591 + }, + { + "epoch": 0.6581165262000733, + "grad_norm": 0.38253641283684975, + "learning_rate": 3.1606855654985476e-06, + "loss": 0.4685, + "step": 3592 + }, + { + "epoch": 0.658299743495786, + "grad_norm": 0.3618932918467087, + "learning_rate": 3.157712301506691e-06, + "loss": 0.455, + "step": 3593 + }, + { + "epoch": 0.6584829607914987, + "grad_norm": 0.5449064334380102, + "learning_rate": 3.154739791114101e-06, + "loss": 0.442, + "step": 3594 + }, + { + "epoch": 0.6586661780872114, + "grad_norm": 0.4202501964497236, + "learning_rate": 3.1517680355366985e-06, + "loss": 0.4935, + "step": 3595 + }, + { + "epoch": 0.6588493953829242, + "grad_norm": 0.40316494653137924, + "learning_rate": 3.148797035990104e-06, + "loss": 0.4941, + "step": 3596 + }, + { + "epoch": 0.6590326126786369, + "grad_norm": 0.5119153385303149, + "learning_rate": 3.145826793689618e-06, + "loss": 0.4902, + "step": 3597 + }, + { + "epoch": 0.6592158299743496, + "grad_norm": 0.3826750892441696, + "learning_rate": 3.1428573098502377e-06, + "loss": 0.4889, + "step": 3598 + }, + { + "epoch": 0.6593990472700623, + "grad_norm": 0.3621392721647233, + "learning_rate": 3.1398885856866513e-06, + "loss": 0.4624, + "step": 3599 + }, + { + "epoch": 0.659582264565775, + "grad_norm": 0.3584064568844213, + "learning_rate": 3.136920622413233e-06, + "loss": 0.4653, + "step": 3600 + }, + { + "epoch": 0.6597654818614878, + "grad_norm": 0.34566302254778253, + "learning_rate": 3.133953421244044e-06, + "loss": 0.4615, + "step": 3601 + }, + { + "epoch": 0.6599486991572004, + "grad_norm": 0.3874955808407621, + "learning_rate": 3.1309869833928396e-06, + "loss": 0.4437, + "step": 3602 + }, + { + "epoch": 0.6601319164529131, + "grad_norm": 0.3285780187411082, + "learning_rate": 3.1280213100730595e-06, + "loss": 0.4648, + "step": 3603 + }, + { + "epoch": 0.6603151337486258, + "grad_norm": 0.36359675720348555, + "learning_rate": 3.1250564024978305e-06, + "loss": 0.476, + "step": 3604 + }, + { + "epoch": 0.6604983510443386, + "grad_norm": 0.3660376161139192, + "learning_rate": 3.1220922618799642e-06, + "loss": 0.4756, + "step": 3605 + }, + { + "epoch": 0.6606815683400513, + "grad_norm": 0.3400712340983726, + "learning_rate": 3.1191288894319627e-06, + "loss": 0.4626, + "step": 3606 + }, + { + "epoch": 0.660864785635764, + "grad_norm": 0.34660446398528555, + "learning_rate": 3.1161662863660125e-06, + "loss": 0.4726, + "step": 3607 + }, + { + "epoch": 0.6610480029314767, + "grad_norm": 0.37898360099511585, + "learning_rate": 3.1132044538939853e-06, + "loss": 0.4672, + "step": 3608 + }, + { + "epoch": 0.6612312202271895, + "grad_norm": 0.3646086298484684, + "learning_rate": 3.110243393227435e-06, + "loss": 0.4916, + "step": 3609 + }, + { + "epoch": 0.6614144375229022, + "grad_norm": 0.3545221361886584, + "learning_rate": 3.1072831055776023e-06, + "loss": 0.4918, + "step": 3610 + }, + { + "epoch": 0.6615976548186149, + "grad_norm": 0.36314042011581393, + "learning_rate": 3.104323592155413e-06, + "loss": 0.4873, + "step": 3611 + }, + { + "epoch": 0.6617808721143276, + "grad_norm": 0.45630613531265113, + "learning_rate": 3.101364854171474e-06, + "loss": 0.4661, + "step": 3612 + }, + { + "epoch": 0.6619640894100403, + "grad_norm": 0.4331128688633357, + "learning_rate": 3.0984068928360743e-06, + "loss": 0.503, + "step": 3613 + }, + { + "epoch": 0.6621473067057531, + "grad_norm": 0.3680508671750564, + "learning_rate": 3.0954497093591863e-06, + "loss": 0.5063, + "step": 3614 + }, + { + "epoch": 0.6623305240014657, + "grad_norm": 0.37461728331324273, + "learning_rate": 3.092493304950467e-06, + "loss": 0.4862, + "step": 3615 + }, + { + "epoch": 0.6625137412971784, + "grad_norm": 0.34622368931668424, + "learning_rate": 3.08953768081925e-06, + "loss": 0.4514, + "step": 3616 + }, + { + "epoch": 0.6626969585928911, + "grad_norm": 0.33003040953059065, + "learning_rate": 3.0865828381745515e-06, + "loss": 0.4526, + "step": 3617 + }, + { + "epoch": 0.6628801758886039, + "grad_norm": 0.3571333496770718, + "learning_rate": 3.0836287782250694e-06, + "loss": 0.4724, + "step": 3618 + }, + { + "epoch": 0.6630633931843166, + "grad_norm": 0.35595962314960083, + "learning_rate": 3.08067550217918e-06, + "loss": 0.4927, + "step": 3619 + }, + { + "epoch": 0.6632466104800293, + "grad_norm": 0.8339189103706414, + "learning_rate": 3.0777230112449386e-06, + "loss": 0.46, + "step": 3620 + }, + { + "epoch": 0.663429827775742, + "grad_norm": 0.3911907340807536, + "learning_rate": 3.0747713066300787e-06, + "loss": 0.5123, + "step": 3621 + }, + { + "epoch": 0.6636130450714548, + "grad_norm": 0.363609262010565, + "learning_rate": 3.0718203895420174e-06, + "loss": 0.4621, + "step": 3622 + }, + { + "epoch": 0.6637962623671675, + "grad_norm": 0.4074492458258354, + "learning_rate": 3.0688702611878435e-06, + "loss": 0.485, + "step": 3623 + }, + { + "epoch": 0.6639794796628802, + "grad_norm": 0.3599190298437237, + "learning_rate": 3.065920922774325e-06, + "loss": 0.4848, + "step": 3624 + }, + { + "epoch": 0.6641626969585929, + "grad_norm": 0.39162387951002525, + "learning_rate": 3.0629723755079068e-06, + "loss": 0.4716, + "step": 3625 + }, + { + "epoch": 0.6643459142543056, + "grad_norm": 0.3656180402897322, + "learning_rate": 3.0600246205947114e-06, + "loss": 0.4713, + "step": 3626 + }, + { + "epoch": 0.6645291315500184, + "grad_norm": 0.3532619420241813, + "learning_rate": 3.057077659240538e-06, + "loss": 0.4234, + "step": 3627 + }, + { + "epoch": 0.664712348845731, + "grad_norm": 0.3875129074414418, + "learning_rate": 3.054131492650857e-06, + "loss": 0.4493, + "step": 3628 + }, + { + "epoch": 0.6648955661414437, + "grad_norm": 0.3676652495284407, + "learning_rate": 3.0511861220308154e-06, + "loss": 0.4803, + "step": 3629 + }, + { + "epoch": 0.6650787834371564, + "grad_norm": 0.3513796811625531, + "learning_rate": 3.0482415485852378e-06, + "loss": 0.4512, + "step": 3630 + }, + { + "epoch": 0.6652620007328692, + "grad_norm": 0.3685400399517221, + "learning_rate": 3.045297773518621e-06, + "loss": 0.4687, + "step": 3631 + }, + { + "epoch": 0.6654452180285819, + "grad_norm": 0.43403002079658626, + "learning_rate": 3.042354798035132e-06, + "loss": 0.5103, + "step": 3632 + }, + { + "epoch": 0.6656284353242946, + "grad_norm": 0.5647428782333035, + "learning_rate": 3.0394126233386144e-06, + "loss": 0.4606, + "step": 3633 + }, + { + "epoch": 0.6658116526200073, + "grad_norm": 0.40573467926341555, + "learning_rate": 3.036471250632582e-06, + "loss": 0.4462, + "step": 3634 + }, + { + "epoch": 0.66599486991572, + "grad_norm": 0.3498915474683279, + "learning_rate": 3.033530681120225e-06, + "loss": 0.4577, + "step": 3635 + }, + { + "epoch": 0.6661780872114328, + "grad_norm": 0.34578109111582106, + "learning_rate": 3.0305909160043977e-06, + "loss": 0.5043, + "step": 3636 + }, + { + "epoch": 0.6663613045071455, + "grad_norm": 0.4737791807620961, + "learning_rate": 3.0276519564876306e-06, + "loss": 0.4952, + "step": 3637 + }, + { + "epoch": 0.6665445218028582, + "grad_norm": 0.36977981521129966, + "learning_rate": 3.0247138037721237e-06, + "loss": 0.4898, + "step": 3638 + }, + { + "epoch": 0.6667277390985709, + "grad_norm": 0.4075260326550853, + "learning_rate": 3.02177645905975e-06, + "loss": 0.4761, + "step": 3639 + }, + { + "epoch": 0.6669109563942837, + "grad_norm": 0.37687815737249974, + "learning_rate": 3.0188399235520417e-06, + "loss": 0.4657, + "step": 3640 + }, + { + "epoch": 0.6670941736899964, + "grad_norm": 0.3391944824739916, + "learning_rate": 3.015904198450211e-06, + "loss": 0.4485, + "step": 3641 + }, + { + "epoch": 0.667277390985709, + "grad_norm": 0.4597155324203258, + "learning_rate": 3.0129692849551357e-06, + "loss": 0.4781, + "step": 3642 + }, + { + "epoch": 0.6674606082814217, + "grad_norm": 0.3718169427643658, + "learning_rate": 3.0100351842673593e-06, + "loss": 0.4723, + "step": 3643 + }, + { + "epoch": 0.6676438255771345, + "grad_norm": 0.3531768798909512, + "learning_rate": 3.0071018975870912e-06, + "loss": 0.4831, + "step": 3644 + }, + { + "epoch": 0.6678270428728472, + "grad_norm": 0.32295965024049833, + "learning_rate": 3.004169426114215e-06, + "loss": 0.4594, + "step": 3645 + }, + { + "epoch": 0.6680102601685599, + "grad_norm": 0.3654691371662637, + "learning_rate": 3.001237771048274e-06, + "loss": 0.4764, + "step": 3646 + }, + { + "epoch": 0.6681934774642726, + "grad_norm": 0.33702086546319027, + "learning_rate": 2.998306933588482e-06, + "loss": 0.4785, + "step": 3647 + }, + { + "epoch": 0.6683766947599853, + "grad_norm": 0.3736833496299851, + "learning_rate": 2.9953769149337133e-06, + "loss": 0.479, + "step": 3648 + }, + { + "epoch": 0.6685599120556981, + "grad_norm": 0.3687097220827378, + "learning_rate": 2.992447716282511e-06, + "loss": 0.4644, + "step": 3649 + }, + { + "epoch": 0.6687431293514108, + "grad_norm": 0.33093257795201414, + "learning_rate": 2.9895193388330836e-06, + "loss": 0.4468, + "step": 3650 + }, + { + "epoch": 0.6689263466471235, + "grad_norm": 0.40345814477015735, + "learning_rate": 2.9865917837833025e-06, + "loss": 0.4575, + "step": 3651 + }, + { + "epoch": 0.6691095639428362, + "grad_norm": 0.36138483287970047, + "learning_rate": 2.983665052330701e-06, + "loss": 0.4673, + "step": 3652 + }, + { + "epoch": 0.669292781238549, + "grad_norm": 0.44660071361125864, + "learning_rate": 2.980739145672475e-06, + "loss": 0.4613, + "step": 3653 + }, + { + "epoch": 0.6694759985342617, + "grad_norm": 0.339100916728272, + "learning_rate": 2.9778140650054885e-06, + "loss": 0.4939, + "step": 3654 + }, + { + "epoch": 0.6696592158299743, + "grad_norm": 0.34620832486159653, + "learning_rate": 2.974889811526263e-06, + "loss": 0.4815, + "step": 3655 + }, + { + "epoch": 0.669842433125687, + "grad_norm": 0.3515974116024798, + "learning_rate": 2.9719663864309817e-06, + "loss": 0.4378, + "step": 3656 + }, + { + "epoch": 0.6700256504213998, + "grad_norm": 0.35908021152128194, + "learning_rate": 2.9690437909154903e-06, + "loss": 0.462, + "step": 3657 + }, + { + "epoch": 0.6702088677171125, + "grad_norm": 0.3792522193832267, + "learning_rate": 2.9661220261752925e-06, + "loss": 0.4762, + "step": 3658 + }, + { + "epoch": 0.6703920850128252, + "grad_norm": 0.3364184351070449, + "learning_rate": 2.963201093405559e-06, + "loss": 0.4518, + "step": 3659 + }, + { + "epoch": 0.6705753023085379, + "grad_norm": 0.33853165503611987, + "learning_rate": 2.960280993801112e-06, + "loss": 0.4649, + "step": 3660 + }, + { + "epoch": 0.6707585196042506, + "grad_norm": 0.371620313504904, + "learning_rate": 2.9573617285564383e-06, + "loss": 0.463, + "step": 3661 + }, + { + "epoch": 0.6709417368999634, + "grad_norm": 0.4183671449577938, + "learning_rate": 2.954443298865679e-06, + "loss": 0.4994, + "step": 3662 + }, + { + "epoch": 0.6711249541956761, + "grad_norm": 0.38398923054172057, + "learning_rate": 2.951525705922641e-06, + "loss": 0.4639, + "step": 3663 + }, + { + "epoch": 0.6713081714913888, + "grad_norm": 0.39645566544686267, + "learning_rate": 2.948608950920777e-06, + "loss": 0.4586, + "step": 3664 + }, + { + "epoch": 0.6714913887871015, + "grad_norm": 0.3525598524858501, + "learning_rate": 2.9456930350532088e-06, + "loss": 0.4712, + "step": 3665 + }, + { + "epoch": 0.6716746060828143, + "grad_norm": 0.38408556771090857, + "learning_rate": 2.942777959512707e-06, + "loss": 0.4474, + "step": 3666 + }, + { + "epoch": 0.671857823378527, + "grad_norm": 0.3644771139644592, + "learning_rate": 2.939863725491706e-06, + "loss": 0.4599, + "step": 3667 + }, + { + "epoch": 0.6720410406742396, + "grad_norm": 0.3574783291828957, + "learning_rate": 2.936950334182286e-06, + "loss": 0.4299, + "step": 3668 + }, + { + "epoch": 0.6722242579699523, + "grad_norm": 0.4124065852299994, + "learning_rate": 2.9340377867761916e-06, + "loss": 0.4988, + "step": 3669 + }, + { + "epoch": 0.6724074752656651, + "grad_norm": 0.3450490351178046, + "learning_rate": 2.931126084464816e-06, + "loss": 0.4901, + "step": 3670 + }, + { + "epoch": 0.6725906925613778, + "grad_norm": 0.389168918652041, + "learning_rate": 2.928215228439212e-06, + "loss": 0.4839, + "step": 3671 + }, + { + "epoch": 0.6727739098570905, + "grad_norm": 0.34540055292142574, + "learning_rate": 2.925305219890079e-06, + "loss": 0.4911, + "step": 3672 + }, + { + "epoch": 0.6729571271528032, + "grad_norm": 0.347865224841731, + "learning_rate": 2.922396060007778e-06, + "loss": 0.4912, + "step": 3673 + }, + { + "epoch": 0.6731403444485159, + "grad_norm": 0.3529199593714104, + "learning_rate": 2.919487749982318e-06, + "loss": 0.4597, + "step": 3674 + }, + { + "epoch": 0.6733235617442287, + "grad_norm": 0.3755895003197089, + "learning_rate": 2.916580291003361e-06, + "loss": 0.4754, + "step": 3675 + }, + { + "epoch": 0.6735067790399414, + "grad_norm": 0.33162124683853705, + "learning_rate": 2.913673684260222e-06, + "loss": 0.4756, + "step": 3676 + }, + { + "epoch": 0.6736899963356541, + "grad_norm": 0.3771007338972965, + "learning_rate": 2.9107679309418647e-06, + "loss": 0.5052, + "step": 3677 + }, + { + "epoch": 0.6738732136313668, + "grad_norm": 0.40152873637637315, + "learning_rate": 2.907863032236906e-06, + "loss": 0.4571, + "step": 3678 + }, + { + "epoch": 0.6740564309270796, + "grad_norm": 0.41354164522253456, + "learning_rate": 2.9049589893336182e-06, + "loss": 0.4691, + "step": 3679 + }, + { + "epoch": 0.6742396482227923, + "grad_norm": 0.33833040573853235, + "learning_rate": 2.90205580341991e-06, + "loss": 0.4323, + "step": 3680 + }, + { + "epoch": 0.6744228655185049, + "grad_norm": 0.39257173234965403, + "learning_rate": 2.8991534756833506e-06, + "loss": 0.4609, + "step": 3681 + }, + { + "epoch": 0.6746060828142176, + "grad_norm": 0.3657002644165077, + "learning_rate": 2.8962520073111578e-06, + "loss": 0.4892, + "step": 3682 + }, + { + "epoch": 0.6747893001099303, + "grad_norm": 0.3365704088860072, + "learning_rate": 2.893351399490194e-06, + "loss": 0.4731, + "step": 3683 + }, + { + "epoch": 0.6749725174056431, + "grad_norm": 0.5207839612722297, + "learning_rate": 2.8904516534069684e-06, + "loss": 0.4979, + "step": 3684 + }, + { + "epoch": 0.6751557347013558, + "grad_norm": 0.33086446522652063, + "learning_rate": 2.887552770247642e-06, + "loss": 0.4648, + "step": 3685 + }, + { + "epoch": 0.6753389519970685, + "grad_norm": 0.407449131563358, + "learning_rate": 2.884654751198023e-06, + "loss": 0.457, + "step": 3686 + }, + { + "epoch": 0.6755221692927812, + "grad_norm": 0.3582272557832074, + "learning_rate": 2.881757597443563e-06, + "loss": 0.4702, + "step": 3687 + }, + { + "epoch": 0.675705386588494, + "grad_norm": 0.3395069200116549, + "learning_rate": 2.8788613101693576e-06, + "loss": 0.4667, + "step": 3688 + }, + { + "epoch": 0.6758886038842067, + "grad_norm": 0.3947311899924817, + "learning_rate": 2.875965890560154e-06, + "loss": 0.4486, + "step": 3689 + }, + { + "epoch": 0.6760718211799194, + "grad_norm": 0.3357945071534466, + "learning_rate": 2.8730713398003436e-06, + "loss": 0.4754, + "step": 3690 + }, + { + "epoch": 0.6762550384756321, + "grad_norm": 0.3943563960508351, + "learning_rate": 2.870177659073958e-06, + "loss": 0.4562, + "step": 3691 + }, + { + "epoch": 0.6764382557713449, + "grad_norm": 0.3381366644305319, + "learning_rate": 2.8672848495646733e-06, + "loss": 0.4694, + "step": 3692 + }, + { + "epoch": 0.6766214730670576, + "grad_norm": 0.3769860154406749, + "learning_rate": 2.8643929124558136e-06, + "loss": 0.4874, + "step": 3693 + }, + { + "epoch": 0.6768046903627702, + "grad_norm": 0.37938707892452844, + "learning_rate": 2.861501848930346e-06, + "loss": 0.4888, + "step": 3694 + }, + { + "epoch": 0.6769879076584829, + "grad_norm": 0.34441409337604084, + "learning_rate": 2.8586116601708753e-06, + "loss": 0.4564, + "step": 3695 + }, + { + "epoch": 0.6771711249541956, + "grad_norm": 0.40988765730588894, + "learning_rate": 2.8557223473596497e-06, + "loss": 0.4521, + "step": 3696 + }, + { + "epoch": 0.6773543422499084, + "grad_norm": 0.38066253375011677, + "learning_rate": 2.852833911678565e-06, + "loss": 0.4517, + "step": 3697 + }, + { + "epoch": 0.6775375595456211, + "grad_norm": 0.40114839316292034, + "learning_rate": 2.8499463543091486e-06, + "loss": 0.4581, + "step": 3698 + }, + { + "epoch": 0.6777207768413338, + "grad_norm": 0.38710086636626223, + "learning_rate": 2.84705967643258e-06, + "loss": 0.4542, + "step": 3699 + }, + { + "epoch": 0.6779039941370465, + "grad_norm": 0.3908802584749809, + "learning_rate": 2.8441738792296662e-06, + "loss": 0.4883, + "step": 3700 + }, + { + "epoch": 0.6780872114327593, + "grad_norm": 0.40525376258769763, + "learning_rate": 2.841288963880866e-06, + "loss": 0.4786, + "step": 3701 + }, + { + "epoch": 0.678270428728472, + "grad_norm": 0.3457724875646536, + "learning_rate": 2.8384049315662686e-06, + "loss": 0.4356, + "step": 3702 + }, + { + "epoch": 0.6784536460241847, + "grad_norm": 0.39545041641540507, + "learning_rate": 2.835521783465608e-06, + "loss": 0.4523, + "step": 3703 + }, + { + "epoch": 0.6786368633198974, + "grad_norm": 0.3612810290783695, + "learning_rate": 2.8326395207582503e-06, + "loss": 0.4796, + "step": 3704 + }, + { + "epoch": 0.6788200806156102, + "grad_norm": 0.3627612940464092, + "learning_rate": 2.829758144623207e-06, + "loss": 0.4951, + "step": 3705 + }, + { + "epoch": 0.6790032979113229, + "grad_norm": 0.3582928178771917, + "learning_rate": 2.8268776562391198e-06, + "loss": 0.4595, + "step": 3706 + }, + { + "epoch": 0.6791865152070355, + "grad_norm": 0.36759753727104305, + "learning_rate": 2.8239980567842734e-06, + "loss": 0.4818, + "step": 3707 + }, + { + "epoch": 0.6793697325027482, + "grad_norm": 0.33183408875574777, + "learning_rate": 2.8211193474365823e-06, + "loss": 0.445, + "step": 3708 + }, + { + "epoch": 0.6795529497984609, + "grad_norm": 0.4106995263611198, + "learning_rate": 2.8182415293736044e-06, + "loss": 0.4981, + "step": 3709 + }, + { + "epoch": 0.6797361670941737, + "grad_norm": 0.3588796895617637, + "learning_rate": 2.815364603772525e-06, + "loss": 0.4693, + "step": 3710 + }, + { + "epoch": 0.6799193843898864, + "grad_norm": 0.41041493230920484, + "learning_rate": 2.8124885718101734e-06, + "loss": 0.4419, + "step": 3711 + }, + { + "epoch": 0.6801026016855991, + "grad_norm": 0.4129861730231944, + "learning_rate": 2.809613434663003e-06, + "loss": 0.4933, + "step": 3712 + }, + { + "epoch": 0.6802858189813118, + "grad_norm": 0.38588408487572534, + "learning_rate": 2.8067391935071114e-06, + "loss": 0.4365, + "step": 3713 + }, + { + "epoch": 0.6804690362770246, + "grad_norm": 0.3245481196880622, + "learning_rate": 2.8038658495182207e-06, + "loss": 0.4395, + "step": 3714 + }, + { + "epoch": 0.6806522535727373, + "grad_norm": 0.48103086858705363, + "learning_rate": 2.800993403871694e-06, + "loss": 0.4991, + "step": 3715 + }, + { + "epoch": 0.68083547086845, + "grad_norm": 0.3509720891321349, + "learning_rate": 2.7981218577425217e-06, + "loss": 0.4581, + "step": 3716 + }, + { + "epoch": 0.6810186881641627, + "grad_norm": 0.3824163074894137, + "learning_rate": 2.7952512123053242e-06, + "loss": 0.4539, + "step": 3717 + }, + { + "epoch": 0.6812019054598755, + "grad_norm": 0.386373829468268, + "learning_rate": 2.792381468734361e-06, + "loss": 0.4916, + "step": 3718 + }, + { + "epoch": 0.6813851227555882, + "grad_norm": 0.3768184046543257, + "learning_rate": 2.789512628203519e-06, + "loss": 0.4613, + "step": 3719 + }, + { + "epoch": 0.6815683400513008, + "grad_norm": 0.3513501402047029, + "learning_rate": 2.786644691886314e-06, + "loss": 0.4742, + "step": 3720 + }, + { + "epoch": 0.6817515573470135, + "grad_norm": 0.3930467936870125, + "learning_rate": 2.783777660955891e-06, + "loss": 0.4657, + "step": 3721 + }, + { + "epoch": 0.6819347746427262, + "grad_norm": 0.3687945503617596, + "learning_rate": 2.780911536585028e-06, + "loss": 0.4815, + "step": 3722 + }, + { + "epoch": 0.682117991938439, + "grad_norm": 0.3511399629548886, + "learning_rate": 2.778046319946135e-06, + "loss": 0.48, + "step": 3723 + }, + { + "epoch": 0.6823012092341517, + "grad_norm": 0.3439979153751878, + "learning_rate": 2.775182012211243e-06, + "loss": 0.4782, + "step": 3724 + }, + { + "epoch": 0.6824844265298644, + "grad_norm": 0.3918711967380485, + "learning_rate": 2.7723186145520136e-06, + "loss": 0.4706, + "step": 3725 + }, + { + "epoch": 0.6826676438255771, + "grad_norm": 0.36481251620257066, + "learning_rate": 2.7694561281397397e-06, + "loss": 0.4701, + "step": 3726 + }, + { + "epoch": 0.6828508611212899, + "grad_norm": 0.38947454107410345, + "learning_rate": 2.766594554145341e-06, + "loss": 0.4718, + "step": 3727 + }, + { + "epoch": 0.6830340784170026, + "grad_norm": 0.38005459572371675, + "learning_rate": 2.7637338937393597e-06, + "loss": 0.4998, + "step": 3728 + }, + { + "epoch": 0.6832172957127153, + "grad_norm": 0.3621566179234073, + "learning_rate": 2.7608741480919653e-06, + "loss": 0.4797, + "step": 3729 + }, + { + "epoch": 0.683400513008428, + "grad_norm": 0.3852371442483278, + "learning_rate": 2.758015318372957e-06, + "loss": 0.486, + "step": 3730 + }, + { + "epoch": 0.6835837303041407, + "grad_norm": 0.3684511691733946, + "learning_rate": 2.75515740575176e-06, + "loss": 0.4433, + "step": 3731 + }, + { + "epoch": 0.6837669475998535, + "grad_norm": 0.3738788714019796, + "learning_rate": 2.7523004113974137e-06, + "loss": 0.4782, + "step": 3732 + }, + { + "epoch": 0.6839501648955661, + "grad_norm": 0.42108472604876307, + "learning_rate": 2.7494443364785933e-06, + "loss": 0.4445, + "step": 3733 + }, + { + "epoch": 0.6841333821912788, + "grad_norm": 0.36823381624177864, + "learning_rate": 2.746589182163596e-06, + "loss": 0.4633, + "step": 3734 + }, + { + "epoch": 0.6843165994869915, + "grad_norm": 0.38881244485791505, + "learning_rate": 2.743734949620339e-06, + "loss": 0.4717, + "step": 3735 + }, + { + "epoch": 0.6844998167827043, + "grad_norm": 0.34549070190842535, + "learning_rate": 2.740881640016361e-06, + "loss": 0.47, + "step": 3736 + }, + { + "epoch": 0.684683034078417, + "grad_norm": 0.37799017457465145, + "learning_rate": 2.738029254518829e-06, + "loss": 0.4643, + "step": 3737 + }, + { + "epoch": 0.6848662513741297, + "grad_norm": 0.3235008582173912, + "learning_rate": 2.7351777942945314e-06, + "loss": 0.4551, + "step": 3738 + }, + { + "epoch": 0.6850494686698424, + "grad_norm": 0.3691861359328076, + "learning_rate": 2.7323272605098718e-06, + "loss": 0.4611, + "step": 3739 + }, + { + "epoch": 0.6852326859655552, + "grad_norm": 0.43348896919126706, + "learning_rate": 2.729477654330879e-06, + "loss": 0.4797, + "step": 3740 + }, + { + "epoch": 0.6854159032612679, + "grad_norm": 0.3751836062806745, + "learning_rate": 2.7266289769232023e-06, + "loss": 0.4739, + "step": 3741 + }, + { + "epoch": 0.6855991205569806, + "grad_norm": 0.41362711288421566, + "learning_rate": 2.7237812294521136e-06, + "loss": 0.4953, + "step": 3742 + }, + { + "epoch": 0.6857823378526933, + "grad_norm": 0.35358888372179276, + "learning_rate": 2.720934413082501e-06, + "loss": 0.4697, + "step": 3743 + }, + { + "epoch": 0.685965555148406, + "grad_norm": 0.3855820303997612, + "learning_rate": 2.7180885289788696e-06, + "loss": 0.4702, + "step": 3744 + }, + { + "epoch": 0.6861487724441188, + "grad_norm": 0.41264098981364733, + "learning_rate": 2.7152435783053475e-06, + "loss": 0.4986, + "step": 3745 + }, + { + "epoch": 0.6863319897398314, + "grad_norm": 0.36966626544271464, + "learning_rate": 2.7123995622256817e-06, + "loss": 0.4846, + "step": 3746 + }, + { + "epoch": 0.6865152070355441, + "grad_norm": 0.3626496310332301, + "learning_rate": 2.709556481903234e-06, + "loss": 0.4632, + "step": 3747 + }, + { + "epoch": 0.6866984243312568, + "grad_norm": 0.5099125627909202, + "learning_rate": 2.7067143385009797e-06, + "loss": 0.4309, + "step": 3748 + }, + { + "epoch": 0.6868816416269696, + "grad_norm": 0.3604408210608384, + "learning_rate": 2.7038731331815195e-06, + "loss": 0.4366, + "step": 3749 + }, + { + "epoch": 0.6870648589226823, + "grad_norm": 0.3383589578234134, + "learning_rate": 2.7010328671070675e-06, + "loss": 0.4685, + "step": 3750 + }, + { + "epoch": 0.687248076218395, + "grad_norm": 0.45657444748105225, + "learning_rate": 2.6981935414394488e-06, + "loss": 0.4782, + "step": 3751 + }, + { + "epoch": 0.6874312935141077, + "grad_norm": 0.4231260291934653, + "learning_rate": 2.695355157340107e-06, + "loss": 0.4836, + "step": 3752 + }, + { + "epoch": 0.6876145108098205, + "grad_norm": 0.3805981738702678, + "learning_rate": 2.6925177159701048e-06, + "loss": 0.5006, + "step": 3753 + }, + { + "epoch": 0.6877977281055332, + "grad_norm": 0.3425881387463667, + "learning_rate": 2.689681218490111e-06, + "loss": 0.464, + "step": 3754 + }, + { + "epoch": 0.6879809454012459, + "grad_norm": 0.36228947299472514, + "learning_rate": 2.686845666060416e-06, + "loss": 0.468, + "step": 3755 + }, + { + "epoch": 0.6881641626969586, + "grad_norm": 0.36398270949547584, + "learning_rate": 2.6840110598409166e-06, + "loss": 0.4876, + "step": 3756 + }, + { + "epoch": 0.6883473799926713, + "grad_norm": 0.33359919715688446, + "learning_rate": 2.6811774009911294e-06, + "loss": 0.464, + "step": 3757 + }, + { + "epoch": 0.6885305972883841, + "grad_norm": 0.341221437474584, + "learning_rate": 2.678344690670177e-06, + "loss": 0.4572, + "step": 3758 + }, + { + "epoch": 0.6887138145840968, + "grad_norm": 0.358835455718573, + "learning_rate": 2.675512930036802e-06, + "loss": 0.4506, + "step": 3759 + }, + { + "epoch": 0.6888970318798094, + "grad_norm": 0.3887364444258361, + "learning_rate": 2.672682120249348e-06, + "loss": 0.488, + "step": 3760 + }, + { + "epoch": 0.6890802491755221, + "grad_norm": 0.3845058420994467, + "learning_rate": 2.6698522624657808e-06, + "loss": 0.4554, + "step": 3761 + }, + { + "epoch": 0.6892634664712349, + "grad_norm": 0.3642399414292326, + "learning_rate": 2.667023357843667e-06, + "loss": 0.5043, + "step": 3762 + }, + { + "epoch": 0.6894466837669476, + "grad_norm": 0.3442406303936571, + "learning_rate": 2.6641954075401904e-06, + "loss": 0.4561, + "step": 3763 + }, + { + "epoch": 0.6896299010626603, + "grad_norm": 0.35755575299036146, + "learning_rate": 2.66136841271214e-06, + "loss": 0.4505, + "step": 3764 + }, + { + "epoch": 0.689813118358373, + "grad_norm": 0.38052624726055784, + "learning_rate": 2.6585423745159174e-06, + "loss": 0.4853, + "step": 3765 + }, + { + "epoch": 0.6899963356540857, + "grad_norm": 0.3363549507928974, + "learning_rate": 2.6557172941075284e-06, + "loss": 0.4729, + "step": 3766 + }, + { + "epoch": 0.6901795529497985, + "grad_norm": 0.3635065429068533, + "learning_rate": 2.6528931726425938e-06, + "loss": 0.4558, + "step": 3767 + }, + { + "epoch": 0.6903627702455112, + "grad_norm": 0.38251880298721247, + "learning_rate": 2.6500700112763356e-06, + "loss": 0.4843, + "step": 3768 + }, + { + "epoch": 0.6905459875412239, + "grad_norm": 0.3699868166960581, + "learning_rate": 2.6472478111635834e-06, + "loss": 0.4911, + "step": 3769 + }, + { + "epoch": 0.6907292048369366, + "grad_norm": 0.40083726336614606, + "learning_rate": 2.6444265734587786e-06, + "loss": 0.4778, + "step": 3770 + }, + { + "epoch": 0.6909124221326494, + "grad_norm": 0.38075230204167265, + "learning_rate": 2.6416062993159673e-06, + "loss": 0.498, + "step": 3771 + }, + { + "epoch": 0.691095639428362, + "grad_norm": 0.4906523195937728, + "learning_rate": 2.6387869898887986e-06, + "loss": 0.4889, + "step": 3772 + }, + { + "epoch": 0.6912788567240747, + "grad_norm": 0.3831131731787244, + "learning_rate": 2.635968646330527e-06, + "loss": 0.4708, + "step": 3773 + }, + { + "epoch": 0.6914620740197874, + "grad_norm": 0.3515802384587336, + "learning_rate": 2.6331512697940144e-06, + "loss": 0.4758, + "step": 3774 + }, + { + "epoch": 0.6916452913155002, + "grad_norm": 0.37697764812496154, + "learning_rate": 2.6303348614317286e-06, + "loss": 0.4801, + "step": 3775 + }, + { + "epoch": 0.6918285086112129, + "grad_norm": 0.3209903659735475, + "learning_rate": 2.6275194223957374e-06, + "loss": 0.4479, + "step": 3776 + }, + { + "epoch": 0.6920117259069256, + "grad_norm": 0.34859891071978905, + "learning_rate": 2.624704953837711e-06, + "loss": 0.4883, + "step": 3777 + }, + { + "epoch": 0.6921949432026383, + "grad_norm": 0.3597534669019847, + "learning_rate": 2.621891456908929e-06, + "loss": 0.4772, + "step": 3778 + }, + { + "epoch": 0.692378160498351, + "grad_norm": 0.34414379091468106, + "learning_rate": 2.6190789327602695e-06, + "loss": 0.459, + "step": 3779 + }, + { + "epoch": 0.6925613777940638, + "grad_norm": 0.358605655316751, + "learning_rate": 2.6162673825422124e-06, + "loss": 0.4358, + "step": 3780 + }, + { + "epoch": 0.6927445950897765, + "grad_norm": 0.3564847100502352, + "learning_rate": 2.6134568074048383e-06, + "loss": 0.501, + "step": 3781 + }, + { + "epoch": 0.6929278123854892, + "grad_norm": 0.4221096845433095, + "learning_rate": 2.6106472084978307e-06, + "loss": 0.4552, + "step": 3782 + }, + { + "epoch": 0.6931110296812019, + "grad_norm": 0.3834129742156116, + "learning_rate": 2.6078385869704794e-06, + "loss": 0.4841, + "step": 3783 + }, + { + "epoch": 0.6932942469769147, + "grad_norm": 0.37876116224077055, + "learning_rate": 2.605030943971659e-06, + "loss": 0.4225, + "step": 3784 + }, + { + "epoch": 0.6934774642726274, + "grad_norm": 0.33034529194712653, + "learning_rate": 2.602224280649859e-06, + "loss": 0.4916, + "step": 3785 + }, + { + "epoch": 0.69366068156834, + "grad_norm": 0.3356487090419772, + "learning_rate": 2.5994185981531617e-06, + "loss": 0.4113, + "step": 3786 + }, + { + "epoch": 0.6938438988640527, + "grad_norm": 0.37330363401017225, + "learning_rate": 2.5966138976292525e-06, + "loss": 0.4901, + "step": 3787 + }, + { + "epoch": 0.6940271161597655, + "grad_norm": 0.3688010851609858, + "learning_rate": 2.5938101802254045e-06, + "loss": 0.482, + "step": 3788 + }, + { + "epoch": 0.6942103334554782, + "grad_norm": 0.44309397358359714, + "learning_rate": 2.5910074470884996e-06, + "loss": 0.4799, + "step": 3789 + }, + { + "epoch": 0.6943935507511909, + "grad_norm": 0.3807439690701102, + "learning_rate": 2.588205699365014e-06, + "loss": 0.4579, + "step": 3790 + }, + { + "epoch": 0.6945767680469036, + "grad_norm": 0.3394042004438676, + "learning_rate": 2.5854049382010193e-06, + "loss": 0.4626, + "step": 3791 + }, + { + "epoch": 0.6947599853426163, + "grad_norm": 0.3866460838493798, + "learning_rate": 2.582605164742183e-06, + "loss": 0.5061, + "step": 3792 + }, + { + "epoch": 0.6949432026383291, + "grad_norm": 0.6739869457451736, + "learning_rate": 2.5798063801337703e-06, + "loss": 0.4739, + "step": 3793 + }, + { + "epoch": 0.6951264199340418, + "grad_norm": 0.3374424755782798, + "learning_rate": 2.5770085855206447e-06, + "loss": 0.4435, + "step": 3794 + }, + { + "epoch": 0.6953096372297545, + "grad_norm": 0.35344268960593894, + "learning_rate": 2.5742117820472585e-06, + "loss": 0.4457, + "step": 3795 + }, + { + "epoch": 0.6954928545254672, + "grad_norm": 0.3539278780158419, + "learning_rate": 2.5714159708576604e-06, + "loss": 0.4662, + "step": 3796 + }, + { + "epoch": 0.69567607182118, + "grad_norm": 0.4654889373205005, + "learning_rate": 2.5686211530954956e-06, + "loss": 0.4475, + "step": 3797 + }, + { + "epoch": 0.6958592891168927, + "grad_norm": 0.3483866384191484, + "learning_rate": 2.565827329904005e-06, + "loss": 0.4583, + "step": 3798 + }, + { + "epoch": 0.6960425064126053, + "grad_norm": 0.35564238697040806, + "learning_rate": 2.5630345024260164e-06, + "loss": 0.4744, + "step": 3799 + }, + { + "epoch": 0.696225723708318, + "grad_norm": 0.33379335510919744, + "learning_rate": 2.5602426718039507e-06, + "loss": 0.4663, + "step": 3800 + }, + { + "epoch": 0.6964089410040308, + "grad_norm": 0.3879006848124751, + "learning_rate": 2.5574518391798276e-06, + "loss": 0.4432, + "step": 3801 + }, + { + "epoch": 0.6965921582997435, + "grad_norm": 0.3745353259464326, + "learning_rate": 2.554662005695255e-06, + "loss": 0.482, + "step": 3802 + }, + { + "epoch": 0.6967753755954562, + "grad_norm": 0.4214484952398853, + "learning_rate": 2.5518731724914296e-06, + "loss": 0.4729, + "step": 3803 + }, + { + "epoch": 0.6969585928911689, + "grad_norm": 0.3504880714869815, + "learning_rate": 2.54908534070914e-06, + "loss": 0.4443, + "step": 3804 + }, + { + "epoch": 0.6971418101868816, + "grad_norm": 0.38355719727950427, + "learning_rate": 2.5462985114887696e-06, + "loss": 0.4446, + "step": 3805 + }, + { + "epoch": 0.6973250274825944, + "grad_norm": 0.36437676145395215, + "learning_rate": 2.543512685970284e-06, + "loss": 0.4833, + "step": 3806 + }, + { + "epoch": 0.6975082447783071, + "grad_norm": 0.3447362656869495, + "learning_rate": 2.5407278652932473e-06, + "loss": 0.4767, + "step": 3807 + }, + { + "epoch": 0.6976914620740198, + "grad_norm": 0.37408240841902085, + "learning_rate": 2.5379440505968027e-06, + "loss": 0.4451, + "step": 3808 + }, + { + "epoch": 0.6978746793697325, + "grad_norm": 0.35645270482914476, + "learning_rate": 2.5351612430196914e-06, + "loss": 0.4529, + "step": 3809 + }, + { + "epoch": 0.6980578966654453, + "grad_norm": 0.3853121816910978, + "learning_rate": 2.532379443700235e-06, + "loss": 0.4579, + "step": 3810 + }, + { + "epoch": 0.698241113961158, + "grad_norm": 0.3649173678181005, + "learning_rate": 2.529598653776349e-06, + "loss": 0.4744, + "step": 3811 + }, + { + "epoch": 0.6984243312568706, + "grad_norm": 0.32668542453414895, + "learning_rate": 2.526818874385529e-06, + "loss": 0.4328, + "step": 3812 + }, + { + "epoch": 0.6986075485525833, + "grad_norm": 0.4036555858766503, + "learning_rate": 2.5240401066648644e-06, + "loss": 0.4854, + "step": 3813 + }, + { + "epoch": 0.698790765848296, + "grad_norm": 0.4019267619776793, + "learning_rate": 2.521262351751025e-06, + "loss": 0.507, + "step": 3814 + }, + { + "epoch": 0.6989739831440088, + "grad_norm": 0.4947615656802971, + "learning_rate": 2.5184856107802713e-06, + "loss": 0.4668, + "step": 3815 + }, + { + "epoch": 0.6991572004397215, + "grad_norm": 0.34782180039973204, + "learning_rate": 2.5157098848884433e-06, + "loss": 0.4816, + "step": 3816 + }, + { + "epoch": 0.6993404177354342, + "grad_norm": 0.3682041973133259, + "learning_rate": 2.5129351752109725e-06, + "loss": 0.465, + "step": 3817 + }, + { + "epoch": 0.6995236350311469, + "grad_norm": 0.37345579131040246, + "learning_rate": 2.510161482882868e-06, + "loss": 0.4619, + "step": 3818 + }, + { + "epoch": 0.6997068523268597, + "grad_norm": 0.3981491412912823, + "learning_rate": 2.5073888090387288e-06, + "loss": 0.4577, + "step": 3819 + }, + { + "epoch": 0.6998900696225724, + "grad_norm": 0.3557242369312865, + "learning_rate": 2.5046171548127335e-06, + "loss": 0.4378, + "step": 3820 + }, + { + "epoch": 0.7000732869182851, + "grad_norm": 0.3475467955199327, + "learning_rate": 2.5018465213386417e-06, + "loss": 0.4575, + "step": 3821 + }, + { + "epoch": 0.7002565042139978, + "grad_norm": 0.3799250774146336, + "learning_rate": 2.4990769097498007e-06, + "loss": 0.4332, + "step": 3822 + }, + { + "epoch": 0.7004397215097106, + "grad_norm": 0.40500568866145925, + "learning_rate": 2.4963083211791396e-06, + "loss": 0.4982, + "step": 3823 + }, + { + "epoch": 0.7006229388054233, + "grad_norm": 0.3269978508130673, + "learning_rate": 2.4935407567591643e-06, + "loss": 0.472, + "step": 3824 + }, + { + "epoch": 0.700806156101136, + "grad_norm": 0.3433257970162682, + "learning_rate": 2.4907742176219622e-06, + "loss": 0.488, + "step": 3825 + }, + { + "epoch": 0.7009893733968486, + "grad_norm": 0.35699368870853326, + "learning_rate": 2.4880087048992063e-06, + "loss": 0.4821, + "step": 3826 + }, + { + "epoch": 0.7011725906925613, + "grad_norm": 0.37284267851339215, + "learning_rate": 2.485244219722148e-06, + "loss": 0.4866, + "step": 3827 + }, + { + "epoch": 0.7013558079882741, + "grad_norm": 0.4047583327732591, + "learning_rate": 2.4824807632216153e-06, + "loss": 0.4638, + "step": 3828 + }, + { + "epoch": 0.7015390252839868, + "grad_norm": 0.35020387747221843, + "learning_rate": 2.479718336528015e-06, + "loss": 0.4749, + "step": 3829 + }, + { + "epoch": 0.7017222425796995, + "grad_norm": 0.3834433962201249, + "learning_rate": 2.4769569407713385e-06, + "loss": 0.4548, + "step": 3830 + }, + { + "epoch": 0.7019054598754122, + "grad_norm": 0.37922987865888286, + "learning_rate": 2.4741965770811514e-06, + "loss": 0.4706, + "step": 3831 + }, + { + "epoch": 0.702088677171125, + "grad_norm": 0.4097363306160758, + "learning_rate": 2.4714372465865982e-06, + "loss": 0.4939, + "step": 3832 + }, + { + "epoch": 0.7022718944668377, + "grad_norm": 0.38096405632188923, + "learning_rate": 2.4686789504163966e-06, + "loss": 0.4794, + "step": 3833 + }, + { + "epoch": 0.7024551117625504, + "grad_norm": 0.37478178387756106, + "learning_rate": 2.4659216896988477e-06, + "loss": 0.4999, + "step": 3834 + }, + { + "epoch": 0.7026383290582631, + "grad_norm": 0.45520764703702454, + "learning_rate": 2.4631654655618293e-06, + "loss": 0.461, + "step": 3835 + }, + { + "epoch": 0.7028215463539759, + "grad_norm": 0.3558083039435706, + "learning_rate": 2.4604102791327846e-06, + "loss": 0.4619, + "step": 3836 + }, + { + "epoch": 0.7030047636496886, + "grad_norm": 0.38001496752817143, + "learning_rate": 2.457656131538744e-06, + "loss": 0.4674, + "step": 3837 + }, + { + "epoch": 0.7031879809454012, + "grad_norm": 0.3188394848956792, + "learning_rate": 2.4549030239063084e-06, + "loss": 0.43, + "step": 3838 + }, + { + "epoch": 0.7033711982411139, + "grad_norm": 0.3351553365562699, + "learning_rate": 2.4521509573616575e-06, + "loss": 0.4696, + "step": 3839 + }, + { + "epoch": 0.7035544155368266, + "grad_norm": 0.4095489677061301, + "learning_rate": 2.449399933030534e-06, + "loss": 0.5064, + "step": 3840 + }, + { + "epoch": 0.7037376328325394, + "grad_norm": 0.3989519333618829, + "learning_rate": 2.446649952038265e-06, + "loss": 0.4568, + "step": 3841 + }, + { + "epoch": 0.7039208501282521, + "grad_norm": 0.42013601901862657, + "learning_rate": 2.443901015509749e-06, + "loss": 0.4832, + "step": 3842 + }, + { + "epoch": 0.7041040674239648, + "grad_norm": 0.33201803636696386, + "learning_rate": 2.4411531245694546e-06, + "loss": 0.4842, + "step": 3843 + }, + { + "epoch": 0.7042872847196775, + "grad_norm": 0.46701253249732194, + "learning_rate": 2.438406280341421e-06, + "loss": 0.4547, + "step": 3844 + }, + { + "epoch": 0.7044705020153903, + "grad_norm": 0.37847070682160494, + "learning_rate": 2.435660483949264e-06, + "loss": 0.4373, + "step": 3845 + }, + { + "epoch": 0.704653719311103, + "grad_norm": 0.41504348463682667, + "learning_rate": 2.432915736516171e-06, + "loss": 0.5079, + "step": 3846 + }, + { + "epoch": 0.7048369366068157, + "grad_norm": 0.37695025503483887, + "learning_rate": 2.4301720391648957e-06, + "loss": 0.4595, + "step": 3847 + }, + { + "epoch": 0.7050201539025284, + "grad_norm": 0.35055356188077097, + "learning_rate": 2.4274293930177628e-06, + "loss": 0.4764, + "step": 3848 + }, + { + "epoch": 0.7052033711982412, + "grad_norm": 0.39196353523551825, + "learning_rate": 2.424687799196671e-06, + "loss": 0.4873, + "step": 3849 + }, + { + "epoch": 0.7053865884939539, + "grad_norm": 0.41730549722802246, + "learning_rate": 2.421947258823088e-06, + "loss": 0.5018, + "step": 3850 + }, + { + "epoch": 0.7055698057896665, + "grad_norm": 0.3641735944661569, + "learning_rate": 2.4192077730180475e-06, + "loss": 0.4793, + "step": 3851 + }, + { + "epoch": 0.7057530230853792, + "grad_norm": 0.4115128953736341, + "learning_rate": 2.4164693429021503e-06, + "loss": 0.4487, + "step": 3852 + }, + { + "epoch": 0.7059362403810919, + "grad_norm": 0.38180106974255745, + "learning_rate": 2.4137319695955706e-06, + "loss": 0.4819, + "step": 3853 + }, + { + "epoch": 0.7061194576768047, + "grad_norm": 0.3762680149784185, + "learning_rate": 2.410995654218051e-06, + "loss": 0.4769, + "step": 3854 + }, + { + "epoch": 0.7063026749725174, + "grad_norm": 0.39658884341466655, + "learning_rate": 2.4082603978888947e-06, + "loss": 0.4741, + "step": 3855 + }, + { + "epoch": 0.7064858922682301, + "grad_norm": 0.35179562448237883, + "learning_rate": 2.4055262017269747e-06, + "loss": 0.4545, + "step": 3856 + }, + { + "epoch": 0.7066691095639428, + "grad_norm": 0.3695261181306668, + "learning_rate": 2.402793066850734e-06, + "loss": 0.4898, + "step": 3857 + }, + { + "epoch": 0.7068523268596556, + "grad_norm": 0.39564675640388264, + "learning_rate": 2.400060994378176e-06, + "loss": 0.4805, + "step": 3858 + }, + { + "epoch": 0.7070355441553683, + "grad_norm": 0.38729094782316986, + "learning_rate": 2.3973299854268743e-06, + "loss": 0.4858, + "step": 3859 + }, + { + "epoch": 0.707218761451081, + "grad_norm": 0.37195210119847283, + "learning_rate": 2.394600041113962e-06, + "loss": 0.457, + "step": 3860 + }, + { + "epoch": 0.7074019787467937, + "grad_norm": 0.42760347455983244, + "learning_rate": 2.391871162556143e-06, + "loss": 0.4667, + "step": 3861 + }, + { + "epoch": 0.7075851960425064, + "grad_norm": 0.36001218627908826, + "learning_rate": 2.3891433508696793e-06, + "loss": 0.4844, + "step": 3862 + }, + { + "epoch": 0.7077684133382192, + "grad_norm": 0.3914735651944279, + "learning_rate": 2.3864166071704026e-06, + "loss": 0.4889, + "step": 3863 + }, + { + "epoch": 0.7079516306339318, + "grad_norm": 0.3298837059612055, + "learning_rate": 2.3836909325737e-06, + "loss": 0.4754, + "step": 3864 + }, + { + "epoch": 0.7081348479296445, + "grad_norm": 0.37885170331444445, + "learning_rate": 2.3809663281945304e-06, + "loss": 0.4619, + "step": 3865 + }, + { + "epoch": 0.7083180652253572, + "grad_norm": 0.3471909899932877, + "learning_rate": 2.3782427951474065e-06, + "loss": 0.4672, + "step": 3866 + }, + { + "epoch": 0.70850128252107, + "grad_norm": 0.39582817919789814, + "learning_rate": 2.3755203345464093e-06, + "loss": 0.4973, + "step": 3867 + }, + { + "epoch": 0.7086844998167827, + "grad_norm": 0.5171192742768901, + "learning_rate": 2.3727989475051756e-06, + "loss": 0.514, + "step": 3868 + }, + { + "epoch": 0.7088677171124954, + "grad_norm": 0.3914387974525994, + "learning_rate": 2.370078635136908e-06, + "loss": 0.4796, + "step": 3869 + }, + { + "epoch": 0.7090509344082081, + "grad_norm": 0.35608752364478324, + "learning_rate": 2.3673593985543647e-06, + "loss": 0.4813, + "step": 3870 + }, + { + "epoch": 0.7092341517039209, + "grad_norm": 0.3407567232006967, + "learning_rate": 2.3646412388698703e-06, + "loss": 0.4411, + "step": 3871 + }, + { + "epoch": 0.7094173689996336, + "grad_norm": 0.4030342455048933, + "learning_rate": 2.361924157195302e-06, + "loss": 0.4286, + "step": 3872 + }, + { + "epoch": 0.7096005862953463, + "grad_norm": 0.3344753724022738, + "learning_rate": 2.359208154642098e-06, + "loss": 0.4764, + "step": 3873 + }, + { + "epoch": 0.709783803591059, + "grad_norm": 0.3877832467293633, + "learning_rate": 2.3564932323212574e-06, + "loss": 0.5046, + "step": 3874 + }, + { + "epoch": 0.7099670208867717, + "grad_norm": 0.4056288760749119, + "learning_rate": 2.3537793913433386e-06, + "loss": 0.4777, + "step": 3875 + }, + { + "epoch": 0.7101502381824845, + "grad_norm": 1.5040346864031384, + "learning_rate": 2.3510666328184527e-06, + "loss": 0.4798, + "step": 3876 + }, + { + "epoch": 0.7103334554781972, + "grad_norm": 0.3310106746801017, + "learning_rate": 2.3483549578562692e-06, + "loss": 0.4546, + "step": 3877 + }, + { + "epoch": 0.7105166727739098, + "grad_norm": 0.39204472145138664, + "learning_rate": 2.3456443675660168e-06, + "loss": 0.4709, + "step": 3878 + }, + { + "epoch": 0.7106998900696225, + "grad_norm": 0.41356445706117884, + "learning_rate": 2.3429348630564814e-06, + "loss": 0.4584, + "step": 3879 + }, + { + "epoch": 0.7108831073653353, + "grad_norm": 0.35419883069363933, + "learning_rate": 2.3402264454360015e-06, + "loss": 0.4787, + "step": 3880 + }, + { + "epoch": 0.711066324661048, + "grad_norm": 0.37747083194603237, + "learning_rate": 2.3375191158124694e-06, + "loss": 0.4558, + "step": 3881 + }, + { + "epoch": 0.7112495419567607, + "grad_norm": 0.36158993228830477, + "learning_rate": 2.3348128752933375e-06, + "loss": 0.4704, + "step": 3882 + }, + { + "epoch": 0.7114327592524734, + "grad_norm": 0.44466267951102173, + "learning_rate": 2.3321077249856123e-06, + "loss": 0.4891, + "step": 3883 + }, + { + "epoch": 0.7116159765481862, + "grad_norm": 0.3908470845675225, + "learning_rate": 2.329403665995851e-06, + "loss": 0.4809, + "step": 3884 + }, + { + "epoch": 0.7117991938438989, + "grad_norm": 0.34947637497569195, + "learning_rate": 2.326700699430162e-06, + "loss": 0.4677, + "step": 3885 + }, + { + "epoch": 0.7119824111396116, + "grad_norm": 0.4034660171845511, + "learning_rate": 2.323998826394215e-06, + "loss": 0.4954, + "step": 3886 + }, + { + "epoch": 0.7121656284353243, + "grad_norm": 0.3902853494836444, + "learning_rate": 2.3212980479932278e-06, + "loss": 0.4708, + "step": 3887 + }, + { + "epoch": 0.712348845731037, + "grad_norm": 0.3826243485087709, + "learning_rate": 2.3185983653319704e-06, + "loss": 0.4577, + "step": 3888 + }, + { + "epoch": 0.7125320630267498, + "grad_norm": 0.34221063508214916, + "learning_rate": 2.315899779514762e-06, + "loss": 0.4828, + "step": 3889 + }, + { + "epoch": 0.7127152803224625, + "grad_norm": 0.31606410866518403, + "learning_rate": 2.313202291645478e-06, + "loss": 0.4421, + "step": 3890 + }, + { + "epoch": 0.7128984976181751, + "grad_norm": 0.39675831775841675, + "learning_rate": 2.3105059028275467e-06, + "loss": 0.4761, + "step": 3891 + }, + { + "epoch": 0.7130817149138878, + "grad_norm": 0.41928597858495215, + "learning_rate": 2.3078106141639347e-06, + "loss": 0.4918, + "step": 3892 + }, + { + "epoch": 0.7132649322096006, + "grad_norm": 0.35834153154382087, + "learning_rate": 2.30511642675717e-06, + "loss": 0.4284, + "step": 3893 + }, + { + "epoch": 0.7134481495053133, + "grad_norm": 0.37770177714774683, + "learning_rate": 2.30242334170933e-06, + "loss": 0.4528, + "step": 3894 + }, + { + "epoch": 0.713631366801026, + "grad_norm": 0.3701259832997652, + "learning_rate": 2.2997313601220354e-06, + "loss": 0.4799, + "step": 3895 + }, + { + "epoch": 0.7138145840967387, + "grad_norm": 0.4149159194213658, + "learning_rate": 2.297040483096456e-06, + "loss": 0.5102, + "step": 3896 + }, + { + "epoch": 0.7139978013924514, + "grad_norm": 0.41350958709860136, + "learning_rate": 2.2943507117333137e-06, + "loss": 0.4682, + "step": 3897 + }, + { + "epoch": 0.7141810186881642, + "grad_norm": 0.41705694501658497, + "learning_rate": 2.2916620471328784e-06, + "loss": 0.4695, + "step": 3898 + }, + { + "epoch": 0.7143642359838769, + "grad_norm": 0.40508098818541893, + "learning_rate": 2.288974490394962e-06, + "loss": 0.4575, + "step": 3899 + }, + { + "epoch": 0.7145474532795896, + "grad_norm": 0.3959996212933778, + "learning_rate": 2.2862880426189266e-06, + "loss": 0.4515, + "step": 3900 + }, + { + "epoch": 0.7147306705753023, + "grad_norm": 0.3670118613753447, + "learning_rate": 2.283602704903681e-06, + "loss": 0.4549, + "step": 3901 + }, + { + "epoch": 0.7149138878710151, + "grad_norm": 0.3518895294151359, + "learning_rate": 2.280918478347681e-06, + "loss": 0.4291, + "step": 3902 + }, + { + "epoch": 0.7150971051667278, + "grad_norm": 0.3296707504158102, + "learning_rate": 2.2782353640489246e-06, + "loss": 0.4803, + "step": 3903 + }, + { + "epoch": 0.7152803224624404, + "grad_norm": 0.38278413538931505, + "learning_rate": 2.275553363104955e-06, + "loss": 0.4306, + "step": 3904 + }, + { + "epoch": 0.7154635397581531, + "grad_norm": 0.4072668034438296, + "learning_rate": 2.272872476612863e-06, + "loss": 0.4642, + "step": 3905 + }, + { + "epoch": 0.7156467570538659, + "grad_norm": 0.3502020381492555, + "learning_rate": 2.2701927056692835e-06, + "loss": 0.4896, + "step": 3906 + }, + { + "epoch": 0.7158299743495786, + "grad_norm": 0.3691555729533692, + "learning_rate": 2.2675140513703924e-06, + "loss": 0.4672, + "step": 3907 + }, + { + "epoch": 0.7160131916452913, + "grad_norm": 0.35830120196254023, + "learning_rate": 2.2648365148119073e-06, + "loss": 0.4606, + "step": 3908 + }, + { + "epoch": 0.716196408941004, + "grad_norm": 0.34986818855700375, + "learning_rate": 2.2621600970890944e-06, + "loss": 0.455, + "step": 3909 + }, + { + "epoch": 0.7163796262367167, + "grad_norm": 0.3870870347746835, + "learning_rate": 2.2594847992967573e-06, + "loss": 0.4714, + "step": 3910 + }, + { + "epoch": 0.7165628435324295, + "grad_norm": 0.37249300671972857, + "learning_rate": 2.2568106225292447e-06, + "loss": 0.4847, + "step": 3911 + }, + { + "epoch": 0.7167460608281422, + "grad_norm": 0.43939952436081475, + "learning_rate": 2.2541375678804435e-06, + "loss": 0.5097, + "step": 3912 + }, + { + "epoch": 0.7169292781238549, + "grad_norm": 0.405532505727656, + "learning_rate": 2.251465636443786e-06, + "loss": 0.4772, + "step": 3913 + }, + { + "epoch": 0.7171124954195676, + "grad_norm": 0.3868217983838645, + "learning_rate": 2.2487948293122385e-06, + "loss": 0.4776, + "step": 3914 + }, + { + "epoch": 0.7172957127152804, + "grad_norm": 0.34727108875203117, + "learning_rate": 2.246125147578316e-06, + "loss": 0.4826, + "step": 3915 + }, + { + "epoch": 0.717478930010993, + "grad_norm": 0.35607943451775403, + "learning_rate": 2.2434565923340633e-06, + "loss": 0.4539, + "step": 3916 + }, + { + "epoch": 0.7176621473067057, + "grad_norm": 0.3600291279263338, + "learning_rate": 2.2407891646710744e-06, + "loss": 0.4679, + "step": 3917 + }, + { + "epoch": 0.7178453646024184, + "grad_norm": 0.3628355380823112, + "learning_rate": 2.2381228656804737e-06, + "loss": 0.4526, + "step": 3918 + }, + { + "epoch": 0.7180285818981312, + "grad_norm": 0.36213995409506666, + "learning_rate": 2.235457696452931e-06, + "loss": 0.4904, + "step": 3919 + }, + { + "epoch": 0.7182117991938439, + "grad_norm": 0.3831546832431041, + "learning_rate": 2.2327936580786454e-06, + "loss": 0.4443, + "step": 3920 + }, + { + "epoch": 0.7183950164895566, + "grad_norm": 0.42205583866499036, + "learning_rate": 2.230130751647363e-06, + "loss": 0.4698, + "step": 3921 + }, + { + "epoch": 0.7185782337852693, + "grad_norm": 0.37799608449143307, + "learning_rate": 2.2274689782483587e-06, + "loss": 0.4333, + "step": 3922 + }, + { + "epoch": 0.718761451080982, + "grad_norm": 0.3731146777944977, + "learning_rate": 2.224808338970452e-06, + "loss": 0.4437, + "step": 3923 + }, + { + "epoch": 0.7189446683766948, + "grad_norm": 0.3745077798290519, + "learning_rate": 2.2221488349019903e-06, + "loss": 0.4889, + "step": 3924 + }, + { + "epoch": 0.7191278856724075, + "grad_norm": 0.38593867796467957, + "learning_rate": 2.2194904671308597e-06, + "loss": 0.4837, + "step": 3925 + }, + { + "epoch": 0.7193111029681202, + "grad_norm": 0.3733427492850883, + "learning_rate": 2.2168332367444856e-06, + "loss": 0.4671, + "step": 3926 + }, + { + "epoch": 0.7194943202638329, + "grad_norm": 0.42288412356060506, + "learning_rate": 2.2141771448298204e-06, + "loss": 0.4987, + "step": 3927 + }, + { + "epoch": 0.7196775375595457, + "grad_norm": 0.40370031860317235, + "learning_rate": 2.2115221924733597e-06, + "loss": 0.4487, + "step": 3928 + }, + { + "epoch": 0.7198607548552584, + "grad_norm": 0.34387620450670553, + "learning_rate": 2.2088683807611245e-06, + "loss": 0.4551, + "step": 3929 + }, + { + "epoch": 0.720043972150971, + "grad_norm": 0.3357205462234579, + "learning_rate": 2.2062157107786755e-06, + "loss": 0.4798, + "step": 3930 + }, + { + "epoch": 0.7202271894466837, + "grad_norm": 0.36225562703789105, + "learning_rate": 2.2035641836111014e-06, + "loss": 0.4302, + "step": 3931 + }, + { + "epoch": 0.7204104067423965, + "grad_norm": 0.34692152266678516, + "learning_rate": 2.2009138003430293e-06, + "loss": 0.4343, + "step": 3932 + }, + { + "epoch": 0.7205936240381092, + "grad_norm": 0.47525673930571805, + "learning_rate": 2.1982645620586115e-06, + "loss": 0.4581, + "step": 3933 + }, + { + "epoch": 0.7207768413338219, + "grad_norm": 0.41490121146448244, + "learning_rate": 2.1956164698415384e-06, + "loss": 0.4804, + "step": 3934 + }, + { + "epoch": 0.7209600586295346, + "grad_norm": 0.3726254826073696, + "learning_rate": 2.192969524775025e-06, + "loss": 0.4929, + "step": 3935 + }, + { + "epoch": 0.7211432759252473, + "grad_norm": 0.39510906163396003, + "learning_rate": 2.1903237279418255e-06, + "loss": 0.4716, + "step": 3936 + }, + { + "epoch": 0.7213264932209601, + "grad_norm": 0.34147798613218494, + "learning_rate": 2.187679080424215e-06, + "loss": 0.4785, + "step": 3937 + }, + { + "epoch": 0.7215097105166728, + "grad_norm": 0.386384897070331, + "learning_rate": 2.1850355833040067e-06, + "loss": 0.4303, + "step": 3938 + }, + { + "epoch": 0.7216929278123855, + "grad_norm": 0.355376967657911, + "learning_rate": 2.182393237662536e-06, + "loss": 0.4635, + "step": 3939 + }, + { + "epoch": 0.7218761451080982, + "grad_norm": 0.36273164237101024, + "learning_rate": 2.1797520445806754e-06, + "loss": 0.4877, + "step": 3940 + }, + { + "epoch": 0.722059362403811, + "grad_norm": 0.38291884206668514, + "learning_rate": 2.177112005138816e-06, + "loss": 0.47, + "step": 3941 + }, + { + "epoch": 0.7222425796995237, + "grad_norm": 0.3746789118909081, + "learning_rate": 2.174473120416887e-06, + "loss": 0.463, + "step": 3942 + }, + { + "epoch": 0.7224257969952363, + "grad_norm": 0.35439101753244096, + "learning_rate": 2.1718353914943397e-06, + "loss": 0.4834, + "step": 3943 + }, + { + "epoch": 0.722609014290949, + "grad_norm": 0.33280492693648706, + "learning_rate": 2.1691988194501506e-06, + "loss": 0.464, + "step": 3944 + }, + { + "epoch": 0.7227922315866617, + "grad_norm": 0.41110827295372265, + "learning_rate": 2.1665634053628277e-06, + "loss": 0.5061, + "step": 3945 + }, + { + "epoch": 0.7229754488823745, + "grad_norm": 0.380399201123713, + "learning_rate": 2.163929150310405e-06, + "loss": 0.4375, + "step": 3946 + }, + { + "epoch": 0.7231586661780872, + "grad_norm": 0.41231434008927476, + "learning_rate": 2.161296055370441e-06, + "loss": 0.4764, + "step": 3947 + }, + { + "epoch": 0.7233418834737999, + "grad_norm": 0.3968883603719042, + "learning_rate": 2.158664121620016e-06, + "loss": 0.4774, + "step": 3948 + }, + { + "epoch": 0.7235251007695126, + "grad_norm": 0.41344274482635385, + "learning_rate": 2.1560333501357415e-06, + "loss": 0.4684, + "step": 3949 + }, + { + "epoch": 0.7237083180652254, + "grad_norm": 0.35912097103796253, + "learning_rate": 2.1534037419937525e-06, + "loss": 0.4966, + "step": 3950 + }, + { + "epoch": 0.7238915353609381, + "grad_norm": 0.45882156092291665, + "learning_rate": 2.150775298269705e-06, + "loss": 0.4807, + "step": 3951 + }, + { + "epoch": 0.7240747526566508, + "grad_norm": 0.37565207755020413, + "learning_rate": 2.1481480200387784e-06, + "loss": 0.4872, + "step": 3952 + }, + { + "epoch": 0.7242579699523635, + "grad_norm": 0.41009415053450377, + "learning_rate": 2.1455219083756785e-06, + "loss": 0.4462, + "step": 3953 + }, + { + "epoch": 0.7244411872480763, + "grad_norm": 0.39760682647944046, + "learning_rate": 2.1428969643546356e-06, + "loss": 0.4739, + "step": 3954 + }, + { + "epoch": 0.724624404543789, + "grad_norm": 0.45134579183184625, + "learning_rate": 2.140273189049396e-06, + "loss": 0.4791, + "step": 3955 + }, + { + "epoch": 0.7248076218395016, + "grad_norm": 0.3471197425655035, + "learning_rate": 2.1376505835332313e-06, + "loss": 0.4814, + "step": 3956 + }, + { + "epoch": 0.7249908391352143, + "grad_norm": 0.34004276621453156, + "learning_rate": 2.1350291488789353e-06, + "loss": 0.4583, + "step": 3957 + }, + { + "epoch": 0.725174056430927, + "grad_norm": 0.3674316268274796, + "learning_rate": 2.1324088861588267e-06, + "loss": 0.4679, + "step": 3958 + }, + { + "epoch": 0.7253572737266398, + "grad_norm": 0.37639881921654234, + "learning_rate": 2.129789796444732e-06, + "loss": 0.4925, + "step": 3959 + }, + { + "epoch": 0.7255404910223525, + "grad_norm": 0.4024732893000549, + "learning_rate": 2.12717188080801e-06, + "loss": 0.4396, + "step": 3960 + }, + { + "epoch": 0.7257237083180652, + "grad_norm": 0.36148082213298177, + "learning_rate": 2.124555140319538e-06, + "loss": 0.4581, + "step": 3961 + }, + { + "epoch": 0.7259069256137779, + "grad_norm": 0.4094140684804306, + "learning_rate": 2.1219395760497077e-06, + "loss": 0.4671, + "step": 3962 + }, + { + "epoch": 0.7260901429094907, + "grad_norm": 0.3663010278600634, + "learning_rate": 2.1193251890684297e-06, + "loss": 0.4639, + "step": 3963 + }, + { + "epoch": 0.7262733602052034, + "grad_norm": 0.4369992565379576, + "learning_rate": 2.1167119804451377e-06, + "loss": 0.4617, + "step": 3964 + }, + { + "epoch": 0.7264565775009161, + "grad_norm": 0.3460240340869184, + "learning_rate": 2.1140999512487826e-06, + "loss": 0.4494, + "step": 3965 + }, + { + "epoch": 0.7266397947966288, + "grad_norm": 0.38418566572409013, + "learning_rate": 2.1114891025478297e-06, + "loss": 0.4776, + "step": 3966 + }, + { + "epoch": 0.7268230120923416, + "grad_norm": 0.41207635226932526, + "learning_rate": 2.108879435410261e-06, + "loss": 0.4894, + "step": 3967 + }, + { + "epoch": 0.7270062293880543, + "grad_norm": 0.3583196609746141, + "learning_rate": 2.1062709509035783e-06, + "loss": 0.4804, + "step": 3968 + }, + { + "epoch": 0.727189446683767, + "grad_norm": 0.3456199452833048, + "learning_rate": 2.103663650094801e-06, + "loss": 0.4973, + "step": 3969 + }, + { + "epoch": 0.7273726639794796, + "grad_norm": 0.34378933684805224, + "learning_rate": 2.1010575340504594e-06, + "loss": 0.449, + "step": 3970 + }, + { + "epoch": 0.7275558812751923, + "grad_norm": 0.35498055821095786, + "learning_rate": 2.0984526038366005e-06, + "loss": 0.4579, + "step": 3971 + }, + { + "epoch": 0.7277390985709051, + "grad_norm": 0.41880826479222205, + "learning_rate": 2.0958488605187883e-06, + "loss": 0.47, + "step": 3972 + }, + { + "epoch": 0.7279223158666178, + "grad_norm": 0.3578587549836757, + "learning_rate": 2.093246305162102e-06, + "loss": 0.4772, + "step": 3973 + }, + { + "epoch": 0.7281055331623305, + "grad_norm": 0.34757539141396526, + "learning_rate": 2.0906449388311317e-06, + "loss": 0.4786, + "step": 3974 + }, + { + "epoch": 0.7282887504580432, + "grad_norm": 0.3797612718167418, + "learning_rate": 2.0880447625899806e-06, + "loss": 0.4954, + "step": 3975 + }, + { + "epoch": 0.728471967753756, + "grad_norm": 0.38903246942370673, + "learning_rate": 2.0854457775022693e-06, + "loss": 0.4714, + "step": 3976 + }, + { + "epoch": 0.7286551850494687, + "grad_norm": 0.34078803659969625, + "learning_rate": 2.0828479846311295e-06, + "loss": 0.4769, + "step": 3977 + }, + { + "epoch": 0.7288384023451814, + "grad_norm": 0.39415553649838064, + "learning_rate": 2.0802513850392036e-06, + "loss": 0.4742, + "step": 3978 + }, + { + "epoch": 0.7290216196408941, + "grad_norm": 0.3358841455697702, + "learning_rate": 2.0776559797886445e-06, + "loss": 0.4613, + "step": 3979 + }, + { + "epoch": 0.7292048369366069, + "grad_norm": 0.8250284818178605, + "learning_rate": 2.075061769941123e-06, + "loss": 0.5048, + "step": 3980 + }, + { + "epoch": 0.7293880542323196, + "grad_norm": 0.5376183820817682, + "learning_rate": 2.0724687565578134e-06, + "loss": 0.4823, + "step": 3981 + }, + { + "epoch": 0.7295712715280323, + "grad_norm": 0.3611278436452834, + "learning_rate": 2.0698769406994066e-06, + "loss": 0.4729, + "step": 3982 + }, + { + "epoch": 0.7297544888237449, + "grad_norm": 0.34343766283787625, + "learning_rate": 2.067286323426098e-06, + "loss": 0.4447, + "step": 3983 + }, + { + "epoch": 0.7299377061194576, + "grad_norm": 0.3645991950693383, + "learning_rate": 2.0646969057976e-06, + "loss": 0.4459, + "step": 3984 + }, + { + "epoch": 0.7301209234151704, + "grad_norm": 0.3656627102370819, + "learning_rate": 2.0621086888731257e-06, + "loss": 0.4533, + "step": 3985 + }, + { + "epoch": 0.7303041407108831, + "grad_norm": 0.3694791196027567, + "learning_rate": 2.059521673711405e-06, + "loss": 0.4487, + "step": 3986 + }, + { + "epoch": 0.7304873580065958, + "grad_norm": 0.3764444533298445, + "learning_rate": 2.0569358613706685e-06, + "loss": 0.4924, + "step": 3987 + }, + { + "epoch": 0.7306705753023085, + "grad_norm": 0.32858535408254086, + "learning_rate": 2.0543512529086634e-06, + "loss": 0.4445, + "step": 3988 + }, + { + "epoch": 0.7308537925980213, + "grad_norm": 0.37648795100517984, + "learning_rate": 2.051767849382636e-06, + "loss": 0.484, + "step": 3989 + }, + { + "epoch": 0.731037009893734, + "grad_norm": 0.3688899941410821, + "learning_rate": 2.049185651849347e-06, + "loss": 0.4874, + "step": 3990 + }, + { + "epoch": 0.7312202271894467, + "grad_norm": 0.3244483523144509, + "learning_rate": 2.046604661365057e-06, + "loss": 0.4444, + "step": 3991 + }, + { + "epoch": 0.7314034444851594, + "grad_norm": 0.3541469099575952, + "learning_rate": 2.04402487898554e-06, + "loss": 0.519, + "step": 3992 + }, + { + "epoch": 0.7315866617808721, + "grad_norm": 0.42664644819369096, + "learning_rate": 2.041446305766068e-06, + "loss": 0.4557, + "step": 3993 + }, + { + "epoch": 0.7317698790765849, + "grad_norm": 0.4203285626240392, + "learning_rate": 2.038868942761426e-06, + "loss": 0.4995, + "step": 3994 + }, + { + "epoch": 0.7319530963722976, + "grad_norm": 0.3509889902121387, + "learning_rate": 2.036292791025899e-06, + "loss": 0.4839, + "step": 3995 + }, + { + "epoch": 0.7321363136680102, + "grad_norm": 0.3364279899808081, + "learning_rate": 2.033717851613276e-06, + "loss": 0.4748, + "step": 3996 + }, + { + "epoch": 0.7323195309637229, + "grad_norm": 0.3893694640258706, + "learning_rate": 2.0311441255768538e-06, + "loss": 0.4594, + "step": 3997 + }, + { + "epoch": 0.7325027482594357, + "grad_norm": 0.36140113659251494, + "learning_rate": 2.028571613969433e-06, + "loss": 0.4472, + "step": 3998 + }, + { + "epoch": 0.7326859655551484, + "grad_norm": 0.3595150609390052, + "learning_rate": 2.0260003178433136e-06, + "loss": 0.4729, + "step": 3999 + }, + { + "epoch": 0.7328691828508611, + "grad_norm": 0.3657197643744775, + "learning_rate": 2.023430238250298e-06, + "loss": 0.4748, + "step": 4000 + }, + { + "epoch": 0.7330524001465738, + "grad_norm": 0.3505033676142658, + "learning_rate": 2.0208613762416963e-06, + "loss": 0.4462, + "step": 4001 + }, + { + "epoch": 0.7332356174422866, + "grad_norm": 0.5841692514255732, + "learning_rate": 2.0182937328683175e-06, + "loss": 0.4952, + "step": 4002 + }, + { + "epoch": 0.7334188347379993, + "grad_norm": 0.4021075121985259, + "learning_rate": 2.0157273091804725e-06, + "loss": 0.489, + "step": 4003 + }, + { + "epoch": 0.733602052033712, + "grad_norm": 0.35490571696421497, + "learning_rate": 2.01316210622797e-06, + "loss": 0.4254, + "step": 4004 + }, + { + "epoch": 0.7337852693294247, + "grad_norm": 0.3665236313827954, + "learning_rate": 2.010598125060124e-06, + "loss": 0.4885, + "step": 4005 + }, + { + "epoch": 0.7339684866251374, + "grad_norm": 0.38965802996637944, + "learning_rate": 2.008035366725748e-06, + "loss": 0.4498, + "step": 4006 + }, + { + "epoch": 0.7341517039208502, + "grad_norm": 0.42966416457191475, + "learning_rate": 2.0054738322731538e-06, + "loss": 0.4795, + "step": 4007 + }, + { + "epoch": 0.7343349212165629, + "grad_norm": 0.33424248444325466, + "learning_rate": 2.0029135227501506e-06, + "loss": 0.4442, + "step": 4008 + }, + { + "epoch": 0.7345181385122755, + "grad_norm": 0.3621404545566873, + "learning_rate": 2.0003544392040515e-06, + "loss": 0.4796, + "step": 4009 + }, + { + "epoch": 0.7347013558079882, + "grad_norm": 0.41861203487654636, + "learning_rate": 1.9977965826816682e-06, + "loss": 0.4507, + "step": 4010 + }, + { + "epoch": 0.734884573103701, + "grad_norm": 0.3598241247193229, + "learning_rate": 1.995239954229301e-06, + "loss": 0.5101, + "step": 4011 + }, + { + "epoch": 0.7350677903994137, + "grad_norm": 0.39240903487141304, + "learning_rate": 1.992684554892758e-06, + "loss": 0.4836, + "step": 4012 + }, + { + "epoch": 0.7352510076951264, + "grad_norm": 0.35814741948156764, + "learning_rate": 1.9901303857173427e-06, + "loss": 0.4518, + "step": 4013 + }, + { + "epoch": 0.7354342249908391, + "grad_norm": 0.40446364976945115, + "learning_rate": 1.9875774477478554e-06, + "loss": 0.4993, + "step": 4014 + }, + { + "epoch": 0.7356174422865519, + "grad_norm": 0.3411751401932077, + "learning_rate": 1.9850257420285853e-06, + "loss": 0.4608, + "step": 4015 + }, + { + "epoch": 0.7358006595822646, + "grad_norm": 0.37456578839867344, + "learning_rate": 1.9824752696033272e-06, + "loss": 0.4845, + "step": 4016 + }, + { + "epoch": 0.7359838768779773, + "grad_norm": 0.40766576507833674, + "learning_rate": 1.979926031515369e-06, + "loss": 0.4681, + "step": 4017 + }, + { + "epoch": 0.73616709417369, + "grad_norm": 0.3767413594158177, + "learning_rate": 1.9773780288074907e-06, + "loss": 0.468, + "step": 4018 + }, + { + "epoch": 0.7363503114694027, + "grad_norm": 0.35839677107164547, + "learning_rate": 1.9748312625219674e-06, + "loss": 0.4823, + "step": 4019 + }, + { + "epoch": 0.7365335287651155, + "grad_norm": 0.3882541817695617, + "learning_rate": 1.972285733700571e-06, + "loss": 0.492, + "step": 4020 + }, + { + "epoch": 0.7367167460608282, + "grad_norm": 0.3794713824787052, + "learning_rate": 1.969741443384568e-06, + "loss": 0.4566, + "step": 4021 + }, + { + "epoch": 0.7368999633565408, + "grad_norm": 0.36678362989099145, + "learning_rate": 1.9671983926147143e-06, + "loss": 0.4528, + "step": 4022 + }, + { + "epoch": 0.7370831806522535, + "grad_norm": 0.3978770489237176, + "learning_rate": 1.964656582431259e-06, + "loss": 0.4845, + "step": 4023 + }, + { + "epoch": 0.7372663979479663, + "grad_norm": 0.3452119681101174, + "learning_rate": 1.962116013873947e-06, + "loss": 0.4587, + "step": 4024 + }, + { + "epoch": 0.737449615243679, + "grad_norm": 0.3648561806801237, + "learning_rate": 1.959576687982016e-06, + "loss": 0.4646, + "step": 4025 + }, + { + "epoch": 0.7376328325393917, + "grad_norm": 0.3969117613318373, + "learning_rate": 1.95703860579419e-06, + "loss": 0.4687, + "step": 4026 + }, + { + "epoch": 0.7378160498351044, + "grad_norm": 0.4850066093035532, + "learning_rate": 1.954501768348687e-06, + "loss": 0.4819, + "step": 4027 + }, + { + "epoch": 0.7379992671308171, + "grad_norm": 0.39266238773846035, + "learning_rate": 1.9519661766832174e-06, + "loss": 0.4641, + "step": 4028 + }, + { + "epoch": 0.7381824844265299, + "grad_norm": 0.31646845520624245, + "learning_rate": 1.9494318318349825e-06, + "loss": 0.4385, + "step": 4029 + }, + { + "epoch": 0.7383657017222426, + "grad_norm": 0.35566510994887, + "learning_rate": 1.9468987348406703e-06, + "loss": 0.474, + "step": 4030 + }, + { + "epoch": 0.7385489190179553, + "grad_norm": 0.3872354487936671, + "learning_rate": 1.944366886736458e-06, + "loss": 0.4657, + "step": 4031 + }, + { + "epoch": 0.738732136313668, + "grad_norm": 0.3804177713529073, + "learning_rate": 1.9418362885580174e-06, + "loss": 0.4618, + "step": 4032 + }, + { + "epoch": 0.7389153536093808, + "grad_norm": 0.35257179808793176, + "learning_rate": 1.939306941340501e-06, + "loss": 0.4571, + "step": 4033 + }, + { + "epoch": 0.7390985709050935, + "grad_norm": 0.38170426112860006, + "learning_rate": 1.9367788461185596e-06, + "loss": 0.4784, + "step": 4034 + }, + { + "epoch": 0.7392817882008061, + "grad_norm": 0.34927534676270283, + "learning_rate": 1.9342520039263206e-06, + "loss": 0.4731, + "step": 4035 + }, + { + "epoch": 0.7394650054965188, + "grad_norm": 0.39851619679887534, + "learning_rate": 1.93172641579741e-06, + "loss": 0.488, + "step": 4036 + }, + { + "epoch": 0.7396482227922316, + "grad_norm": 0.3241777037878778, + "learning_rate": 1.9292020827649303e-06, + "loss": 0.4538, + "step": 4037 + }, + { + "epoch": 0.7398314400879443, + "grad_norm": 0.9937331158551702, + "learning_rate": 1.9266790058614805e-06, + "loss": 0.4691, + "step": 4038 + }, + { + "epoch": 0.740014657383657, + "grad_norm": 0.33552160080546933, + "learning_rate": 1.924157186119136e-06, + "loss": 0.4594, + "step": 4039 + }, + { + "epoch": 0.7401978746793697, + "grad_norm": 0.32776676443434455, + "learning_rate": 1.9216366245694673e-06, + "loss": 0.4456, + "step": 4040 + }, + { + "epoch": 0.7403810919750824, + "grad_norm": 0.36270164983310027, + "learning_rate": 1.9191173222435223e-06, + "loss": 0.4541, + "step": 4041 + }, + { + "epoch": 0.7405643092707952, + "grad_norm": 0.34997719170583524, + "learning_rate": 1.916599280171841e-06, + "loss": 0.4682, + "step": 4042 + }, + { + "epoch": 0.7407475265665079, + "grad_norm": 0.34361373471099194, + "learning_rate": 1.9140824993844396e-06, + "loss": 0.4686, + "step": 4043 + }, + { + "epoch": 0.7409307438622206, + "grad_norm": 0.39433678985217036, + "learning_rate": 1.9115669809108278e-06, + "loss": 0.4672, + "step": 4044 + }, + { + "epoch": 0.7411139611579333, + "grad_norm": 0.4019491878484139, + "learning_rate": 1.9090527257799903e-06, + "loss": 0.4964, + "step": 4045 + }, + { + "epoch": 0.7412971784536461, + "grad_norm": 0.39612212196419777, + "learning_rate": 1.9065397350204013e-06, + "loss": 0.4277, + "step": 4046 + }, + { + "epoch": 0.7414803957493588, + "grad_norm": 0.6062012031418743, + "learning_rate": 1.9040280096600144e-06, + "loss": 0.4383, + "step": 4047 + }, + { + "epoch": 0.7416636130450714, + "grad_norm": 0.43624618906610063, + "learning_rate": 1.9015175507262656e-06, + "loss": 0.4849, + "step": 4048 + }, + { + "epoch": 0.7418468303407841, + "grad_norm": 0.35337558329567226, + "learning_rate": 1.899008359246074e-06, + "loss": 0.4386, + "step": 4049 + }, + { + "epoch": 0.7420300476364969, + "grad_norm": 0.44416886111322623, + "learning_rate": 1.8965004362458423e-06, + "loss": 0.4181, + "step": 4050 + }, + { + "epoch": 0.7422132649322096, + "grad_norm": 0.4095745408930984, + "learning_rate": 1.8939937827514509e-06, + "loss": 0.4758, + "step": 4051 + }, + { + "epoch": 0.7423964822279223, + "grad_norm": 0.36684395263301556, + "learning_rate": 1.8914883997882594e-06, + "loss": 0.4812, + "step": 4052 + }, + { + "epoch": 0.742579699523635, + "grad_norm": 0.3521988380162774, + "learning_rate": 1.8889842883811127e-06, + "loss": 0.4476, + "step": 4053 + }, + { + "epoch": 0.7427629168193477, + "grad_norm": 0.37018266065758226, + "learning_rate": 1.8864814495543344e-06, + "loss": 0.4638, + "step": 4054 + }, + { + "epoch": 0.7429461341150605, + "grad_norm": 0.34315046413477657, + "learning_rate": 1.8839798843317247e-06, + "loss": 0.4521, + "step": 4055 + }, + { + "epoch": 0.7431293514107732, + "grad_norm": 0.48373817790218515, + "learning_rate": 1.8814795937365632e-06, + "loss": 0.4851, + "step": 4056 + }, + { + "epoch": 0.7433125687064859, + "grad_norm": 0.4796010085771966, + "learning_rate": 1.8789805787916105e-06, + "loss": 0.4599, + "step": 4057 + }, + { + "epoch": 0.7434957860021986, + "grad_norm": 0.351900140603146, + "learning_rate": 1.8764828405191066e-06, + "loss": 0.459, + "step": 4058 + }, + { + "epoch": 0.7436790032979114, + "grad_norm": 0.365491528800529, + "learning_rate": 1.8739863799407644e-06, + "loss": 0.4868, + "step": 4059 + }, + { + "epoch": 0.7438622205936241, + "grad_norm": 0.38751813349730896, + "learning_rate": 1.8714911980777757e-06, + "loss": 0.4315, + "step": 4060 + }, + { + "epoch": 0.7440454378893367, + "grad_norm": 0.3580242193371107, + "learning_rate": 1.8689972959508113e-06, + "loss": 0.4805, + "step": 4061 + }, + { + "epoch": 0.7442286551850494, + "grad_norm": 0.4509351696375182, + "learning_rate": 1.866504674580021e-06, + "loss": 0.4709, + "step": 4062 + }, + { + "epoch": 0.7444118724807622, + "grad_norm": 0.3674083726141402, + "learning_rate": 1.8640133349850203e-06, + "loss": 0.4782, + "step": 4063 + }, + { + "epoch": 0.7445950897764749, + "grad_norm": 0.40375539880733496, + "learning_rate": 1.8615232781849107e-06, + "loss": 0.4537, + "step": 4064 + }, + { + "epoch": 0.7447783070721876, + "grad_norm": 0.34494006412033174, + "learning_rate": 1.8590345051982657e-06, + "loss": 0.4352, + "step": 4065 + }, + { + "epoch": 0.7449615243679003, + "grad_norm": 0.36816174185686534, + "learning_rate": 1.8565470170431365e-06, + "loss": 0.4585, + "step": 4066 + }, + { + "epoch": 0.745144741663613, + "grad_norm": 0.36293283553495814, + "learning_rate": 1.8540608147370386e-06, + "loss": 0.4874, + "step": 4067 + }, + { + "epoch": 0.7453279589593258, + "grad_norm": 0.3619828856712631, + "learning_rate": 1.8515758992969724e-06, + "loss": 0.483, + "step": 4068 + }, + { + "epoch": 0.7455111762550385, + "grad_norm": 0.4092780416874802, + "learning_rate": 1.8490922717394094e-06, + "loss": 0.4872, + "step": 4069 + }, + { + "epoch": 0.7456943935507512, + "grad_norm": 0.3413826003401774, + "learning_rate": 1.846609933080291e-06, + "loss": 0.4741, + "step": 4070 + }, + { + "epoch": 0.7458776108464639, + "grad_norm": 0.43626629769081204, + "learning_rate": 1.8441288843350325e-06, + "loss": 0.4793, + "step": 4071 + }, + { + "epoch": 0.7460608281421767, + "grad_norm": 0.44933278447276864, + "learning_rate": 1.8416491265185233e-06, + "loss": 0.447, + "step": 4072 + }, + { + "epoch": 0.7462440454378894, + "grad_norm": 0.4298809770026864, + "learning_rate": 1.839170660645126e-06, + "loss": 0.3847, + "step": 4073 + }, + { + "epoch": 0.746427262733602, + "grad_norm": 0.3917091027821756, + "learning_rate": 1.8366934877286702e-06, + "loss": 0.4648, + "step": 4074 + }, + { + "epoch": 0.7466104800293147, + "grad_norm": 0.3388982273699895, + "learning_rate": 1.8342176087824576e-06, + "loss": 0.448, + "step": 4075 + }, + { + "epoch": 0.7467936973250274, + "grad_norm": 0.4978974652300739, + "learning_rate": 1.8317430248192636e-06, + "loss": 0.4953, + "step": 4076 + }, + { + "epoch": 0.7469769146207402, + "grad_norm": 0.3324605261370153, + "learning_rate": 1.8292697368513335e-06, + "loss": 0.48, + "step": 4077 + }, + { + "epoch": 0.7471601319164529, + "grad_norm": 0.3205253612405013, + "learning_rate": 1.82679774589038e-06, + "loss": 0.4819, + "step": 4078 + }, + { + "epoch": 0.7473433492121656, + "grad_norm": 0.5099369257618026, + "learning_rate": 1.8243270529475848e-06, + "loss": 0.481, + "step": 4079 + }, + { + "epoch": 0.7475265665078783, + "grad_norm": 0.3736030875841353, + "learning_rate": 1.8218576590336017e-06, + "loss": 0.4668, + "step": 4080 + }, + { + "epoch": 0.7477097838035911, + "grad_norm": 0.39388685721322414, + "learning_rate": 1.819389565158553e-06, + "loss": 0.456, + "step": 4081 + }, + { + "epoch": 0.7478930010993038, + "grad_norm": 0.3735944138776285, + "learning_rate": 1.8169227723320265e-06, + "loss": 0.4298, + "step": 4082 + }, + { + "epoch": 0.7480762183950165, + "grad_norm": 0.3767625966098096, + "learning_rate": 1.814457281563078e-06, + "loss": 0.4767, + "step": 4083 + }, + { + "epoch": 0.7482594356907292, + "grad_norm": 0.3857979642189771, + "learning_rate": 1.8119930938602342e-06, + "loss": 0.4628, + "step": 4084 + }, + { + "epoch": 0.748442652986442, + "grad_norm": 0.34594760810350345, + "learning_rate": 1.8095302102314843e-06, + "loss": 0.4662, + "step": 4085 + }, + { + "epoch": 0.7486258702821547, + "grad_norm": 0.39093168935002987, + "learning_rate": 1.8070686316842883e-06, + "loss": 0.4419, + "step": 4086 + }, + { + "epoch": 0.7488090875778673, + "grad_norm": 0.3289049801215156, + "learning_rate": 1.8046083592255676e-06, + "loss": 0.4332, + "step": 4087 + }, + { + "epoch": 0.74899230487358, + "grad_norm": 0.41846397995876716, + "learning_rate": 1.8021493938617152e-06, + "loss": 0.4947, + "step": 4088 + }, + { + "epoch": 0.7491755221692927, + "grad_norm": 0.37197853344196835, + "learning_rate": 1.7996917365985827e-06, + "loss": 0.4386, + "step": 4089 + }, + { + "epoch": 0.7493587394650055, + "grad_norm": 0.3746935512002778, + "learning_rate": 1.7972353884414945e-06, + "loss": 0.4684, + "step": 4090 + }, + { + "epoch": 0.7495419567607182, + "grad_norm": 0.3976089073534478, + "learning_rate": 1.7947803503952298e-06, + "loss": 0.4731, + "step": 4091 + }, + { + "epoch": 0.7497251740564309, + "grad_norm": 0.3779397969449795, + "learning_rate": 1.7923266234640424e-06, + "loss": 0.4855, + "step": 4092 + }, + { + "epoch": 0.7499083913521436, + "grad_norm": 0.40852890198802416, + "learning_rate": 1.7898742086516403e-06, + "loss": 0.4458, + "step": 4093 + }, + { + "epoch": 0.7500916086478564, + "grad_norm": 0.3648978647311167, + "learning_rate": 1.7874231069612024e-06, + "loss": 0.4566, + "step": 4094 + }, + { + "epoch": 0.7502748259435691, + "grad_norm": 0.5140950428078369, + "learning_rate": 1.7849733193953644e-06, + "loss": 0.4224, + "step": 4095 + }, + { + "epoch": 0.7504580432392818, + "grad_norm": 0.35408155389517404, + "learning_rate": 1.7825248469562301e-06, + "loss": 0.4585, + "step": 4096 + }, + { + "epoch": 0.7506412605349945, + "grad_norm": 0.35397433582325016, + "learning_rate": 1.7800776906453587e-06, + "loss": 0.4519, + "step": 4097 + }, + { + "epoch": 0.7508244778307073, + "grad_norm": 0.3558271730882, + "learning_rate": 1.777631851463779e-06, + "loss": 0.4792, + "step": 4098 + }, + { + "epoch": 0.75100769512642, + "grad_norm": 0.48459300225858803, + "learning_rate": 1.7751873304119743e-06, + "loss": 0.4756, + "step": 4099 + }, + { + "epoch": 0.7511909124221327, + "grad_norm": 0.47309599249739553, + "learning_rate": 1.77274412848989e-06, + "loss": 0.4762, + "step": 4100 + }, + { + "epoch": 0.7513741297178453, + "grad_norm": 0.402871823866162, + "learning_rate": 1.7703022466969344e-06, + "loss": 0.4702, + "step": 4101 + }, + { + "epoch": 0.751557347013558, + "grad_norm": 0.3469509938178955, + "learning_rate": 1.767861686031976e-06, + "loss": 0.4578, + "step": 4102 + }, + { + "epoch": 0.7517405643092708, + "grad_norm": 0.42604410490007893, + "learning_rate": 1.7654224474933402e-06, + "loss": 0.4669, + "step": 4103 + }, + { + "epoch": 0.7519237816049835, + "grad_norm": 0.4221122705392122, + "learning_rate": 1.7629845320788108e-06, + "loss": 0.4669, + "step": 4104 + }, + { + "epoch": 0.7521069989006962, + "grad_norm": 0.34060667846008097, + "learning_rate": 1.7605479407856336e-06, + "loss": 0.4598, + "step": 4105 + }, + { + "epoch": 0.7522902161964089, + "grad_norm": 0.3594956302533286, + "learning_rate": 1.7581126746105138e-06, + "loss": 0.4643, + "step": 4106 + }, + { + "epoch": 0.7524734334921217, + "grad_norm": 0.3418664593987484, + "learning_rate": 1.7556787345496102e-06, + "loss": 0.4796, + "step": 4107 + }, + { + "epoch": 0.7526566507878344, + "grad_norm": 0.3578518020006305, + "learning_rate": 1.7532461215985396e-06, + "loss": 0.4615, + "step": 4108 + }, + { + "epoch": 0.7528398680835471, + "grad_norm": 0.3703679763002126, + "learning_rate": 1.7508148367523792e-06, + "loss": 0.4872, + "step": 4109 + }, + { + "epoch": 0.7530230853792598, + "grad_norm": 0.34518380968447654, + "learning_rate": 1.7483848810056626e-06, + "loss": 0.4249, + "step": 4110 + }, + { + "epoch": 0.7532063026749726, + "grad_norm": 0.40315432836820303, + "learning_rate": 1.7459562553523768e-06, + "loss": 0.4928, + "step": 4111 + }, + { + "epoch": 0.7533895199706853, + "grad_norm": 0.3328304893922875, + "learning_rate": 1.7435289607859646e-06, + "loss": 0.4698, + "step": 4112 + }, + { + "epoch": 0.753572737266398, + "grad_norm": 0.36319375008876503, + "learning_rate": 1.741102998299327e-06, + "loss": 0.4621, + "step": 4113 + }, + { + "epoch": 0.7537559545621106, + "grad_norm": 0.4311280052755772, + "learning_rate": 1.7386783688848225e-06, + "loss": 0.4603, + "step": 4114 + }, + { + "epoch": 0.7539391718578233, + "grad_norm": 0.3375728277767886, + "learning_rate": 1.7362550735342575e-06, + "loss": 0.4525, + "step": 4115 + }, + { + "epoch": 0.7541223891535361, + "grad_norm": 0.35346560552592615, + "learning_rate": 1.7338331132388952e-06, + "loss": 0.4704, + "step": 4116 + }, + { + "epoch": 0.7543056064492488, + "grad_norm": 0.32490533390383314, + "learning_rate": 1.7314124889894552e-06, + "loss": 0.464, + "step": 4117 + }, + { + "epoch": 0.7544888237449615, + "grad_norm": 0.4028596236830994, + "learning_rate": 1.7289932017761118e-06, + "loss": 0.4471, + "step": 4118 + }, + { + "epoch": 0.7546720410406742, + "grad_norm": 0.34021366956503474, + "learning_rate": 1.7265752525884833e-06, + "loss": 0.4539, + "step": 4119 + }, + { + "epoch": 0.754855258336387, + "grad_norm": 0.33789436101613624, + "learning_rate": 1.7241586424156497e-06, + "loss": 0.4686, + "step": 4120 + }, + { + "epoch": 0.7550384756320997, + "grad_norm": 0.3523250942278014, + "learning_rate": 1.7217433722461429e-06, + "loss": 0.4613, + "step": 4121 + }, + { + "epoch": 0.7552216929278124, + "grad_norm": 0.33966744608529686, + "learning_rate": 1.719329443067942e-06, + "loss": 0.4957, + "step": 4122 + }, + { + "epoch": 0.7554049102235251, + "grad_norm": 0.3774558497554812, + "learning_rate": 1.7169168558684784e-06, + "loss": 0.4939, + "step": 4123 + }, + { + "epoch": 0.7555881275192378, + "grad_norm": 0.397659556575842, + "learning_rate": 1.714505611634637e-06, + "loss": 0.4634, + "step": 4124 + }, + { + "epoch": 0.7557713448149506, + "grad_norm": 0.3849861681355583, + "learning_rate": 1.7120957113527548e-06, + "loss": 0.4554, + "step": 4125 + }, + { + "epoch": 0.7559545621106633, + "grad_norm": 0.4422282038794372, + "learning_rate": 1.7096871560086149e-06, + "loss": 0.4658, + "step": 4126 + }, + { + "epoch": 0.7561377794063759, + "grad_norm": 0.3782777153652058, + "learning_rate": 1.707279946587449e-06, + "loss": 0.4397, + "step": 4127 + }, + { + "epoch": 0.7563209967020886, + "grad_norm": 0.35840358360089986, + "learning_rate": 1.7048740840739436e-06, + "loss": 0.4651, + "step": 4128 + }, + { + "epoch": 0.7565042139978014, + "grad_norm": 0.3580145314485348, + "learning_rate": 1.7024695694522337e-06, + "loss": 0.4868, + "step": 4129 + }, + { + "epoch": 0.7566874312935141, + "grad_norm": 0.4588081825112617, + "learning_rate": 1.7000664037058989e-06, + "loss": 0.4688, + "step": 4130 + }, + { + "epoch": 0.7568706485892268, + "grad_norm": 0.3925924294224504, + "learning_rate": 1.6976645878179677e-06, + "loss": 0.4774, + "step": 4131 + }, + { + "epoch": 0.7570538658849395, + "grad_norm": 0.3264787491206452, + "learning_rate": 1.6952641227709188e-06, + "loss": 0.4689, + "step": 4132 + }, + { + "epoch": 0.7572370831806523, + "grad_norm": 0.34950702301533965, + "learning_rate": 1.6928650095466803e-06, + "loss": 0.4713, + "step": 4133 + }, + { + "epoch": 0.757420300476365, + "grad_norm": 0.38322660772581035, + "learning_rate": 1.6904672491266217e-06, + "loss": 0.4922, + "step": 4134 + }, + { + "epoch": 0.7576035177720777, + "grad_norm": 0.36738334003208656, + "learning_rate": 1.6880708424915604e-06, + "loss": 0.4862, + "step": 4135 + }, + { + "epoch": 0.7577867350677904, + "grad_norm": 0.40877370078560427, + "learning_rate": 1.6856757906217652e-06, + "loss": 0.4977, + "step": 4136 + }, + { + "epoch": 0.7579699523635031, + "grad_norm": 0.33709017913109657, + "learning_rate": 1.6832820944969435e-06, + "loss": 0.4987, + "step": 4137 + }, + { + "epoch": 0.7581531696592159, + "grad_norm": 0.36675015101417785, + "learning_rate": 1.680889755096255e-06, + "loss": 0.4654, + "step": 4138 + }, + { + "epoch": 0.7583363869549286, + "grad_norm": 0.39254403639733965, + "learning_rate": 1.6784987733982978e-06, + "loss": 0.471, + "step": 4139 + }, + { + "epoch": 0.7585196042506412, + "grad_norm": 0.3722507602228711, + "learning_rate": 1.6761091503811204e-06, + "loss": 0.4886, + "step": 4140 + }, + { + "epoch": 0.7587028215463539, + "grad_norm": 0.5932535007558865, + "learning_rate": 1.6737208870222105e-06, + "loss": 0.4683, + "step": 4141 + }, + { + "epoch": 0.7588860388420667, + "grad_norm": 0.3839548933850747, + "learning_rate": 1.6713339842985048e-06, + "loss": 0.4956, + "step": 4142 + }, + { + "epoch": 0.7590692561377794, + "grad_norm": 0.46206809633520224, + "learning_rate": 1.6689484431863785e-06, + "loss": 0.4615, + "step": 4143 + }, + { + "epoch": 0.7592524734334921, + "grad_norm": 0.45184103157453875, + "learning_rate": 1.6665642646616543e-06, + "loss": 0.5115, + "step": 4144 + }, + { + "epoch": 0.7594356907292048, + "grad_norm": 0.35105233040858413, + "learning_rate": 1.6641814496995923e-06, + "loss": 0.4552, + "step": 4145 + }, + { + "epoch": 0.7596189080249176, + "grad_norm": 0.4067670583647117, + "learning_rate": 1.6617999992749018e-06, + "loss": 0.423, + "step": 4146 + }, + { + "epoch": 0.7598021253206303, + "grad_norm": 0.3320446448998926, + "learning_rate": 1.6594199143617252e-06, + "loss": 0.4698, + "step": 4147 + }, + { + "epoch": 0.759985342616343, + "grad_norm": 0.3456835617701573, + "learning_rate": 1.6570411959336558e-06, + "loss": 0.4348, + "step": 4148 + }, + { + "epoch": 0.7601685599120557, + "grad_norm": 0.35509330605599865, + "learning_rate": 1.6546638449637187e-06, + "loss": 0.4897, + "step": 4149 + }, + { + "epoch": 0.7603517772077684, + "grad_norm": 0.39661692300679996, + "learning_rate": 1.6522878624243881e-06, + "loss": 0.4486, + "step": 4150 + }, + { + "epoch": 0.7605349945034812, + "grad_norm": 0.3464123244994105, + "learning_rate": 1.6499132492875712e-06, + "loss": 0.453, + "step": 4151 + }, + { + "epoch": 0.7607182117991939, + "grad_norm": 0.3539553427373948, + "learning_rate": 1.6475400065246217e-06, + "loss": 0.4907, + "step": 4152 + }, + { + "epoch": 0.7609014290949065, + "grad_norm": 0.3533453346303954, + "learning_rate": 1.6451681351063258e-06, + "loss": 0.4509, + "step": 4153 + }, + { + "epoch": 0.7610846463906192, + "grad_norm": 0.40404966516344104, + "learning_rate": 1.6427976360029151e-06, + "loss": 0.4804, + "step": 4154 + }, + { + "epoch": 0.761267863686332, + "grad_norm": 0.3750616991341421, + "learning_rate": 1.6404285101840567e-06, + "loss": 0.4466, + "step": 4155 + }, + { + "epoch": 0.7614510809820447, + "grad_norm": 0.4456100894903707, + "learning_rate": 1.6380607586188528e-06, + "loss": 0.4605, + "step": 4156 + }, + { + "epoch": 0.7616342982777574, + "grad_norm": 0.4352429161943705, + "learning_rate": 1.635694382275851e-06, + "loss": 0.4791, + "step": 4157 + }, + { + "epoch": 0.7618175155734701, + "grad_norm": 0.36615705419495553, + "learning_rate": 1.633329382123031e-06, + "loss": 0.4606, + "step": 4158 + }, + { + "epoch": 0.7620007328691829, + "grad_norm": 0.3801812307833013, + "learning_rate": 1.6309657591278117e-06, + "loss": 0.4785, + "step": 4159 + }, + { + "epoch": 0.7621839501648956, + "grad_norm": 0.38130134781278163, + "learning_rate": 1.628603514257045e-06, + "loss": 0.446, + "step": 4160 + }, + { + "epoch": 0.7623671674606083, + "grad_norm": 0.37170322054903127, + "learning_rate": 1.6262426484770239e-06, + "loss": 0.4857, + "step": 4161 + }, + { + "epoch": 0.762550384756321, + "grad_norm": 0.38757657832913417, + "learning_rate": 1.6238831627534768e-06, + "loss": 0.4301, + "step": 4162 + }, + { + "epoch": 0.7627336020520337, + "grad_norm": 0.3353446683382116, + "learning_rate": 1.621525058051564e-06, + "loss": 0.4387, + "step": 4163 + }, + { + "epoch": 0.7629168193477465, + "grad_norm": 0.36850221103539726, + "learning_rate": 1.6191683353358823e-06, + "loss": 0.4746, + "step": 4164 + }, + { + "epoch": 0.7631000366434592, + "grad_norm": 0.34589362198209417, + "learning_rate": 1.6168129955704641e-06, + "loss": 0.4619, + "step": 4165 + }, + { + "epoch": 0.7632832539391718, + "grad_norm": 0.36619617381064873, + "learning_rate": 1.614459039718778e-06, + "loss": 0.4985, + "step": 4166 + }, + { + "epoch": 0.7634664712348845, + "grad_norm": 0.34564674778188265, + "learning_rate": 1.6121064687437239e-06, + "loss": 0.4494, + "step": 4167 + }, + { + "epoch": 0.7636496885305973, + "grad_norm": 0.3641431122407145, + "learning_rate": 1.609755283607632e-06, + "loss": 0.4913, + "step": 4168 + }, + { + "epoch": 0.76383290582631, + "grad_norm": 0.3858309338681685, + "learning_rate": 1.6074054852722721e-06, + "loss": 0.4599, + "step": 4169 + }, + { + "epoch": 0.7640161231220227, + "grad_norm": 0.4003061856987761, + "learning_rate": 1.6050570746988469e-06, + "loss": 0.4348, + "step": 4170 + }, + { + "epoch": 0.7641993404177354, + "grad_norm": 0.33650621796585883, + "learning_rate": 1.6027100528479816e-06, + "loss": 0.4761, + "step": 4171 + }, + { + "epoch": 0.7643825577134481, + "grad_norm": 0.3269855918796489, + "learning_rate": 1.6003644206797436e-06, + "loss": 0.4811, + "step": 4172 + }, + { + "epoch": 0.7645657750091609, + "grad_norm": 0.3529860976319853, + "learning_rate": 1.5980201791536298e-06, + "loss": 0.4661, + "step": 4173 + }, + { + "epoch": 0.7647489923048736, + "grad_norm": 0.3644002955222909, + "learning_rate": 1.5956773292285655e-06, + "loss": 0.4748, + "step": 4174 + }, + { + "epoch": 0.7649322096005863, + "grad_norm": 0.314339830950356, + "learning_rate": 1.593335871862906e-06, + "loss": 0.4082, + "step": 4175 + }, + { + "epoch": 0.765115426896299, + "grad_norm": 0.41446388760044756, + "learning_rate": 1.590995808014441e-06, + "loss": 0.4712, + "step": 4176 + }, + { + "epoch": 0.7652986441920118, + "grad_norm": 0.3812122482641003, + "learning_rate": 1.5886571386403898e-06, + "loss": 0.5068, + "step": 4177 + }, + { + "epoch": 0.7654818614877245, + "grad_norm": 0.3543458154295496, + "learning_rate": 1.5863198646973976e-06, + "loss": 0.4696, + "step": 4178 + }, + { + "epoch": 0.7656650787834371, + "grad_norm": 0.40835369049952497, + "learning_rate": 1.5839839871415403e-06, + "loss": 0.4509, + "step": 4179 + }, + { + "epoch": 0.7658482960791498, + "grad_norm": 0.385025035615343, + "learning_rate": 1.5816495069283234e-06, + "loss": 0.4814, + "step": 4180 + }, + { + "epoch": 0.7660315133748626, + "grad_norm": 0.3514251086381649, + "learning_rate": 1.5793164250126835e-06, + "loss": 0.4617, + "step": 4181 + }, + { + "epoch": 0.7662147306705753, + "grad_norm": 0.4013212525588526, + "learning_rate": 1.5769847423489797e-06, + "loss": 0.5075, + "step": 4182 + }, + { + "epoch": 0.766397947966288, + "grad_norm": 0.32102329335479796, + "learning_rate": 1.5746544598910007e-06, + "loss": 0.4834, + "step": 4183 + }, + { + "epoch": 0.7665811652620007, + "grad_norm": 0.35662707150689005, + "learning_rate": 1.5723255785919634e-06, + "loss": 0.4789, + "step": 4184 + }, + { + "epoch": 0.7667643825577134, + "grad_norm": 0.33365195770538464, + "learning_rate": 1.569998099404514e-06, + "loss": 0.4808, + "step": 4185 + }, + { + "epoch": 0.7669475998534262, + "grad_norm": 0.3442342328140645, + "learning_rate": 1.5676720232807202e-06, + "loss": 0.4782, + "step": 4186 + }, + { + "epoch": 0.7671308171491389, + "grad_norm": 0.454415288803408, + "learning_rate": 1.5653473511720762e-06, + "loss": 0.4583, + "step": 4187 + }, + { + "epoch": 0.7673140344448516, + "grad_norm": 0.30923761265789995, + "learning_rate": 1.5630240840295058e-06, + "loss": 0.4145, + "step": 4188 + }, + { + "epoch": 0.7674972517405643, + "grad_norm": 0.4047111527877341, + "learning_rate": 1.560702222803357e-06, + "loss": 0.4656, + "step": 4189 + }, + { + "epoch": 0.7676804690362771, + "grad_norm": 0.3903732999900964, + "learning_rate": 1.5583817684433998e-06, + "loss": 0.4426, + "step": 4190 + }, + { + "epoch": 0.7678636863319898, + "grad_norm": 0.35865130783591553, + "learning_rate": 1.5560627218988294e-06, + "loss": 0.4828, + "step": 4191 + }, + { + "epoch": 0.7680469036277024, + "grad_norm": 0.33980355275347973, + "learning_rate": 1.5537450841182694e-06, + "loss": 0.4604, + "step": 4192 + }, + { + "epoch": 0.7682301209234151, + "grad_norm": 0.4319578260358002, + "learning_rate": 1.5514288560497608e-06, + "loss": 0.4803, + "step": 4193 + }, + { + "epoch": 0.7684133382191279, + "grad_norm": 0.39328076739054296, + "learning_rate": 1.5491140386407739e-06, + "loss": 0.5101, + "step": 4194 + }, + { + "epoch": 0.7685965555148406, + "grad_norm": 0.3735467963804508, + "learning_rate": 1.5468006328381968e-06, + "loss": 0.4466, + "step": 4195 + }, + { + "epoch": 0.7687797728105533, + "grad_norm": 0.3733647141412181, + "learning_rate": 1.5444886395883446e-06, + "loss": 0.5102, + "step": 4196 + }, + { + "epoch": 0.768962990106266, + "grad_norm": 0.442033262728887, + "learning_rate": 1.5421780598369507e-06, + "loss": 0.505, + "step": 4197 + }, + { + "epoch": 0.7691462074019787, + "grad_norm": 0.3592033145540987, + "learning_rate": 1.539868894529174e-06, + "loss": 0.4549, + "step": 4198 + }, + { + "epoch": 0.7693294246976915, + "grad_norm": 0.5395142317646816, + "learning_rate": 1.5375611446095905e-06, + "loss": 0.4287, + "step": 4199 + }, + { + "epoch": 0.7695126419934042, + "grad_norm": 0.3626109983260122, + "learning_rate": 1.5352548110222033e-06, + "loss": 0.4512, + "step": 4200 + }, + { + "epoch": 0.7696958592891169, + "grad_norm": 0.3737644873913779, + "learning_rate": 1.5329498947104287e-06, + "loss": 0.4359, + "step": 4201 + }, + { + "epoch": 0.7698790765848296, + "grad_norm": 0.36092771862388157, + "learning_rate": 1.5306463966171103e-06, + "loss": 0.4575, + "step": 4202 + }, + { + "epoch": 0.7700622938805424, + "grad_norm": 0.3311537515837117, + "learning_rate": 1.5283443176845053e-06, + "loss": 0.4253, + "step": 4203 + }, + { + "epoch": 0.7702455111762551, + "grad_norm": 0.4015094204372302, + "learning_rate": 1.5260436588542965e-06, + "loss": 0.4626, + "step": 4204 + }, + { + "epoch": 0.7704287284719677, + "grad_norm": 0.42399651635104585, + "learning_rate": 1.5237444210675795e-06, + "loss": 0.4721, + "step": 4205 + }, + { + "epoch": 0.7706119457676804, + "grad_norm": 0.3769294067423412, + "learning_rate": 1.5214466052648758e-06, + "loss": 0.4454, + "step": 4206 + }, + { + "epoch": 0.7707951630633931, + "grad_norm": 0.40721140381541165, + "learning_rate": 1.5191502123861185e-06, + "loss": 0.4635, + "step": 4207 + }, + { + "epoch": 0.7709783803591059, + "grad_norm": 0.37894938125189215, + "learning_rate": 1.5168552433706608e-06, + "loss": 0.4532, + "step": 4208 + }, + { + "epoch": 0.7711615976548186, + "grad_norm": 0.3583657848512605, + "learning_rate": 1.5145616991572747e-06, + "loss": 0.4371, + "step": 4209 + }, + { + "epoch": 0.7713448149505313, + "grad_norm": 0.3661693636774428, + "learning_rate": 1.5122695806841514e-06, + "loss": 0.4598, + "step": 4210 + }, + { + "epoch": 0.771528032246244, + "grad_norm": 0.3872875724422321, + "learning_rate": 1.509978888888894e-06, + "loss": 0.4822, + "step": 4211 + }, + { + "epoch": 0.7717112495419568, + "grad_norm": 0.38763223126774926, + "learning_rate": 1.5076896247085226e-06, + "loss": 0.4907, + "step": 4212 + }, + { + "epoch": 0.7718944668376695, + "grad_norm": 0.35769111954228433, + "learning_rate": 1.505401789079477e-06, + "loss": 0.4743, + "step": 4213 + }, + { + "epoch": 0.7720776841333822, + "grad_norm": 0.368590737774548, + "learning_rate": 1.503115382937612e-06, + "loss": 0.4714, + "step": 4214 + }, + { + "epoch": 0.7722609014290949, + "grad_norm": 0.4025154190887305, + "learning_rate": 1.500830407218195e-06, + "loss": 0.4548, + "step": 4215 + }, + { + "epoch": 0.7724441187248077, + "grad_norm": 0.4242777210067457, + "learning_rate": 1.4985468628559074e-06, + "loss": 0.4974, + "step": 4216 + }, + { + "epoch": 0.7726273360205204, + "grad_norm": 0.31382590051865583, + "learning_rate": 1.4962647507848489e-06, + "loss": 0.4106, + "step": 4217 + }, + { + "epoch": 0.772810553316233, + "grad_norm": 0.3475998819815029, + "learning_rate": 1.493984071938534e-06, + "loss": 0.468, + "step": 4218 + }, + { + "epoch": 0.7729937706119457, + "grad_norm": 0.34653937296882287, + "learning_rate": 1.4917048272498862e-06, + "loss": 0.4448, + "step": 4219 + }, + { + "epoch": 0.7731769879076584, + "grad_norm": 0.37297449403274, + "learning_rate": 1.4894270176512431e-06, + "loss": 0.42, + "step": 4220 + }, + { + "epoch": 0.7733602052033712, + "grad_norm": 0.3991617672322749, + "learning_rate": 1.4871506440743594e-06, + "loss": 0.4385, + "step": 4221 + }, + { + "epoch": 0.7735434224990839, + "grad_norm": 0.3824646545144206, + "learning_rate": 1.4848757074504022e-06, + "loss": 0.4696, + "step": 4222 + }, + { + "epoch": 0.7737266397947966, + "grad_norm": 0.31946947825954536, + "learning_rate": 1.482602208709943e-06, + "loss": 0.4291, + "step": 4223 + }, + { + "epoch": 0.7739098570905093, + "grad_norm": 0.32384181752907243, + "learning_rate": 1.4803301487829724e-06, + "loss": 0.4732, + "step": 4224 + }, + { + "epoch": 0.7740930743862221, + "grad_norm": 0.3413473241442703, + "learning_rate": 1.4780595285988926e-06, + "loss": 0.45, + "step": 4225 + }, + { + "epoch": 0.7742762916819348, + "grad_norm": 0.3926313230738948, + "learning_rate": 1.4757903490865161e-06, + "loss": 0.4797, + "step": 4226 + }, + { + "epoch": 0.7744595089776475, + "grad_norm": 0.4057802372281518, + "learning_rate": 1.4735226111740603e-06, + "loss": 0.4713, + "step": 4227 + }, + { + "epoch": 0.7746427262733602, + "grad_norm": 0.3452989925591827, + "learning_rate": 1.471256315789159e-06, + "loss": 0.4456, + "step": 4228 + }, + { + "epoch": 0.774825943569073, + "grad_norm": 0.3587563978512209, + "learning_rate": 1.4689914638588576e-06, + "loss": 0.4718, + "step": 4229 + }, + { + "epoch": 0.7750091608647857, + "grad_norm": 0.36869921606926875, + "learning_rate": 1.466728056309606e-06, + "loss": 0.4826, + "step": 4230 + }, + { + "epoch": 0.7751923781604984, + "grad_norm": 0.46525249605615593, + "learning_rate": 1.4644660940672628e-06, + "loss": 0.4451, + "step": 4231 + }, + { + "epoch": 0.775375595456211, + "grad_norm": 0.3790786257091947, + "learning_rate": 1.4622055780570998e-06, + "loss": 0.4966, + "step": 4232 + }, + { + "epoch": 0.7755588127519237, + "grad_norm": 0.40297734518599043, + "learning_rate": 1.4599465092037972e-06, + "loss": 0.4553, + "step": 4233 + }, + { + "epoch": 0.7757420300476365, + "grad_norm": 0.32038981146950085, + "learning_rate": 1.457688888431439e-06, + "loss": 0.4586, + "step": 4234 + }, + { + "epoch": 0.7759252473433492, + "grad_norm": 0.3411044943658708, + "learning_rate": 1.4554327166635173e-06, + "loss": 0.4438, + "step": 4235 + }, + { + "epoch": 0.7761084646390619, + "grad_norm": 0.3810814204685612, + "learning_rate": 1.4531779948229352e-06, + "loss": 0.5268, + "step": 4236 + }, + { + "epoch": 0.7762916819347746, + "grad_norm": 0.3846917874458302, + "learning_rate": 1.450924723832004e-06, + "loss": 0.4913, + "step": 4237 + }, + { + "epoch": 0.7764748992304874, + "grad_norm": 0.34488136187377955, + "learning_rate": 1.448672904612432e-06, + "loss": 0.4635, + "step": 4238 + }, + { + "epoch": 0.7766581165262001, + "grad_norm": 0.34879248585022016, + "learning_rate": 1.4464225380853424e-06, + "loss": 0.4917, + "step": 4239 + }, + { + "epoch": 0.7768413338219128, + "grad_norm": 0.3512419617121299, + "learning_rate": 1.4441736251712618e-06, + "loss": 0.4873, + "step": 4240 + }, + { + "epoch": 0.7770245511176255, + "grad_norm": 0.3614433426020675, + "learning_rate": 1.4419261667901263e-06, + "loss": 0.4771, + "step": 4241 + }, + { + "epoch": 0.7772077684133383, + "grad_norm": 0.3484288404288865, + "learning_rate": 1.4396801638612651e-06, + "loss": 0.467, + "step": 4242 + }, + { + "epoch": 0.777390985709051, + "grad_norm": 0.36531882658799253, + "learning_rate": 1.4374356173034232e-06, + "loss": 0.4651, + "step": 4243 + }, + { + "epoch": 0.7775742030047637, + "grad_norm": 0.36660115110869784, + "learning_rate": 1.4351925280347483e-06, + "loss": 0.4382, + "step": 4244 + }, + { + "epoch": 0.7777574203004763, + "grad_norm": 0.373638355172436, + "learning_rate": 1.432950896972788e-06, + "loss": 0.4687, + "step": 4245 + }, + { + "epoch": 0.777940637596189, + "grad_norm": 0.3842026017085754, + "learning_rate": 1.4307107250344943e-06, + "loss": 0.4786, + "step": 4246 + }, + { + "epoch": 0.7781238548919018, + "grad_norm": 0.39885735283836576, + "learning_rate": 1.4284720131362251e-06, + "loss": 0.4565, + "step": 4247 + }, + { + "epoch": 0.7783070721876145, + "grad_norm": 0.32413558436040535, + "learning_rate": 1.4262347621937407e-06, + "loss": 0.4188, + "step": 4248 + }, + { + "epoch": 0.7784902894833272, + "grad_norm": 0.3199825820361419, + "learning_rate": 1.4239989731222008e-06, + "loss": 0.4753, + "step": 4249 + }, + { + "epoch": 0.7786735067790399, + "grad_norm": 0.3866048150203384, + "learning_rate": 1.4217646468361678e-06, + "loss": 0.4719, + "step": 4250 + }, + { + "epoch": 0.7788567240747527, + "grad_norm": 0.33398457863101333, + "learning_rate": 1.4195317842496081e-06, + "loss": 0.4501, + "step": 4251 + }, + { + "epoch": 0.7790399413704654, + "grad_norm": 0.37438075278743793, + "learning_rate": 1.41730038627589e-06, + "loss": 0.486, + "step": 4252 + }, + { + "epoch": 0.7792231586661781, + "grad_norm": 0.3491717754374234, + "learning_rate": 1.4150704538277787e-06, + "loss": 0.4569, + "step": 4253 + }, + { + "epoch": 0.7794063759618908, + "grad_norm": 0.2925615976398129, + "learning_rate": 1.4128419878174404e-06, + "loss": 0.4572, + "step": 4254 + }, + { + "epoch": 0.7795895932576035, + "grad_norm": 0.36550934487284253, + "learning_rate": 1.4106149891564446e-06, + "loss": 0.4496, + "step": 4255 + }, + { + "epoch": 0.7797728105533163, + "grad_norm": 0.33090577008304595, + "learning_rate": 1.4083894587557611e-06, + "loss": 0.478, + "step": 4256 + }, + { + "epoch": 0.779956027849029, + "grad_norm": 0.39536060323763084, + "learning_rate": 1.4061653975257543e-06, + "loss": 0.4555, + "step": 4257 + }, + { + "epoch": 0.7801392451447416, + "grad_norm": 0.33484107350738446, + "learning_rate": 1.40394280637619e-06, + "loss": 0.487, + "step": 4258 + }, + { + "epoch": 0.7803224624404543, + "grad_norm": 0.3963614209290271, + "learning_rate": 1.4017216862162358e-06, + "loss": 0.4752, + "step": 4259 + }, + { + "epoch": 0.7805056797361671, + "grad_norm": 0.3819513315128331, + "learning_rate": 1.3995020379544511e-06, + "loss": 0.4418, + "step": 4260 + }, + { + "epoch": 0.7806888970318798, + "grad_norm": 0.3818285149584103, + "learning_rate": 1.3972838624988e-06, + "loss": 0.4868, + "step": 4261 + }, + { + "epoch": 0.7808721143275925, + "grad_norm": 0.3597367747320405, + "learning_rate": 1.3950671607566379e-06, + "loss": 0.4256, + "step": 4262 + }, + { + "epoch": 0.7810553316233052, + "grad_norm": 0.3730154734054765, + "learning_rate": 1.3928519336347246e-06, + "loss": 0.436, + "step": 4263 + }, + { + "epoch": 0.781238548919018, + "grad_norm": 0.3646722112923301, + "learning_rate": 1.3906381820392078e-06, + "loss": 0.4764, + "step": 4264 + }, + { + "epoch": 0.7814217662147307, + "grad_norm": 0.32776218110740013, + "learning_rate": 1.3884259068756405e-06, + "loss": 0.4931, + "step": 4265 + }, + { + "epoch": 0.7816049835104434, + "grad_norm": 0.36847017159944145, + "learning_rate": 1.386215109048965e-06, + "loss": 0.4492, + "step": 4266 + }, + { + "epoch": 0.7817882008061561, + "grad_norm": 0.3715592107491827, + "learning_rate": 1.3840057894635239e-06, + "loss": 0.4985, + "step": 4267 + }, + { + "epoch": 0.7819714181018688, + "grad_norm": 0.36107516640793513, + "learning_rate": 1.3817979490230504e-06, + "loss": 0.4565, + "step": 4268 + }, + { + "epoch": 0.7821546353975816, + "grad_norm": 0.41991905805528, + "learning_rate": 1.3795915886306788e-06, + "loss": 0.4747, + "step": 4269 + }, + { + "epoch": 0.7823378526932943, + "grad_norm": 0.43970955495464925, + "learning_rate": 1.3773867091889314e-06, + "loss": 0.4588, + "step": 4270 + }, + { + "epoch": 0.782521069989007, + "grad_norm": 0.3788182062414767, + "learning_rate": 1.375183311599731e-06, + "loss": 0.4669, + "step": 4271 + }, + { + "epoch": 0.7827042872847196, + "grad_norm": 0.37028250258956347, + "learning_rate": 1.3729813967643874e-06, + "loss": 0.4784, + "step": 4272 + }, + { + "epoch": 0.7828875045804324, + "grad_norm": 0.3867900156936939, + "learning_rate": 1.3707809655836118e-06, + "loss": 0.4897, + "step": 4273 + }, + { + "epoch": 0.7830707218761451, + "grad_norm": 0.38057234870413315, + "learning_rate": 1.3685820189575016e-06, + "loss": 0.4581, + "step": 4274 + }, + { + "epoch": 0.7832539391718578, + "grad_norm": 0.39588496774229404, + "learning_rate": 1.3663845577855489e-06, + "loss": 0.485, + "step": 4275 + }, + { + "epoch": 0.7834371564675705, + "grad_norm": 0.399184532051107, + "learning_rate": 1.3641885829666402e-06, + "loss": 0.509, + "step": 4276 + }, + { + "epoch": 0.7836203737632833, + "grad_norm": 0.3306604176396543, + "learning_rate": 1.361994095399054e-06, + "loss": 0.4773, + "step": 4277 + }, + { + "epoch": 0.783803591058996, + "grad_norm": 0.4116985333231267, + "learning_rate": 1.3598010959804581e-06, + "loss": 0.4967, + "step": 4278 + }, + { + "epoch": 0.7839868083547087, + "grad_norm": 0.38816292483749887, + "learning_rate": 1.3576095856079103e-06, + "loss": 0.5005, + "step": 4279 + }, + { + "epoch": 0.7841700256504214, + "grad_norm": 0.37339570240461945, + "learning_rate": 1.3554195651778635e-06, + "loss": 0.4727, + "step": 4280 + }, + { + "epoch": 0.7843532429461341, + "grad_norm": 0.4139138005983736, + "learning_rate": 1.353231035586161e-06, + "loss": 0.4577, + "step": 4281 + }, + { + "epoch": 0.7845364602418469, + "grad_norm": 0.380168189362039, + "learning_rate": 1.3510439977280321e-06, + "loss": 0.4941, + "step": 4282 + }, + { + "epoch": 0.7847196775375596, + "grad_norm": 0.35535538182553233, + "learning_rate": 1.348858452498098e-06, + "loss": 0.4814, + "step": 4283 + }, + { + "epoch": 0.7849028948332722, + "grad_norm": 1.687367525349789, + "learning_rate": 1.3466744007903703e-06, + "loss": 0.4494, + "step": 4284 + }, + { + "epoch": 0.7850861121289849, + "grad_norm": 0.3616108945114174, + "learning_rate": 1.3444918434982508e-06, + "loss": 0.463, + "step": 4285 + }, + { + "epoch": 0.7852693294246977, + "grad_norm": 0.3273505313207282, + "learning_rate": 1.3423107815145258e-06, + "loss": 0.4695, + "step": 4286 + }, + { + "epoch": 0.7854525467204104, + "grad_norm": 0.37043645219640614, + "learning_rate": 1.3401312157313718e-06, + "loss": 0.4673, + "step": 4287 + }, + { + "epoch": 0.7856357640161231, + "grad_norm": 0.42054382225617815, + "learning_rate": 1.337953147040355e-06, + "loss": 0.4816, + "step": 4288 + }, + { + "epoch": 0.7858189813118358, + "grad_norm": 0.3725066069026344, + "learning_rate": 1.3357765763324292e-06, + "loss": 0.4588, + "step": 4289 + }, + { + "epoch": 0.7860021986075486, + "grad_norm": 0.44984578148010945, + "learning_rate": 1.3336015044979334e-06, + "loss": 0.4619, + "step": 4290 + }, + { + "epoch": 0.7861854159032613, + "grad_norm": 0.3251094395904671, + "learning_rate": 1.3314279324265922e-06, + "loss": 0.467, + "step": 4291 + }, + { + "epoch": 0.786368633198974, + "grad_norm": 0.3785278657991632, + "learning_rate": 1.3292558610075202e-06, + "loss": 0.4533, + "step": 4292 + }, + { + "epoch": 0.7865518504946867, + "grad_norm": 0.33247256342503656, + "learning_rate": 1.3270852911292205e-06, + "loss": 0.4425, + "step": 4293 + }, + { + "epoch": 0.7867350677903994, + "grad_norm": 0.3402317258710267, + "learning_rate": 1.324916223679571e-06, + "loss": 0.4593, + "step": 4294 + }, + { + "epoch": 0.7869182850861122, + "grad_norm": 0.41370625054005633, + "learning_rate": 1.322748659545846e-06, + "loss": 0.4747, + "step": 4295 + }, + { + "epoch": 0.7871015023818249, + "grad_norm": 0.32955713174828827, + "learning_rate": 1.3205825996147032e-06, + "loss": 0.4594, + "step": 4296 + }, + { + "epoch": 0.7872847196775375, + "grad_norm": 0.3691023172258719, + "learning_rate": 1.31841804477218e-06, + "loss": 0.4353, + "step": 4297 + }, + { + "epoch": 0.7874679369732502, + "grad_norm": 0.44060467171937595, + "learning_rate": 1.3162549959037008e-06, + "loss": 0.4525, + "step": 4298 + }, + { + "epoch": 0.787651154268963, + "grad_norm": 0.4016940695728928, + "learning_rate": 1.3140934538940754e-06, + "loss": 0.476, + "step": 4299 + }, + { + "epoch": 0.7878343715646757, + "grad_norm": 0.3665046025808825, + "learning_rate": 1.3119334196274973e-06, + "loss": 0.4811, + "step": 4300 + }, + { + "epoch": 0.7880175888603884, + "grad_norm": 0.8868487082909676, + "learning_rate": 1.3097748939875405e-06, + "loss": 0.4712, + "step": 4301 + }, + { + "epoch": 0.7882008061561011, + "grad_norm": 0.36745576578124317, + "learning_rate": 1.307617877857162e-06, + "loss": 0.4754, + "step": 4302 + }, + { + "epoch": 0.7883840234518138, + "grad_norm": 0.3730248217971587, + "learning_rate": 1.3054623721187044e-06, + "loss": 0.4307, + "step": 4303 + }, + { + "epoch": 0.7885672407475266, + "grad_norm": 0.42274610605900004, + "learning_rate": 1.3033083776538913e-06, + "loss": 0.4425, + "step": 4304 + }, + { + "epoch": 0.7887504580432393, + "grad_norm": 0.3655393687186923, + "learning_rate": 1.3011558953438275e-06, + "loss": 0.4522, + "step": 4305 + }, + { + "epoch": 0.788933675338952, + "grad_norm": 0.37414296465784624, + "learning_rate": 1.2990049260689968e-06, + "loss": 0.4443, + "step": 4306 + }, + { + "epoch": 0.7891168926346647, + "grad_norm": 0.3417181289402449, + "learning_rate": 1.2968554707092684e-06, + "loss": 0.454, + "step": 4307 + }, + { + "epoch": 0.7893001099303775, + "grad_norm": 0.5079584196429107, + "learning_rate": 1.2947075301438917e-06, + "loss": 0.4828, + "step": 4308 + }, + { + "epoch": 0.7894833272260902, + "grad_norm": 0.3318668850813015, + "learning_rate": 1.2925611052514936e-06, + "loss": 0.4544, + "step": 4309 + }, + { + "epoch": 0.7896665445218028, + "grad_norm": 0.4344267616965659, + "learning_rate": 1.2904161969100821e-06, + "loss": 0.4715, + "step": 4310 + }, + { + "epoch": 0.7898497618175155, + "grad_norm": 0.3687021074624977, + "learning_rate": 1.2882728059970467e-06, + "loss": 0.4606, + "step": 4311 + }, + { + "epoch": 0.7900329791132283, + "grad_norm": 0.4643355087722742, + "learning_rate": 1.2861309333891531e-06, + "loss": 0.4799, + "step": 4312 + }, + { + "epoch": 0.790216196408941, + "grad_norm": 0.37056139711117714, + "learning_rate": 1.2839905799625502e-06, + "loss": 0.4973, + "step": 4313 + }, + { + "epoch": 0.7903994137046537, + "grad_norm": 0.3223767446325161, + "learning_rate": 1.2818517465927598e-06, + "loss": 0.4344, + "step": 4314 + }, + { + "epoch": 0.7905826310003664, + "grad_norm": 0.4233916390777898, + "learning_rate": 1.2797144341546886e-06, + "loss": 0.462, + "step": 4315 + }, + { + "epoch": 0.7907658482960791, + "grad_norm": 0.36142948649438467, + "learning_rate": 1.2775786435226134e-06, + "loss": 0.42, + "step": 4316 + }, + { + "epoch": 0.7909490655917919, + "grad_norm": 0.36382775202555373, + "learning_rate": 1.2754443755701968e-06, + "loss": 0.4645, + "step": 4317 + }, + { + "epoch": 0.7911322828875046, + "grad_norm": 0.3608224174816189, + "learning_rate": 1.2733116311704708e-06, + "loss": 0.4939, + "step": 4318 + }, + { + "epoch": 0.7913155001832173, + "grad_norm": 0.35092615184678505, + "learning_rate": 1.271180411195851e-06, + "loss": 0.4863, + "step": 4319 + }, + { + "epoch": 0.79149871747893, + "grad_norm": 0.34005370769868437, + "learning_rate": 1.269050716518123e-06, + "loss": 0.4446, + "step": 4320 + }, + { + "epoch": 0.7916819347746428, + "grad_norm": 0.38812334405000143, + "learning_rate": 1.2669225480084552e-06, + "loss": 0.4712, + "step": 4321 + }, + { + "epoch": 0.7918651520703555, + "grad_norm": 0.43360901277203484, + "learning_rate": 1.264795906537385e-06, + "loss": 0.4871, + "step": 4322 + }, + { + "epoch": 0.7920483693660682, + "grad_norm": 0.4188394515602809, + "learning_rate": 1.262670792974831e-06, + "loss": 0.4442, + "step": 4323 + }, + { + "epoch": 0.7922315866617808, + "grad_norm": 0.34094786611303396, + "learning_rate": 1.260547208190081e-06, + "loss": 0.4715, + "step": 4324 + }, + { + "epoch": 0.7924148039574936, + "grad_norm": 0.34629672440714093, + "learning_rate": 1.2584251530518043e-06, + "loss": 0.4594, + "step": 4325 + }, + { + "epoch": 0.7925980212532063, + "grad_norm": 0.3521557556579978, + "learning_rate": 1.2563046284280377e-06, + "loss": 0.4993, + "step": 4326 + }, + { + "epoch": 0.792781238548919, + "grad_norm": 0.4782575659235454, + "learning_rate": 1.2541856351861976e-06, + "loss": 0.4679, + "step": 4327 + }, + { + "epoch": 0.7929644558446317, + "grad_norm": 0.3343947162744187, + "learning_rate": 1.2520681741930678e-06, + "loss": 0.4495, + "step": 4328 + }, + { + "epoch": 0.7931476731403444, + "grad_norm": 0.40705124339909976, + "learning_rate": 1.2499522463148134e-06, + "loss": 0.4408, + "step": 4329 + }, + { + "epoch": 0.7933308904360572, + "grad_norm": 0.3728257661218358, + "learning_rate": 1.2478378524169653e-06, + "loss": 0.4461, + "step": 4330 + }, + { + "epoch": 0.7935141077317699, + "grad_norm": 0.38649163744573345, + "learning_rate": 1.2457249933644289e-06, + "loss": 0.4601, + "step": 4331 + }, + { + "epoch": 0.7936973250274826, + "grad_norm": 0.3646925001518044, + "learning_rate": 1.243613670021483e-06, + "loss": 0.461, + "step": 4332 + }, + { + "epoch": 0.7938805423231953, + "grad_norm": 0.3830273257319351, + "learning_rate": 1.241503883251779e-06, + "loss": 0.4755, + "step": 4333 + }, + { + "epoch": 0.7940637596189081, + "grad_norm": 0.39895799208872956, + "learning_rate": 1.2393956339183378e-06, + "loss": 0.4621, + "step": 4334 + }, + { + "epoch": 0.7942469769146208, + "grad_norm": 0.36110765523086696, + "learning_rate": 1.2372889228835494e-06, + "loss": 0.4538, + "step": 4335 + }, + { + "epoch": 0.7944301942103335, + "grad_norm": 0.3710208588554927, + "learning_rate": 1.2351837510091792e-06, + "loss": 0.4653, + "step": 4336 + }, + { + "epoch": 0.7946134115060461, + "grad_norm": 0.3544845317172887, + "learning_rate": 1.2330801191563617e-06, + "loss": 0.4794, + "step": 4337 + }, + { + "epoch": 0.7947966288017588, + "grad_norm": 0.510757540441394, + "learning_rate": 1.2309780281856005e-06, + "loss": 0.4627, + "step": 4338 + }, + { + "epoch": 0.7949798460974716, + "grad_norm": 0.430705271678027, + "learning_rate": 1.2288774789567659e-06, + "loss": 0.4503, + "step": 4339 + }, + { + "epoch": 0.7951630633931843, + "grad_norm": 0.3551683574176806, + "learning_rate": 1.226778472329103e-06, + "loss": 0.4605, + "step": 4340 + }, + { + "epoch": 0.795346280688897, + "grad_norm": 0.33622015949413203, + "learning_rate": 1.2246810091612255e-06, + "loss": 0.4531, + "step": 4341 + }, + { + "epoch": 0.7955294979846097, + "grad_norm": 0.3230378019256525, + "learning_rate": 1.2225850903111115e-06, + "loss": 0.4492, + "step": 4342 + }, + { + "epoch": 0.7957127152803225, + "grad_norm": 0.3902465336192029, + "learning_rate": 1.220490716636108e-06, + "loss": 0.4669, + "step": 4343 + }, + { + "epoch": 0.7958959325760352, + "grad_norm": 0.401011733885209, + "learning_rate": 1.2183978889929338e-06, + "loss": 0.4729, + "step": 4344 + }, + { + "epoch": 0.7960791498717479, + "grad_norm": 0.3633379973811808, + "learning_rate": 1.216306608237675e-06, + "loss": 0.4775, + "step": 4345 + }, + { + "epoch": 0.7962623671674606, + "grad_norm": 0.34420397215542725, + "learning_rate": 1.2142168752257788e-06, + "loss": 0.4539, + "step": 4346 + }, + { + "epoch": 0.7964455844631734, + "grad_norm": 0.3637954212352133, + "learning_rate": 1.212128690812065e-06, + "loss": 0.4478, + "step": 4347 + }, + { + "epoch": 0.7966288017588861, + "grad_norm": 0.36818404312358316, + "learning_rate": 1.210042055850721e-06, + "loss": 0.4554, + "step": 4348 + }, + { + "epoch": 0.7968120190545988, + "grad_norm": 0.3693162755249909, + "learning_rate": 1.2079569711952949e-06, + "loss": 0.4309, + "step": 4349 + }, + { + "epoch": 0.7969952363503114, + "grad_norm": 0.36625951590115813, + "learning_rate": 1.2058734376987036e-06, + "loss": 0.4371, + "step": 4350 + }, + { + "epoch": 0.7971784536460241, + "grad_norm": 0.3712938003355549, + "learning_rate": 1.2037914562132303e-06, + "loss": 0.4913, + "step": 4351 + }, + { + "epoch": 0.7973616709417369, + "grad_norm": 0.3596821413630336, + "learning_rate": 1.2017110275905242e-06, + "loss": 0.4726, + "step": 4352 + }, + { + "epoch": 0.7975448882374496, + "grad_norm": 0.4450409267238231, + "learning_rate": 1.1996321526815952e-06, + "loss": 0.4767, + "step": 4353 + }, + { + "epoch": 0.7977281055331623, + "grad_norm": 0.34879634269104026, + "learning_rate": 1.1975548323368201e-06, + "loss": 0.4872, + "step": 4354 + }, + { + "epoch": 0.797911322828875, + "grad_norm": 0.3711969884064517, + "learning_rate": 1.1954790674059401e-06, + "loss": 0.4876, + "step": 4355 + }, + { + "epoch": 0.7980945401245878, + "grad_norm": 0.3814608762106385, + "learning_rate": 1.1934048587380614e-06, + "loss": 0.4751, + "step": 4356 + }, + { + "epoch": 0.7982777574203005, + "grad_norm": 0.36503469451782006, + "learning_rate": 1.1913322071816508e-06, + "loss": 0.4403, + "step": 4357 + }, + { + "epoch": 0.7984609747160132, + "grad_norm": 0.40272608111184344, + "learning_rate": 1.1892611135845377e-06, + "loss": 0.4938, + "step": 4358 + }, + { + "epoch": 0.7986441920117259, + "grad_norm": 0.4756493978315566, + "learning_rate": 1.1871915787939176e-06, + "loss": 0.4554, + "step": 4359 + }, + { + "epoch": 0.7988274093074387, + "grad_norm": 0.4464658741669984, + "learning_rate": 1.1851236036563475e-06, + "loss": 0.483, + "step": 4360 + }, + { + "epoch": 0.7990106266031514, + "grad_norm": 0.38328718284689556, + "learning_rate": 1.1830571890177445e-06, + "loss": 0.5021, + "step": 4361 + }, + { + "epoch": 0.799193843898864, + "grad_norm": 0.37065782957864374, + "learning_rate": 1.1809923357233865e-06, + "loss": 0.4665, + "step": 4362 + }, + { + "epoch": 0.7993770611945767, + "grad_norm": 0.3841809804800641, + "learning_rate": 1.1789290446179168e-06, + "loss": 0.4807, + "step": 4363 + }, + { + "epoch": 0.7995602784902894, + "grad_norm": 0.3806962844430292, + "learning_rate": 1.1768673165453388e-06, + "loss": 0.4773, + "step": 4364 + }, + { + "epoch": 0.7997434957860022, + "grad_norm": 0.4281880314321846, + "learning_rate": 1.1748071523490128e-06, + "loss": 0.4803, + "step": 4365 + }, + { + "epoch": 0.7999267130817149, + "grad_norm": 0.34405712330235033, + "learning_rate": 1.1727485528716614e-06, + "loss": 0.481, + "step": 4366 + }, + { + "epoch": 0.8001099303774276, + "grad_norm": 0.3941810159437789, + "learning_rate": 1.1706915189553697e-06, + "loss": 0.4362, + "step": 4367 + }, + { + "epoch": 0.8002931476731403, + "grad_norm": 0.4665845163260568, + "learning_rate": 1.1686360514415785e-06, + "loss": 0.4885, + "step": 4368 + }, + { + "epoch": 0.8004763649688531, + "grad_norm": 0.32379661690882694, + "learning_rate": 1.1665821511710906e-06, + "loss": 0.4326, + "step": 4369 + }, + { + "epoch": 0.8006595822645658, + "grad_norm": 0.3880909485352397, + "learning_rate": 1.164529818984066e-06, + "loss": 0.4491, + "step": 4370 + }, + { + "epoch": 0.8008427995602785, + "grad_norm": 0.40272810037657775, + "learning_rate": 1.1624790557200255e-06, + "loss": 0.5119, + "step": 4371 + }, + { + "epoch": 0.8010260168559912, + "grad_norm": 0.3793125409371189, + "learning_rate": 1.1604298622178434e-06, + "loss": 0.4753, + "step": 4372 + }, + { + "epoch": 0.801209234151704, + "grad_norm": 0.4062136455850809, + "learning_rate": 1.158382239315759e-06, + "loss": 0.4569, + "step": 4373 + }, + { + "epoch": 0.8013924514474167, + "grad_norm": 0.3795142877597605, + "learning_rate": 1.1563361878513623e-06, + "loss": 0.4316, + "step": 4374 + }, + { + "epoch": 0.8015756687431294, + "grad_norm": 0.40072224290143865, + "learning_rate": 1.154291708661605e-06, + "loss": 0.4813, + "step": 4375 + }, + { + "epoch": 0.801758886038842, + "grad_norm": 0.366659585611398, + "learning_rate": 1.1522488025827927e-06, + "loss": 0.4479, + "step": 4376 + }, + { + "epoch": 0.8019421033345547, + "grad_norm": 0.3367499766091904, + "learning_rate": 1.150207470450591e-06, + "loss": 0.4414, + "step": 4377 + }, + { + "epoch": 0.8021253206302675, + "grad_norm": 0.666426226371213, + "learning_rate": 1.1481677131000168e-06, + "loss": 0.4763, + "step": 4378 + }, + { + "epoch": 0.8023085379259802, + "grad_norm": 0.3820411181072656, + "learning_rate": 1.1461295313654486e-06, + "loss": 0.4435, + "step": 4379 + }, + { + "epoch": 0.8024917552216929, + "grad_norm": 0.39359593527223125, + "learning_rate": 1.1440929260806138e-06, + "loss": 0.4692, + "step": 4380 + }, + { + "epoch": 0.8026749725174056, + "grad_norm": 0.36105377660188126, + "learning_rate": 1.1420578980786017e-06, + "loss": 0.4911, + "step": 4381 + }, + { + "epoch": 0.8028581898131184, + "grad_norm": 0.39957962024831023, + "learning_rate": 1.1400244481918522e-06, + "loss": 0.4635, + "step": 4382 + }, + { + "epoch": 0.8030414071088311, + "grad_norm": 0.35380277463987786, + "learning_rate": 1.1379925772521588e-06, + "loss": 0.4914, + "step": 4383 + }, + { + "epoch": 0.8032246244045438, + "grad_norm": 0.5062571261684051, + "learning_rate": 1.1359622860906728e-06, + "loss": 0.4401, + "step": 4384 + }, + { + "epoch": 0.8034078417002565, + "grad_norm": 0.42014717561460646, + "learning_rate": 1.1339335755378977e-06, + "loss": 0.4812, + "step": 4385 + }, + { + "epoch": 0.8035910589959692, + "grad_norm": 0.3300525420829916, + "learning_rate": 1.1319064464236905e-06, + "loss": 0.45, + "step": 4386 + }, + { + "epoch": 0.803774276291682, + "grad_norm": 0.3662942909996141, + "learning_rate": 1.129880899577258e-06, + "loss": 0.4606, + "step": 4387 + }, + { + "epoch": 0.8039574935873947, + "grad_norm": 0.35610961461431373, + "learning_rate": 1.1278569358271658e-06, + "loss": 0.4412, + "step": 4388 + }, + { + "epoch": 0.8041407108831073, + "grad_norm": 0.3493862728552464, + "learning_rate": 1.1258345560013284e-06, + "loss": 0.466, + "step": 4389 + }, + { + "epoch": 0.80432392817882, + "grad_norm": 0.36281108434312453, + "learning_rate": 1.1238137609270133e-06, + "loss": 0.472, + "step": 4390 + }, + { + "epoch": 0.8045071454745328, + "grad_norm": 0.3449032370268849, + "learning_rate": 1.1217945514308366e-06, + "loss": 0.436, + "step": 4391 + }, + { + "epoch": 0.8046903627702455, + "grad_norm": 0.5043542193677666, + "learning_rate": 1.1197769283387704e-06, + "loss": 0.4626, + "step": 4392 + }, + { + "epoch": 0.8048735800659582, + "grad_norm": 0.3217464944762706, + "learning_rate": 1.1177608924761375e-06, + "loss": 0.4531, + "step": 4393 + }, + { + "epoch": 0.8050567973616709, + "grad_norm": 0.3811625731036834, + "learning_rate": 1.1157464446676086e-06, + "loss": 0.4852, + "step": 4394 + }, + { + "epoch": 0.8052400146573837, + "grad_norm": 0.4047205089953529, + "learning_rate": 1.1137335857372045e-06, + "loss": 0.4609, + "step": 4395 + }, + { + "epoch": 0.8054232319530964, + "grad_norm": 0.42685010940805795, + "learning_rate": 1.1117223165082985e-06, + "loss": 0.4462, + "step": 4396 + }, + { + "epoch": 0.8056064492488091, + "grad_norm": 0.33879794317670375, + "learning_rate": 1.1097126378036171e-06, + "loss": 0.4355, + "step": 4397 + }, + { + "epoch": 0.8057896665445218, + "grad_norm": 0.34659367032239924, + "learning_rate": 1.1077045504452249e-06, + "loss": 0.4718, + "step": 4398 + }, + { + "epoch": 0.8059728838402345, + "grad_norm": 0.37306376584072337, + "learning_rate": 1.1056980552545466e-06, + "loss": 0.453, + "step": 4399 + }, + { + "epoch": 0.8061561011359473, + "grad_norm": 0.391379392652362, + "learning_rate": 1.1036931530523504e-06, + "loss": 0.5196, + "step": 4400 + }, + { + "epoch": 0.80633931843166, + "grad_norm": 0.3156649695122798, + "learning_rate": 1.1016898446587576e-06, + "loss": 0.4317, + "step": 4401 + }, + { + "epoch": 0.8065225357273726, + "grad_norm": 0.3452125669221769, + "learning_rate": 1.0996881308932278e-06, + "loss": 0.4837, + "step": 4402 + }, + { + "epoch": 0.8067057530230853, + "grad_norm": 0.41606500874857577, + "learning_rate": 1.097688012574578e-06, + "loss": 0.4683, + "step": 4403 + }, + { + "epoch": 0.8068889703187981, + "grad_norm": 0.3910554743052055, + "learning_rate": 1.0956894905209697e-06, + "loss": 0.4776, + "step": 4404 + }, + { + "epoch": 0.8070721876145108, + "grad_norm": 0.3463825987056052, + "learning_rate": 1.0936925655499097e-06, + "loss": 0.4431, + "step": 4405 + }, + { + "epoch": 0.8072554049102235, + "grad_norm": 0.48967145753575547, + "learning_rate": 1.091697238478252e-06, + "loss": 0.4691, + "step": 4406 + }, + { + "epoch": 0.8074386222059362, + "grad_norm": 0.35388487088191084, + "learning_rate": 1.0897035101221976e-06, + "loss": 0.4703, + "step": 4407 + }, + { + "epoch": 0.807621839501649, + "grad_norm": 0.38250770882105284, + "learning_rate": 1.0877113812972955e-06, + "loss": 0.4773, + "step": 4408 + }, + { + "epoch": 0.8078050567973617, + "grad_norm": 0.3728025054746775, + "learning_rate": 1.0857208528184376e-06, + "loss": 0.488, + "step": 4409 + }, + { + "epoch": 0.8079882740930744, + "grad_norm": 0.375253461538163, + "learning_rate": 1.0837319254998595e-06, + "loss": 0.4845, + "step": 4410 + }, + { + "epoch": 0.8081714913887871, + "grad_norm": 0.3905933282011512, + "learning_rate": 1.0817446001551467e-06, + "loss": 0.4657, + "step": 4411 + }, + { + "epoch": 0.8083547086844998, + "grad_norm": 0.36763332160464973, + "learning_rate": 1.079758877597228e-06, + "loss": 0.4852, + "step": 4412 + }, + { + "epoch": 0.8085379259802126, + "grad_norm": 0.351866760115047, + "learning_rate": 1.0777747586383736e-06, + "loss": 0.4795, + "step": 4413 + }, + { + "epoch": 0.8087211432759253, + "grad_norm": 0.5675280442034399, + "learning_rate": 1.0757922440901997e-06, + "loss": 0.4791, + "step": 4414 + }, + { + "epoch": 0.808904360571638, + "grad_norm": 0.3754197011748734, + "learning_rate": 1.0738113347636663e-06, + "loss": 0.4852, + "step": 4415 + }, + { + "epoch": 0.8090875778673506, + "grad_norm": 0.30723623558454144, + "learning_rate": 1.071832031469079e-06, + "loss": 0.461, + "step": 4416 + }, + { + "epoch": 0.8092707951630634, + "grad_norm": 0.3488234134989342, + "learning_rate": 1.0698543350160817e-06, + "loss": 0.4665, + "step": 4417 + }, + { + "epoch": 0.8094540124587761, + "grad_norm": 0.3401454346374473, + "learning_rate": 1.0678782462136633e-06, + "loss": 0.4104, + "step": 4418 + }, + { + "epoch": 0.8096372297544888, + "grad_norm": 0.3866140445566972, + "learning_rate": 1.0659037658701576e-06, + "loss": 0.4837, + "step": 4419 + }, + { + "epoch": 0.8098204470502015, + "grad_norm": 0.3322659761645293, + "learning_rate": 1.0639308947932353e-06, + "loss": 0.4314, + "step": 4420 + }, + { + "epoch": 0.8100036643459143, + "grad_norm": 0.3558313119140645, + "learning_rate": 1.061959633789914e-06, + "loss": 0.461, + "step": 4421 + }, + { + "epoch": 0.810186881641627, + "grad_norm": 0.3330464951183908, + "learning_rate": 1.0599899836665478e-06, + "loss": 0.4575, + "step": 4422 + }, + { + "epoch": 0.8103700989373397, + "grad_norm": 0.42928773831421463, + "learning_rate": 1.0580219452288371e-06, + "loss": 0.4769, + "step": 4423 + }, + { + "epoch": 0.8105533162330524, + "grad_norm": 0.32977757779393846, + "learning_rate": 1.0560555192818179e-06, + "loss": 0.4988, + "step": 4424 + }, + { + "epoch": 0.8107365335287651, + "grad_norm": 0.3582487689445677, + "learning_rate": 1.0540907066298706e-06, + "loss": 0.4771, + "step": 4425 + }, + { + "epoch": 0.8109197508244779, + "grad_norm": 0.40665178748695113, + "learning_rate": 1.052127508076713e-06, + "loss": 0.4136, + "step": 4426 + }, + { + "epoch": 0.8111029681201906, + "grad_norm": 0.3885616522577978, + "learning_rate": 1.0501659244254053e-06, + "loss": 0.4503, + "step": 4427 + }, + { + "epoch": 0.8112861854159032, + "grad_norm": 0.3743728300256291, + "learning_rate": 1.0482059564783432e-06, + "loss": 0.4797, + "step": 4428 + }, + { + "epoch": 0.8114694027116159, + "grad_norm": 0.39798624771474667, + "learning_rate": 1.046247605037266e-06, + "loss": 0.4783, + "step": 4429 + }, + { + "epoch": 0.8116526200073287, + "grad_norm": 0.4399111193212314, + "learning_rate": 1.0442908709032474e-06, + "loss": 0.4621, + "step": 4430 + }, + { + "epoch": 0.8118358373030414, + "grad_norm": 0.36391524408557696, + "learning_rate": 1.0423357548767044e-06, + "loss": 0.4689, + "step": 4431 + }, + { + "epoch": 0.8120190545987541, + "grad_norm": 0.3521884153330929, + "learning_rate": 1.0403822577573863e-06, + "loss": 0.4381, + "step": 4432 + }, + { + "epoch": 0.8122022718944668, + "grad_norm": 0.36384756956744413, + "learning_rate": 1.0384303803443862e-06, + "loss": 0.4742, + "step": 4433 + }, + { + "epoch": 0.8123854891901795, + "grad_norm": 0.36914020353091515, + "learning_rate": 1.0364801234361304e-06, + "loss": 0.5122, + "step": 4434 + }, + { + "epoch": 0.8125687064858923, + "grad_norm": 0.36959898699837357, + "learning_rate": 1.0345314878303826e-06, + "loss": 0.4588, + "step": 4435 + }, + { + "epoch": 0.812751923781605, + "grad_norm": 0.3582657175686495, + "learning_rate": 1.0325844743242447e-06, + "loss": 0.4863, + "step": 4436 + }, + { + "epoch": 0.8129351410773177, + "grad_norm": 0.3302257889303132, + "learning_rate": 1.0306390837141571e-06, + "loss": 0.4773, + "step": 4437 + }, + { + "epoch": 0.8131183583730304, + "grad_norm": 0.334571725214738, + "learning_rate": 1.0286953167958924e-06, + "loss": 0.4558, + "step": 4438 + }, + { + "epoch": 0.8133015756687432, + "grad_norm": 0.3469409174438552, + "learning_rate": 1.0267531743645588e-06, + "loss": 0.4703, + "step": 4439 + }, + { + "epoch": 0.8134847929644559, + "grad_norm": 0.32931549191190873, + "learning_rate": 1.0248126572146034e-06, + "loss": 0.4472, + "step": 4440 + }, + { + "epoch": 0.8136680102601686, + "grad_norm": 0.4480920133629509, + "learning_rate": 1.022873766139808e-06, + "loss": 0.4988, + "step": 4441 + }, + { + "epoch": 0.8138512275558812, + "grad_norm": 0.3817080914390894, + "learning_rate": 1.0209365019332874e-06, + "loss": 0.4824, + "step": 4442 + }, + { + "epoch": 0.814034444851594, + "grad_norm": 0.37044396015598, + "learning_rate": 1.019000865387489e-06, + "loss": 0.4895, + "step": 4443 + }, + { + "epoch": 0.8142176621473067, + "grad_norm": 0.4093703441079783, + "learning_rate": 1.0170668572941994e-06, + "loss": 0.4423, + "step": 4444 + }, + { + "epoch": 0.8144008794430194, + "grad_norm": 0.3848055668142181, + "learning_rate": 1.0151344784445377e-06, + "loss": 0.4673, + "step": 4445 + }, + { + "epoch": 0.8145840967387321, + "grad_norm": 0.3849793780680973, + "learning_rate": 1.0132037296289543e-06, + "loss": 0.4373, + "step": 4446 + }, + { + "epoch": 0.8147673140344448, + "grad_norm": 0.45161902879935206, + "learning_rate": 1.0112746116372325e-06, + "loss": 0.4859, + "step": 4447 + }, + { + "epoch": 0.8149505313301576, + "grad_norm": 0.417825453068041, + "learning_rate": 1.0093471252584913e-06, + "loss": 0.4605, + "step": 4448 + }, + { + "epoch": 0.8151337486258703, + "grad_norm": 0.3711653049927354, + "learning_rate": 1.0074212712811837e-06, + "loss": 0.4823, + "step": 4449 + }, + { + "epoch": 0.815316965921583, + "grad_norm": 0.38442640351733903, + "learning_rate": 1.0054970504930871e-06, + "loss": 0.4853, + "step": 4450 + }, + { + "epoch": 0.8155001832172957, + "grad_norm": 0.3751117008126234, + "learning_rate": 1.0035744636813188e-06, + "loss": 0.4409, + "step": 4451 + }, + { + "epoch": 0.8156834005130085, + "grad_norm": 0.3518804625006102, + "learning_rate": 1.0016535116323244e-06, + "loss": 0.4352, + "step": 4452 + }, + { + "epoch": 0.8158666178087212, + "grad_norm": 0.35352305826439295, + "learning_rate": 9.997341951318845e-07, + "loss": 0.4716, + "step": 4453 + }, + { + "epoch": 0.8160498351044339, + "grad_norm": 0.33286345392841504, + "learning_rate": 9.97816514965102e-07, + "loss": 0.4819, + "step": 4454 + }, + { + "epoch": 0.8162330524001465, + "grad_norm": 0.3351324583095673, + "learning_rate": 9.959004719164183e-07, + "loss": 0.4325, + "step": 4455 + }, + { + "epoch": 0.8164162696958593, + "grad_norm": 0.36938213527741026, + "learning_rate": 9.939860667696044e-07, + "loss": 0.4501, + "step": 4456 + }, + { + "epoch": 0.816599486991572, + "grad_norm": 0.35056915856927895, + "learning_rate": 9.920733003077581e-07, + "loss": 0.4652, + "step": 4457 + }, + { + "epoch": 0.8167827042872847, + "grad_norm": 0.3323779365144009, + "learning_rate": 9.901621733133077e-07, + "loss": 0.4811, + "step": 4458 + }, + { + "epoch": 0.8169659215829974, + "grad_norm": 0.3646028401284863, + "learning_rate": 9.882526865680125e-07, + "loss": 0.4938, + "step": 4459 + }, + { + "epoch": 0.8171491388787101, + "grad_norm": 0.4385436430907302, + "learning_rate": 9.863448408529607e-07, + "loss": 0.4743, + "step": 4460 + }, + { + "epoch": 0.8173323561744229, + "grad_norm": 0.41099955048823666, + "learning_rate": 9.84438636948568e-07, + "loss": 0.4752, + "step": 4461 + }, + { + "epoch": 0.8175155734701356, + "grad_norm": 0.43989353460252445, + "learning_rate": 9.82534075634578e-07, + "loss": 0.4928, + "step": 4462 + }, + { + "epoch": 0.8176987907658483, + "grad_norm": 0.3118585525166271, + "learning_rate": 9.806311576900634e-07, + "loss": 0.4488, + "step": 4463 + }, + { + "epoch": 0.817882008061561, + "grad_norm": 0.3648061199592301, + "learning_rate": 9.787298838934268e-07, + "loss": 0.4478, + "step": 4464 + }, + { + "epoch": 0.8180652253572738, + "grad_norm": 0.35776395194326077, + "learning_rate": 9.768302550223946e-07, + "loss": 0.4529, + "step": 4465 + }, + { + "epoch": 0.8182484426529865, + "grad_norm": 0.38026279339849606, + "learning_rate": 9.74932271854021e-07, + "loss": 0.496, + "step": 4466 + }, + { + "epoch": 0.8184316599486992, + "grad_norm": 0.4189226914369542, + "learning_rate": 9.730359351646885e-07, + "loss": 0.4488, + "step": 4467 + }, + { + "epoch": 0.8186148772444118, + "grad_norm": 0.36426634253454604, + "learning_rate": 9.711412457301067e-07, + "loss": 0.473, + "step": 4468 + }, + { + "epoch": 0.8187980945401245, + "grad_norm": 0.3545936825028976, + "learning_rate": 9.692482043253093e-07, + "loss": 0.4704, + "step": 4469 + }, + { + "epoch": 0.8189813118358373, + "grad_norm": 0.37455159789300885, + "learning_rate": 9.673568117246547e-07, + "loss": 0.4326, + "step": 4470 + }, + { + "epoch": 0.81916452913155, + "grad_norm": 0.39841516260954707, + "learning_rate": 9.654670687018325e-07, + "loss": 0.4638, + "step": 4471 + }, + { + "epoch": 0.8193477464272627, + "grad_norm": 0.3735166775800908, + "learning_rate": 9.635789760298502e-07, + "loss": 0.4438, + "step": 4472 + }, + { + "epoch": 0.8195309637229754, + "grad_norm": 0.48541384381778846, + "learning_rate": 9.616925344810474e-07, + "loss": 0.4749, + "step": 4473 + }, + { + "epoch": 0.8197141810186882, + "grad_norm": 0.3317020963961296, + "learning_rate": 9.598077448270815e-07, + "loss": 0.4487, + "step": 4474 + }, + { + "epoch": 0.8198973983144009, + "grad_norm": 0.35687935050035474, + "learning_rate": 9.579246078389404e-07, + "loss": 0.4818, + "step": 4475 + }, + { + "epoch": 0.8200806156101136, + "grad_norm": 0.3715318685578431, + "learning_rate": 9.5604312428693e-07, + "loss": 0.4616, + "step": 4476 + }, + { + "epoch": 0.8202638329058263, + "grad_norm": 0.43281036847700827, + "learning_rate": 9.54163294940686e-07, + "loss": 0.4631, + "step": 4477 + }, + { + "epoch": 0.8204470502015391, + "grad_norm": 0.43091325828842253, + "learning_rate": 9.522851205691613e-07, + "loss": 0.4656, + "step": 4478 + }, + { + "epoch": 0.8206302674972518, + "grad_norm": 0.3432212999111417, + "learning_rate": 9.504086019406372e-07, + "loss": 0.4569, + "step": 4479 + }, + { + "epoch": 0.8208134847929645, + "grad_norm": 0.35137409679079856, + "learning_rate": 9.48533739822714e-07, + "loss": 0.413, + "step": 4480 + }, + { + "epoch": 0.8209967020886771, + "grad_norm": 0.3571516582130752, + "learning_rate": 9.466605349823171e-07, + "loss": 0.4852, + "step": 4481 + }, + { + "epoch": 0.8211799193843898, + "grad_norm": 0.4093744350311809, + "learning_rate": 9.447889881856903e-07, + "loss": 0.4609, + "step": 4482 + }, + { + "epoch": 0.8213631366801026, + "grad_norm": 0.43496638081906147, + "learning_rate": 9.42919100198404e-07, + "loss": 0.4511, + "step": 4483 + }, + { + "epoch": 0.8215463539758153, + "grad_norm": 0.3635962455991155, + "learning_rate": 9.410508717853456e-07, + "loss": 0.4513, + "step": 4484 + }, + { + "epoch": 0.821729571271528, + "grad_norm": 0.3775517329332157, + "learning_rate": 9.391843037107273e-07, + "loss": 0.4879, + "step": 4485 + }, + { + "epoch": 0.8219127885672407, + "grad_norm": 0.3730331453968138, + "learning_rate": 9.373193967380795e-07, + "loss": 0.4761, + "step": 4486 + }, + { + "epoch": 0.8220960058629535, + "grad_norm": 0.41143431557152677, + "learning_rate": 9.354561516302529e-07, + "loss": 0.4894, + "step": 4487 + }, + { + "epoch": 0.8222792231586662, + "grad_norm": 0.3764448817274266, + "learning_rate": 9.335945691494208e-07, + "loss": 0.4609, + "step": 4488 + }, + { + "epoch": 0.8224624404543789, + "grad_norm": 0.4240698309497419, + "learning_rate": 9.317346500570756e-07, + "loss": 0.453, + "step": 4489 + }, + { + "epoch": 0.8226456577500916, + "grad_norm": 0.3625116175891358, + "learning_rate": 9.298763951140289e-07, + "loss": 0.4856, + "step": 4490 + }, + { + "epoch": 0.8228288750458044, + "grad_norm": 0.3821827959286687, + "learning_rate": 9.28019805080409e-07, + "loss": 0.4754, + "step": 4491 + }, + { + "epoch": 0.8230120923415171, + "grad_norm": 0.3591236541472167, + "learning_rate": 9.261648807156676e-07, + "loss": 0.4822, + "step": 4492 + }, + { + "epoch": 0.8231953096372298, + "grad_norm": 0.49996044918520816, + "learning_rate": 9.243116227785736e-07, + "loss": 0.4548, + "step": 4493 + }, + { + "epoch": 0.8233785269329424, + "grad_norm": 0.33266958827477816, + "learning_rate": 9.224600320272137e-07, + "loss": 0.4442, + "step": 4494 + }, + { + "epoch": 0.8235617442286551, + "grad_norm": 0.3953298488403479, + "learning_rate": 9.206101092189901e-07, + "loss": 0.4776, + "step": 4495 + }, + { + "epoch": 0.8237449615243679, + "grad_norm": 0.38455646293128437, + "learning_rate": 9.18761855110627e-07, + "loss": 0.4832, + "step": 4496 + }, + { + "epoch": 0.8239281788200806, + "grad_norm": 0.3886959733118104, + "learning_rate": 9.169152704581657e-07, + "loss": 0.4787, + "step": 4497 + }, + { + "epoch": 0.8241113961157933, + "grad_norm": 0.34746571008482313, + "learning_rate": 9.150703560169616e-07, + "loss": 0.457, + "step": 4498 + }, + { + "epoch": 0.824294613411506, + "grad_norm": 0.3696277530932602, + "learning_rate": 9.132271125416875e-07, + "loss": 0.4945, + "step": 4499 + }, + { + "epoch": 0.8244778307072188, + "grad_norm": 0.40609660061425584, + "learning_rate": 9.113855407863343e-07, + "loss": 0.4925, + "step": 4500 + }, + { + "epoch": 0.8246610480029315, + "grad_norm": 0.36927918768875456, + "learning_rate": 9.095456415042103e-07, + "loss": 0.4543, + "step": 4501 + }, + { + "epoch": 0.8248442652986442, + "grad_norm": 0.3198398572576675, + "learning_rate": 9.077074154479366e-07, + "loss": 0.4603, + "step": 4502 + }, + { + "epoch": 0.8250274825943569, + "grad_norm": 0.4192273429523626, + "learning_rate": 9.05870863369449e-07, + "loss": 0.4554, + "step": 4503 + }, + { + "epoch": 0.8252106998900697, + "grad_norm": 0.3732913276398482, + "learning_rate": 9.040359860200026e-07, + "loss": 0.491, + "step": 4504 + }, + { + "epoch": 0.8253939171857824, + "grad_norm": 0.33394412586942546, + "learning_rate": 9.022027841501679e-07, + "loss": 0.4603, + "step": 4505 + }, + { + "epoch": 0.8255771344814951, + "grad_norm": 0.31145562770619056, + "learning_rate": 9.003712585098217e-07, + "loss": 0.4424, + "step": 4506 + }, + { + "epoch": 0.8257603517772077, + "grad_norm": 0.3616345159244858, + "learning_rate": 8.985414098481643e-07, + "loss": 0.4727, + "step": 4507 + }, + { + "epoch": 0.8259435690729204, + "grad_norm": 0.39150282387562557, + "learning_rate": 8.967132389137079e-07, + "loss": 0.5034, + "step": 4508 + }, + { + "epoch": 0.8261267863686332, + "grad_norm": 0.39868311148958474, + "learning_rate": 8.948867464542754e-07, + "loss": 0.4795, + "step": 4509 + }, + { + "epoch": 0.8263100036643459, + "grad_norm": 0.3562244393654728, + "learning_rate": 8.930619332170038e-07, + "loss": 0.4456, + "step": 4510 + }, + { + "epoch": 0.8264932209600586, + "grad_norm": 0.3264601268533289, + "learning_rate": 8.912387999483463e-07, + "loss": 0.4307, + "step": 4511 + }, + { + "epoch": 0.8266764382557713, + "grad_norm": 0.4693347130690825, + "learning_rate": 8.894173473940676e-07, + "loss": 0.4578, + "step": 4512 + }, + { + "epoch": 0.8268596555514841, + "grad_norm": 0.4032452932999747, + "learning_rate": 8.875975762992428e-07, + "loss": 0.456, + "step": 4513 + }, + { + "epoch": 0.8270428728471968, + "grad_norm": 0.3560971563578487, + "learning_rate": 8.857794874082598e-07, + "loss": 0.4762, + "step": 4514 + }, + { + "epoch": 0.8272260901429095, + "grad_norm": 0.37176617363152936, + "learning_rate": 8.839630814648204e-07, + "loss": 0.4354, + "step": 4515 + }, + { + "epoch": 0.8274093074386222, + "grad_norm": 0.3730906962033253, + "learning_rate": 8.821483592119379e-07, + "loss": 0.4811, + "step": 4516 + }, + { + "epoch": 0.8275925247343349, + "grad_norm": 0.3808298852463878, + "learning_rate": 8.803353213919341e-07, + "loss": 0.4803, + "step": 4517 + }, + { + "epoch": 0.8277757420300477, + "grad_norm": 0.36005801175121405, + "learning_rate": 8.785239687464431e-07, + "loss": 0.485, + "step": 4518 + }, + { + "epoch": 0.8279589593257604, + "grad_norm": 0.36393977014250445, + "learning_rate": 8.767143020164104e-07, + "loss": 0.4864, + "step": 4519 + }, + { + "epoch": 0.828142176621473, + "grad_norm": 0.3760418887156118, + "learning_rate": 8.749063219420928e-07, + "loss": 0.4389, + "step": 4520 + }, + { + "epoch": 0.8283253939171857, + "grad_norm": 0.33289552891952595, + "learning_rate": 8.731000292630548e-07, + "loss": 0.4529, + "step": 4521 + }, + { + "epoch": 0.8285086112128985, + "grad_norm": 0.3514387850350512, + "learning_rate": 8.712954247181704e-07, + "loss": 0.4899, + "step": 4522 + }, + { + "epoch": 0.8286918285086112, + "grad_norm": 0.3880882911687274, + "learning_rate": 8.694925090456268e-07, + "loss": 0.4583, + "step": 4523 + }, + { + "epoch": 0.8288750458043239, + "grad_norm": 0.3930160698314478, + "learning_rate": 8.67691282982916e-07, + "loss": 0.4321, + "step": 4524 + }, + { + "epoch": 0.8290582631000366, + "grad_norm": 0.364602747991511, + "learning_rate": 8.658917472668421e-07, + "loss": 0.4724, + "step": 4525 + }, + { + "epoch": 0.8292414803957494, + "grad_norm": 0.36882224030459226, + "learning_rate": 8.640939026335144e-07, + "loss": 0.4855, + "step": 4526 + }, + { + "epoch": 0.8294246976914621, + "grad_norm": 0.3236084948360584, + "learning_rate": 8.622977498183555e-07, + "loss": 0.4674, + "step": 4527 + }, + { + "epoch": 0.8296079149871748, + "grad_norm": 0.3756416592206287, + "learning_rate": 8.605032895560894e-07, + "loss": 0.4703, + "step": 4528 + }, + { + "epoch": 0.8297911322828875, + "grad_norm": 0.35000816914720373, + "learning_rate": 8.587105225807546e-07, + "loss": 0.4661, + "step": 4529 + }, + { + "epoch": 0.8299743495786002, + "grad_norm": 0.37679187067257325, + "learning_rate": 8.569194496256905e-07, + "loss": 0.481, + "step": 4530 + }, + { + "epoch": 0.830157566874313, + "grad_norm": 0.3590180280771207, + "learning_rate": 8.551300714235494e-07, + "loss": 0.4742, + "step": 4531 + }, + { + "epoch": 0.8303407841700257, + "grad_norm": 0.3595502464365226, + "learning_rate": 8.533423887062858e-07, + "loss": 0.4822, + "step": 4532 + }, + { + "epoch": 0.8305240014657383, + "grad_norm": 0.39993908061073036, + "learning_rate": 8.515564022051636e-07, + "loss": 0.4947, + "step": 4533 + }, + { + "epoch": 0.830707218761451, + "grad_norm": 0.3616132991433677, + "learning_rate": 8.497721126507502e-07, + "loss": 0.4789, + "step": 4534 + }, + { + "epoch": 0.8308904360571638, + "grad_norm": 0.3765828950687697, + "learning_rate": 8.479895207729227e-07, + "loss": 0.4765, + "step": 4535 + }, + { + "epoch": 0.8310736533528765, + "grad_norm": 0.36266920101078504, + "learning_rate": 8.462086273008585e-07, + "loss": 0.4773, + "step": 4536 + }, + { + "epoch": 0.8312568706485892, + "grad_norm": 0.3457741644742753, + "learning_rate": 8.444294329630464e-07, + "loss": 0.4504, + "step": 4537 + }, + { + "epoch": 0.8314400879443019, + "grad_norm": 4.365216891837018, + "learning_rate": 8.426519384872733e-07, + "loss": 0.4294, + "step": 4538 + }, + { + "epoch": 0.8316233052400147, + "grad_norm": 0.3824653477334578, + "learning_rate": 8.408761446006381e-07, + "loss": 0.4627, + "step": 4539 + }, + { + "epoch": 0.8318065225357274, + "grad_norm": 0.34948308071354306, + "learning_rate": 8.391020520295384e-07, + "loss": 0.4661, + "step": 4540 + }, + { + "epoch": 0.8319897398314401, + "grad_norm": 0.34127634807978796, + "learning_rate": 8.373296614996773e-07, + "loss": 0.4382, + "step": 4541 + }, + { + "epoch": 0.8321729571271528, + "grad_norm": 0.3451205951701214, + "learning_rate": 8.355589737360636e-07, + "loss": 0.4543, + "step": 4542 + }, + { + "epoch": 0.8323561744228655, + "grad_norm": 0.3638514465050463, + "learning_rate": 8.337899894630064e-07, + "loss": 0.4736, + "step": 4543 + }, + { + "epoch": 0.8325393917185783, + "grad_norm": 0.40569003841764567, + "learning_rate": 8.320227094041222e-07, + "loss": 0.4962, + "step": 4544 + }, + { + "epoch": 0.832722609014291, + "grad_norm": 0.4019440010035879, + "learning_rate": 8.302571342823251e-07, + "loss": 0.4465, + "step": 4545 + }, + { + "epoch": 0.8329058263100036, + "grad_norm": 0.36318742919496133, + "learning_rate": 8.284932648198374e-07, + "loss": 0.482, + "step": 4546 + }, + { + "epoch": 0.8330890436057163, + "grad_norm": 0.3341546485213257, + "learning_rate": 8.267311017381779e-07, + "loss": 0.4519, + "step": 4547 + }, + { + "epoch": 0.8332722609014291, + "grad_norm": 0.3690645647843955, + "learning_rate": 8.249706457581741e-07, + "loss": 0.4529, + "step": 4548 + }, + { + "epoch": 0.8334554781971418, + "grad_norm": 0.3965501911600018, + "learning_rate": 8.232118975999476e-07, + "loss": 0.4442, + "step": 4549 + }, + { + "epoch": 0.8336386954928545, + "grad_norm": 0.4539094325456942, + "learning_rate": 8.214548579829285e-07, + "loss": 0.4607, + "step": 4550 + }, + { + "epoch": 0.8338219127885672, + "grad_norm": 0.3614299321405626, + "learning_rate": 8.196995276258424e-07, + "loss": 0.4841, + "step": 4551 + }, + { + "epoch": 0.83400513008428, + "grad_norm": 0.3650558358082691, + "learning_rate": 8.1794590724672e-07, + "loss": 0.4491, + "step": 4552 + }, + { + "epoch": 0.8341883473799927, + "grad_norm": 0.37021208019651697, + "learning_rate": 8.161939975628891e-07, + "loss": 0.4621, + "step": 4553 + }, + { + "epoch": 0.8343715646757054, + "grad_norm": 0.3187081455195085, + "learning_rate": 8.14443799290981e-07, + "loss": 0.4505, + "step": 4554 + }, + { + "epoch": 0.8345547819714181, + "grad_norm": 0.34998813799408895, + "learning_rate": 8.126953131469229e-07, + "loss": 0.4604, + "step": 4555 + }, + { + "epoch": 0.8347379992671308, + "grad_norm": 0.35863274556981756, + "learning_rate": 8.109485398459466e-07, + "loss": 0.4683, + "step": 4556 + }, + { + "epoch": 0.8349212165628436, + "grad_norm": 0.3872812157147719, + "learning_rate": 8.092034801025789e-07, + "loss": 0.4677, + "step": 4557 + }, + { + "epoch": 0.8351044338585563, + "grad_norm": 0.3537014025205882, + "learning_rate": 8.074601346306465e-07, + "loss": 0.4403, + "step": 4558 + }, + { + "epoch": 0.835287651154269, + "grad_norm": 0.3257981845848479, + "learning_rate": 8.057185041432775e-07, + "loss": 0.4784, + "step": 4559 + }, + { + "epoch": 0.8354708684499816, + "grad_norm": 0.34120774188761716, + "learning_rate": 8.039785893528973e-07, + "loss": 0.4537, + "step": 4560 + }, + { + "epoch": 0.8356540857456944, + "grad_norm": 0.38295155691945104, + "learning_rate": 8.022403909712272e-07, + "loss": 0.4642, + "step": 4561 + }, + { + "epoch": 0.8358373030414071, + "grad_norm": 0.37025660534210014, + "learning_rate": 8.005039097092876e-07, + "loss": 0.4711, + "step": 4562 + }, + { + "epoch": 0.8360205203371198, + "grad_norm": 0.38506132031294427, + "learning_rate": 7.987691462773983e-07, + "loss": 0.4603, + "step": 4563 + }, + { + "epoch": 0.8362037376328325, + "grad_norm": 0.3906299247233729, + "learning_rate": 7.970361013851758e-07, + "loss": 0.4569, + "step": 4564 + }, + { + "epoch": 0.8363869549285452, + "grad_norm": 0.37287175398490147, + "learning_rate": 7.953047757415322e-07, + "loss": 0.5081, + "step": 4565 + }, + { + "epoch": 0.836570172224258, + "grad_norm": 0.3954122070934997, + "learning_rate": 7.935751700546751e-07, + "loss": 0.4915, + "step": 4566 + }, + { + "epoch": 0.8367533895199707, + "grad_norm": 0.33981736742323343, + "learning_rate": 7.918472850321124e-07, + "loss": 0.444, + "step": 4567 + }, + { + "epoch": 0.8369366068156834, + "grad_norm": 0.4118368681892205, + "learning_rate": 7.901211213806475e-07, + "loss": 0.4653, + "step": 4568 + }, + { + "epoch": 0.8371198241113961, + "grad_norm": 0.362321331702295, + "learning_rate": 7.883966798063764e-07, + "loss": 0.5057, + "step": 4569 + }, + { + "epoch": 0.8373030414071089, + "grad_norm": 0.40249917563877835, + "learning_rate": 7.866739610146923e-07, + "loss": 0.4566, + "step": 4570 + }, + { + "epoch": 0.8374862587028216, + "grad_norm": 0.33727746860853636, + "learning_rate": 7.84952965710285e-07, + "loss": 0.4452, + "step": 4571 + }, + { + "epoch": 0.8376694759985343, + "grad_norm": 0.34323136464763987, + "learning_rate": 7.832336945971402e-07, + "loss": 0.4733, + "step": 4572 + }, + { + "epoch": 0.8378526932942469, + "grad_norm": 0.31278841887861725, + "learning_rate": 7.815161483785328e-07, + "loss": 0.4385, + "step": 4573 + }, + { + "epoch": 0.8380359105899597, + "grad_norm": 0.34935283325144817, + "learning_rate": 7.79800327757037e-07, + "loss": 0.476, + "step": 4574 + }, + { + "epoch": 0.8382191278856724, + "grad_norm": 0.35699351656051326, + "learning_rate": 7.780862334345218e-07, + "loss": 0.4683, + "step": 4575 + }, + { + "epoch": 0.8384023451813851, + "grad_norm": 0.35724251812079494, + "learning_rate": 7.763738661121467e-07, + "loss": 0.4616, + "step": 4576 + }, + { + "epoch": 0.8385855624770978, + "grad_norm": 0.6989739009443403, + "learning_rate": 7.746632264903652e-07, + "loss": 0.4465, + "step": 4577 + }, + { + "epoch": 0.8387687797728105, + "grad_norm": 0.36127231861511705, + "learning_rate": 7.729543152689262e-07, + "loss": 0.477, + "step": 4578 + }, + { + "epoch": 0.8389519970685233, + "grad_norm": 0.3761840834901936, + "learning_rate": 7.712471331468718e-07, + "loss": 0.4401, + "step": 4579 + }, + { + "epoch": 0.839135214364236, + "grad_norm": 0.3168256929816936, + "learning_rate": 7.695416808225342e-07, + "loss": 0.4645, + "step": 4580 + }, + { + "epoch": 0.8393184316599487, + "grad_norm": 0.3781685502784412, + "learning_rate": 7.678379589935381e-07, + "loss": 0.467, + "step": 4581 + }, + { + "epoch": 0.8395016489556614, + "grad_norm": 0.38774315231911183, + "learning_rate": 7.661359683568032e-07, + "loss": 0.4744, + "step": 4582 + }, + { + "epoch": 0.8396848662513742, + "grad_norm": 0.3757045095906377, + "learning_rate": 7.644357096085398e-07, + "loss": 0.4523, + "step": 4583 + }, + { + "epoch": 0.8398680835470869, + "grad_norm": 0.3467435366761741, + "learning_rate": 7.627371834442481e-07, + "loss": 0.4668, + "step": 4584 + }, + { + "epoch": 0.8400513008427996, + "grad_norm": 0.3448221469385914, + "learning_rate": 7.610403905587205e-07, + "loss": 0.4542, + "step": 4585 + }, + { + "epoch": 0.8402345181385122, + "grad_norm": 0.3808606706880448, + "learning_rate": 7.593453316460408e-07, + "loss": 0.4499, + "step": 4586 + }, + { + "epoch": 0.840417735434225, + "grad_norm": 0.3216050914197737, + "learning_rate": 7.576520073995858e-07, + "loss": 0.4237, + "step": 4587 + }, + { + "epoch": 0.8406009527299377, + "grad_norm": 0.36041465524446253, + "learning_rate": 7.559604185120184e-07, + "loss": 0.4637, + "step": 4588 + }, + { + "epoch": 0.8407841700256504, + "grad_norm": 0.38480181308486905, + "learning_rate": 7.542705656752924e-07, + "loss": 0.432, + "step": 4589 + }, + { + "epoch": 0.8409673873213631, + "grad_norm": 0.3697932660124222, + "learning_rate": 7.525824495806539e-07, + "loss": 0.4634, + "step": 4590 + }, + { + "epoch": 0.8411506046170758, + "grad_norm": 0.35969965173141866, + "learning_rate": 7.508960709186386e-07, + "loss": 0.4463, + "step": 4591 + }, + { + "epoch": 0.8413338219127886, + "grad_norm": 0.38970691229065313, + "learning_rate": 7.492114303790693e-07, + "loss": 0.4727, + "step": 4592 + }, + { + "epoch": 0.8415170392085013, + "grad_norm": 0.40204280376514445, + "learning_rate": 7.475285286510569e-07, + "loss": 0.4842, + "step": 4593 + }, + { + "epoch": 0.841700256504214, + "grad_norm": 0.39678581293302634, + "learning_rate": 7.458473664230054e-07, + "loss": 0.5168, + "step": 4594 + }, + { + "epoch": 0.8418834737999267, + "grad_norm": 0.41301539955038935, + "learning_rate": 7.441679443826022e-07, + "loss": 0.4753, + "step": 4595 + }, + { + "epoch": 0.8420666910956395, + "grad_norm": 0.3277047194415458, + "learning_rate": 7.424902632168285e-07, + "loss": 0.4779, + "step": 4596 + }, + { + "epoch": 0.8422499083913522, + "grad_norm": 0.36117154118790784, + "learning_rate": 7.408143236119469e-07, + "loss": 0.4528, + "step": 4597 + }, + { + "epoch": 0.8424331256870649, + "grad_norm": 0.36900471245720073, + "learning_rate": 7.391401262535141e-07, + "loss": 0.4835, + "step": 4598 + }, + { + "epoch": 0.8426163429827775, + "grad_norm": 0.33552736713541453, + "learning_rate": 7.37467671826368e-07, + "loss": 0.496, + "step": 4599 + }, + { + "epoch": 0.8427995602784902, + "grad_norm": 0.40437208542001996, + "learning_rate": 7.357969610146387e-07, + "loss": 0.4269, + "step": 4600 + }, + { + "epoch": 0.842982777574203, + "grad_norm": 0.3520114205169631, + "learning_rate": 7.341279945017393e-07, + "loss": 0.4534, + "step": 4601 + }, + { + "epoch": 0.8431659948699157, + "grad_norm": 0.37756325907570126, + "learning_rate": 7.324607729703731e-07, + "loss": 0.4719, + "step": 4602 + }, + { + "epoch": 0.8433492121656284, + "grad_norm": 0.41618750381433833, + "learning_rate": 7.307952971025245e-07, + "loss": 0.4486, + "step": 4603 + }, + { + "epoch": 0.8435324294613411, + "grad_norm": 0.372896456871397, + "learning_rate": 7.291315675794703e-07, + "loss": 0.4437, + "step": 4604 + }, + { + "epoch": 0.8437156467570539, + "grad_norm": 0.35730637379891195, + "learning_rate": 7.274695850817654e-07, + "loss": 0.455, + "step": 4605 + }, + { + "epoch": 0.8438988640527666, + "grad_norm": 0.3864061068477958, + "learning_rate": 7.258093502892577e-07, + "loss": 0.4579, + "step": 4606 + }, + { + "epoch": 0.8440820813484793, + "grad_norm": 0.36397130054529137, + "learning_rate": 7.241508638810741e-07, + "loss": 0.4661, + "step": 4607 + }, + { + "epoch": 0.844265298644192, + "grad_norm": 0.39478183132434946, + "learning_rate": 7.224941265356311e-07, + "loss": 0.4572, + "step": 4608 + }, + { + "epoch": 0.8444485159399048, + "grad_norm": 0.35238010256874025, + "learning_rate": 7.208391389306257e-07, + "loss": 0.4665, + "step": 4609 + }, + { + "epoch": 0.8446317332356175, + "grad_norm": 0.39054270140359354, + "learning_rate": 7.191859017430403e-07, + "loss": 0.4688, + "step": 4610 + }, + { + "epoch": 0.8448149505313302, + "grad_norm": 0.3722872319474878, + "learning_rate": 7.175344156491432e-07, + "loss": 0.4762, + "step": 4611 + }, + { + "epoch": 0.8449981678270428, + "grad_norm": 0.39812658488136404, + "learning_rate": 7.158846813244857e-07, + "loss": 0.4306, + "step": 4612 + }, + { + "epoch": 0.8451813851227555, + "grad_norm": 0.3235410643580283, + "learning_rate": 7.14236699443901e-07, + "loss": 0.4748, + "step": 4613 + }, + { + "epoch": 0.8453646024184683, + "grad_norm": 0.3645862100537474, + "learning_rate": 7.125904706815051e-07, + "loss": 0.4708, + "step": 4614 + }, + { + "epoch": 0.845547819714181, + "grad_norm": 0.41444318001972247, + "learning_rate": 7.109459957106995e-07, + "loss": 0.45, + "step": 4615 + }, + { + "epoch": 0.8457310370098937, + "grad_norm": 0.33835473447341924, + "learning_rate": 7.093032752041678e-07, + "loss": 0.4642, + "step": 4616 + }, + { + "epoch": 0.8459142543056064, + "grad_norm": 0.37409991259381564, + "learning_rate": 7.076623098338742e-07, + "loss": 0.4533, + "step": 4617 + }, + { + "epoch": 0.8460974716013192, + "grad_norm": 0.35507840202053215, + "learning_rate": 7.060231002710643e-07, + "loss": 0.4641, + "step": 4618 + }, + { + "epoch": 0.8462806888970319, + "grad_norm": 0.37641675654011486, + "learning_rate": 7.043856471862692e-07, + "loss": 0.4574, + "step": 4619 + }, + { + "epoch": 0.8464639061927446, + "grad_norm": 0.38077617847905443, + "learning_rate": 7.027499512492985e-07, + "loss": 0.4936, + "step": 4620 + }, + { + "epoch": 0.8466471234884573, + "grad_norm": 0.38213562762661146, + "learning_rate": 7.01116013129245e-07, + "loss": 0.4767, + "step": 4621 + }, + { + "epoch": 0.8468303407841701, + "grad_norm": 0.44818862839111945, + "learning_rate": 6.994838334944781e-07, + "loss": 0.471, + "step": 4622 + }, + { + "epoch": 0.8470135580798828, + "grad_norm": 0.40287277556776896, + "learning_rate": 6.97853413012653e-07, + "loss": 0.5018, + "step": 4623 + }, + { + "epoch": 0.8471967753755955, + "grad_norm": 0.3475657194452658, + "learning_rate": 6.96224752350706e-07, + "loss": 0.4854, + "step": 4624 + }, + { + "epoch": 0.8473799926713081, + "grad_norm": 0.49153236671871353, + "learning_rate": 6.945978521748458e-07, + "loss": 0.4745, + "step": 4625 + }, + { + "epoch": 0.8475632099670208, + "grad_norm": 0.3627963856399352, + "learning_rate": 6.929727131505687e-07, + "loss": 0.4666, + "step": 4626 + }, + { + "epoch": 0.8477464272627336, + "grad_norm": 0.37178156831060366, + "learning_rate": 6.913493359426476e-07, + "loss": 0.4569, + "step": 4627 + }, + { + "epoch": 0.8479296445584463, + "grad_norm": 0.3935110501040928, + "learning_rate": 6.897277212151376e-07, + "loss": 0.4559, + "step": 4628 + }, + { + "epoch": 0.848112861854159, + "grad_norm": 0.37119806150514023, + "learning_rate": 6.881078696313658e-07, + "loss": 0.4737, + "step": 4629 + }, + { + "epoch": 0.8482960791498717, + "grad_norm": 0.36940982032393316, + "learning_rate": 6.864897818539445e-07, + "loss": 0.4726, + "step": 4630 + }, + { + "epoch": 0.8484792964455845, + "grad_norm": 0.4236420908531626, + "learning_rate": 6.848734585447641e-07, + "loss": 0.4897, + "step": 4631 + }, + { + "epoch": 0.8486625137412972, + "grad_norm": 0.4137058543064466, + "learning_rate": 6.832589003649909e-07, + "loss": 0.4858, + "step": 4632 + }, + { + "epoch": 0.8488457310370099, + "grad_norm": 0.35295538051409014, + "learning_rate": 6.816461079750675e-07, + "loss": 0.4847, + "step": 4633 + }, + { + "epoch": 0.8490289483327226, + "grad_norm": 0.3680249235559555, + "learning_rate": 6.800350820347196e-07, + "loss": 0.4559, + "step": 4634 + }, + { + "epoch": 0.8492121656284354, + "grad_norm": 0.35440622742913136, + "learning_rate": 6.784258232029473e-07, + "loss": 0.4633, + "step": 4635 + }, + { + "epoch": 0.8493953829241481, + "grad_norm": 0.35474342649832563, + "learning_rate": 6.768183321380267e-07, + "loss": 0.4842, + "step": 4636 + }, + { + "epoch": 0.8495786002198608, + "grad_norm": 0.3447445863478664, + "learning_rate": 6.75212609497512e-07, + "loss": 0.4537, + "step": 4637 + }, + { + "epoch": 0.8497618175155734, + "grad_norm": 0.39696044908830347, + "learning_rate": 6.736086559382343e-07, + "loss": 0.4833, + "step": 4638 + }, + { + "epoch": 0.8499450348112861, + "grad_norm": 0.40950492515990067, + "learning_rate": 6.720064721163022e-07, + "loss": 0.4451, + "step": 4639 + }, + { + "epoch": 0.8501282521069989, + "grad_norm": 0.3716416588918617, + "learning_rate": 6.704060586870981e-07, + "loss": 0.4945, + "step": 4640 + }, + { + "epoch": 0.8503114694027116, + "grad_norm": 0.3539582725690085, + "learning_rate": 6.688074163052793e-07, + "loss": 0.4801, + "step": 4641 + }, + { + "epoch": 0.8504946866984243, + "grad_norm": 0.35801593050898506, + "learning_rate": 6.672105456247824e-07, + "loss": 0.4274, + "step": 4642 + }, + { + "epoch": 0.850677903994137, + "grad_norm": 0.3769100024892945, + "learning_rate": 6.656154472988174e-07, + "loss": 0.4546, + "step": 4643 + }, + { + "epoch": 0.8508611212898498, + "grad_norm": 0.3347319601400921, + "learning_rate": 6.640221219798692e-07, + "loss": 0.4661, + "step": 4644 + }, + { + "epoch": 0.8510443385855625, + "grad_norm": 0.35012115439903235, + "learning_rate": 6.624305703196959e-07, + "loss": 0.462, + "step": 4645 + }, + { + "epoch": 0.8512275558812752, + "grad_norm": 0.6464519568025817, + "learning_rate": 6.608407929693333e-07, + "loss": 0.4871, + "step": 4646 + }, + { + "epoch": 0.8514107731769879, + "grad_norm": 0.3239331939209271, + "learning_rate": 6.592527905790885e-07, + "loss": 0.4801, + "step": 4647 + }, + { + "epoch": 0.8515939904727006, + "grad_norm": 0.35402847628223827, + "learning_rate": 6.576665637985452e-07, + "loss": 0.4689, + "step": 4648 + }, + { + "epoch": 0.8517772077684134, + "grad_norm": 0.36588071309071396, + "learning_rate": 6.560821132765577e-07, + "loss": 0.4798, + "step": 4649 + }, + { + "epoch": 0.8519604250641261, + "grad_norm": 0.36842514750128835, + "learning_rate": 6.544994396612569e-07, + "loss": 0.481, + "step": 4650 + }, + { + "epoch": 0.8521436423598387, + "grad_norm": 0.5433760869689138, + "learning_rate": 6.529185436000435e-07, + "loss": 0.4491, + "step": 4651 + }, + { + "epoch": 0.8523268596555514, + "grad_norm": 0.36308528984837474, + "learning_rate": 6.51339425739595e-07, + "loss": 0.5027, + "step": 4652 + }, + { + "epoch": 0.8525100769512642, + "grad_norm": 0.43046695940095536, + "learning_rate": 6.497620867258575e-07, + "loss": 0.4979, + "step": 4653 + }, + { + "epoch": 0.8526932942469769, + "grad_norm": 0.36860487992630503, + "learning_rate": 6.481865272040532e-07, + "loss": 0.4715, + "step": 4654 + }, + { + "epoch": 0.8528765115426896, + "grad_norm": 0.36651683339123137, + "learning_rate": 6.466127478186729e-07, + "loss": 0.4731, + "step": 4655 + }, + { + "epoch": 0.8530597288384023, + "grad_norm": 0.3609707087616487, + "learning_rate": 6.450407492134825e-07, + "loss": 0.4746, + "step": 4656 + }, + { + "epoch": 0.8532429461341151, + "grad_norm": 0.3596774652911952, + "learning_rate": 6.434705320315155e-07, + "loss": 0.4583, + "step": 4657 + }, + { + "epoch": 0.8534261634298278, + "grad_norm": 0.380847382228284, + "learning_rate": 6.419020969150819e-07, + "loss": 0.4691, + "step": 4658 + }, + { + "epoch": 0.8536093807255405, + "grad_norm": 0.33756954312322307, + "learning_rate": 6.403354445057569e-07, + "loss": 0.4692, + "step": 4659 + }, + { + "epoch": 0.8537925980212532, + "grad_norm": 0.3519944492706372, + "learning_rate": 6.387705754443918e-07, + "loss": 0.4476, + "step": 4660 + }, + { + "epoch": 0.8539758153169659, + "grad_norm": 0.35212978264435935, + "learning_rate": 6.372074903711056e-07, + "loss": 0.4932, + "step": 4661 + }, + { + "epoch": 0.8541590326126787, + "grad_norm": 0.3419603578191104, + "learning_rate": 6.356461899252863e-07, + "loss": 0.441, + "step": 4662 + }, + { + "epoch": 0.8543422499083914, + "grad_norm": 0.36161456417265275, + "learning_rate": 6.340866747455949e-07, + "loss": 0.4429, + "step": 4663 + }, + { + "epoch": 0.854525467204104, + "grad_norm": 0.3254519388964185, + "learning_rate": 6.325289454699612e-07, + "loss": 0.4607, + "step": 4664 + }, + { + "epoch": 0.8547086844998167, + "grad_norm": 0.3828469440501933, + "learning_rate": 6.309730027355842e-07, + "loss": 0.4534, + "step": 4665 + }, + { + "epoch": 0.8548919017955295, + "grad_norm": 0.4481534406657637, + "learning_rate": 6.294188471789297e-07, + "loss": 0.4883, + "step": 4666 + }, + { + "epoch": 0.8550751190912422, + "grad_norm": 0.40650799766285306, + "learning_rate": 6.278664794357369e-07, + "loss": 0.5197, + "step": 4667 + }, + { + "epoch": 0.8552583363869549, + "grad_norm": 0.3557778886699918, + "learning_rate": 6.263159001410118e-07, + "loss": 0.4468, + "step": 4668 + }, + { + "epoch": 0.8554415536826676, + "grad_norm": 0.3155663993825966, + "learning_rate": 6.247671099290281e-07, + "loss": 0.4315, + "step": 4669 + }, + { + "epoch": 0.8556247709783804, + "grad_norm": 0.4241376334024175, + "learning_rate": 6.232201094333262e-07, + "loss": 0.4659, + "step": 4670 + }, + { + "epoch": 0.8558079882740931, + "grad_norm": 0.39926460741940456, + "learning_rate": 6.216748992867178e-07, + "loss": 0.4646, + "step": 4671 + }, + { + "epoch": 0.8559912055698058, + "grad_norm": 0.37448213112394435, + "learning_rate": 6.201314801212822e-07, + "loss": 0.4629, + "step": 4672 + }, + { + "epoch": 0.8561744228655185, + "grad_norm": 0.32841351785338857, + "learning_rate": 6.185898525683625e-07, + "loss": 0.4651, + "step": 4673 + }, + { + "epoch": 0.8563576401612312, + "grad_norm": 0.35687683363139366, + "learning_rate": 6.170500172585708e-07, + "loss": 0.473, + "step": 4674 + }, + { + "epoch": 0.856540857456944, + "grad_norm": 0.5940246428460522, + "learning_rate": 6.155119748217874e-07, + "loss": 0.4631, + "step": 4675 + }, + { + "epoch": 0.8567240747526567, + "grad_norm": 0.36018176214900893, + "learning_rate": 6.139757258871604e-07, + "loss": 0.5154, + "step": 4676 + }, + { + "epoch": 0.8569072920483694, + "grad_norm": 0.4687679294939073, + "learning_rate": 6.124412710830968e-07, + "loss": 0.4594, + "step": 4677 + }, + { + "epoch": 0.857090509344082, + "grad_norm": 0.35388482226740736, + "learning_rate": 6.109086110372781e-07, + "loss": 0.4517, + "step": 4678 + }, + { + "epoch": 0.8572737266397948, + "grad_norm": 0.36328352595149627, + "learning_rate": 6.093777463766481e-07, + "loss": 0.4633, + "step": 4679 + }, + { + "epoch": 0.8574569439355075, + "grad_norm": 0.3614749246413115, + "learning_rate": 6.078486777274184e-07, + "loss": 0.4672, + "step": 4680 + }, + { + "epoch": 0.8576401612312202, + "grad_norm": 0.35822452744121946, + "learning_rate": 6.063214057150602e-07, + "loss": 0.4466, + "step": 4681 + }, + { + "epoch": 0.8578233785269329, + "grad_norm": 0.443689158246001, + "learning_rate": 6.047959309643153e-07, + "loss": 0.4841, + "step": 4682 + }, + { + "epoch": 0.8580065958226457, + "grad_norm": 0.3664501289628841, + "learning_rate": 6.032722540991897e-07, + "loss": 0.4413, + "step": 4683 + }, + { + "epoch": 0.8581898131183584, + "grad_norm": 0.37392123974160224, + "learning_rate": 6.017503757429527e-07, + "loss": 0.498, + "step": 4684 + }, + { + "epoch": 0.8583730304140711, + "grad_norm": 0.3442340298384154, + "learning_rate": 6.002302965181362e-07, + "loss": 0.4873, + "step": 4685 + }, + { + "epoch": 0.8585562477097838, + "grad_norm": 0.38482719467034343, + "learning_rate": 5.9871201704654e-07, + "loss": 0.4601, + "step": 4686 + }, + { + "epoch": 0.8587394650054965, + "grad_norm": 0.38816231667636386, + "learning_rate": 5.971955379492256e-07, + "loss": 0.4804, + "step": 4687 + }, + { + "epoch": 0.8589226823012093, + "grad_norm": 0.42786975057807836, + "learning_rate": 5.95680859846518e-07, + "loss": 0.4349, + "step": 4688 + }, + { + "epoch": 0.859105899596922, + "grad_norm": 0.34136049119828255, + "learning_rate": 5.941679833580044e-07, + "loss": 0.4437, + "step": 4689 + }, + { + "epoch": 0.8592891168926347, + "grad_norm": 0.3978177457577374, + "learning_rate": 5.92656909102538e-07, + "loss": 0.4866, + "step": 4690 + }, + { + "epoch": 0.8594723341883473, + "grad_norm": 0.3329572161638651, + "learning_rate": 5.911476376982333e-07, + "loss": 0.4767, + "step": 4691 + }, + { + "epoch": 0.8596555514840601, + "grad_norm": 0.3429759165018252, + "learning_rate": 5.896401697624665e-07, + "loss": 0.49, + "step": 4692 + }, + { + "epoch": 0.8598387687797728, + "grad_norm": 0.356355519401416, + "learning_rate": 5.881345059118759e-07, + "loss": 0.4549, + "step": 4693 + }, + { + "epoch": 0.8600219860754855, + "grad_norm": 0.4113799697444653, + "learning_rate": 5.866306467623633e-07, + "loss": 0.484, + "step": 4694 + }, + { + "epoch": 0.8602052033711982, + "grad_norm": 0.4101610555768186, + "learning_rate": 5.851285929290934e-07, + "loss": 0.4533, + "step": 4695 + }, + { + "epoch": 0.8603884206669109, + "grad_norm": 0.3338706073643328, + "learning_rate": 5.83628345026489e-07, + "loss": 0.4376, + "step": 4696 + }, + { + "epoch": 0.8605716379626237, + "grad_norm": 0.3934260519301321, + "learning_rate": 5.821299036682354e-07, + "loss": 0.4617, + "step": 4697 + }, + { + "epoch": 0.8607548552583364, + "grad_norm": 0.44033773382119257, + "learning_rate": 5.806332694672806e-07, + "loss": 0.476, + "step": 4698 + }, + { + "epoch": 0.8609380725540491, + "grad_norm": 0.5362324156637753, + "learning_rate": 5.79138443035831e-07, + "loss": 0.4608, + "step": 4699 + }, + { + "epoch": 0.8611212898497618, + "grad_norm": 0.3689328897257396, + "learning_rate": 5.776454249853558e-07, + "loss": 0.4467, + "step": 4700 + }, + { + "epoch": 0.8613045071454746, + "grad_norm": 0.36106098966365785, + "learning_rate": 5.761542159265821e-07, + "loss": 0.4736, + "step": 4701 + }, + { + "epoch": 0.8614877244411873, + "grad_norm": 0.30462929642826386, + "learning_rate": 5.746648164695001e-07, + "loss": 0.4491, + "step": 4702 + }, + { + "epoch": 0.8616709417369, + "grad_norm": 0.3660624254995241, + "learning_rate": 5.731772272233554e-07, + "loss": 0.462, + "step": 4703 + }, + { + "epoch": 0.8618541590326126, + "grad_norm": 0.3916188199259428, + "learning_rate": 5.71691448796658e-07, + "loss": 0.5185, + "step": 4704 + }, + { + "epoch": 0.8620373763283254, + "grad_norm": 0.42360402026125726, + "learning_rate": 5.702074817971731e-07, + "loss": 0.4757, + "step": 4705 + }, + { + "epoch": 0.8622205936240381, + "grad_norm": 0.30443437754622965, + "learning_rate": 5.687253268319287e-07, + "loss": 0.4759, + "step": 4706 + }, + { + "epoch": 0.8624038109197508, + "grad_norm": 0.3762594663400903, + "learning_rate": 5.67244984507207e-07, + "loss": 0.4832, + "step": 4707 + }, + { + "epoch": 0.8625870282154635, + "grad_norm": 0.3828874654360133, + "learning_rate": 5.657664554285536e-07, + "loss": 0.4486, + "step": 4708 + }, + { + "epoch": 0.8627702455111762, + "grad_norm": 0.32478632825544107, + "learning_rate": 5.642897402007674e-07, + "loss": 0.483, + "step": 4709 + }, + { + "epoch": 0.862953462806889, + "grad_norm": 0.37189531684212246, + "learning_rate": 5.62814839427911e-07, + "loss": 0.4484, + "step": 4710 + }, + { + "epoch": 0.8631366801026017, + "grad_norm": 0.38212786115010916, + "learning_rate": 5.613417537132992e-07, + "loss": 0.4665, + "step": 4711 + }, + { + "epoch": 0.8633198973983144, + "grad_norm": 0.39179768082163485, + "learning_rate": 5.598704836595093e-07, + "loss": 0.4744, + "step": 4712 + }, + { + "epoch": 0.8635031146940271, + "grad_norm": 0.35748445849400146, + "learning_rate": 5.584010298683718e-07, + "loss": 0.4703, + "step": 4713 + }, + { + "epoch": 0.8636863319897399, + "grad_norm": 0.3360163273163187, + "learning_rate": 5.569333929409749e-07, + "loss": 0.4756, + "step": 4714 + }, + { + "epoch": 0.8638695492854526, + "grad_norm": 0.3447769988570555, + "learning_rate": 5.554675734776666e-07, + "loss": 0.4874, + "step": 4715 + }, + { + "epoch": 0.8640527665811653, + "grad_norm": 0.3571674314958612, + "learning_rate": 5.540035720780496e-07, + "loss": 0.4896, + "step": 4716 + }, + { + "epoch": 0.864235983876878, + "grad_norm": 0.3806123348007476, + "learning_rate": 5.52541389340982e-07, + "loss": 0.4954, + "step": 4717 + }, + { + "epoch": 0.8644192011725907, + "grad_norm": 0.4451181887254169, + "learning_rate": 5.510810258645783e-07, + "loss": 0.4895, + "step": 4718 + }, + { + "epoch": 0.8646024184683034, + "grad_norm": 0.35671002251334094, + "learning_rate": 5.496224822462093e-07, + "loss": 0.4778, + "step": 4719 + }, + { + "epoch": 0.8647856357640161, + "grad_norm": 0.3258167210449754, + "learning_rate": 5.481657590825029e-07, + "loss": 0.4835, + "step": 4720 + }, + { + "epoch": 0.8649688530597288, + "grad_norm": 0.38838912919272484, + "learning_rate": 5.467108569693403e-07, + "loss": 0.4835, + "step": 4721 + }, + { + "epoch": 0.8651520703554415, + "grad_norm": 0.3790292808401653, + "learning_rate": 5.452577765018563e-07, + "loss": 0.4752, + "step": 4722 + }, + { + "epoch": 0.8653352876511543, + "grad_norm": 0.4244261923396075, + "learning_rate": 5.43806518274444e-07, + "loss": 0.4674, + "step": 4723 + }, + { + "epoch": 0.865518504946867, + "grad_norm": 0.34567247706024967, + "learning_rate": 5.423570828807512e-07, + "loss": 0.4708, + "step": 4724 + }, + { + "epoch": 0.8657017222425797, + "grad_norm": 0.34885877088811224, + "learning_rate": 5.409094709136764e-07, + "loss": 0.4722, + "step": 4725 + }, + { + "epoch": 0.8658849395382924, + "grad_norm": 0.4690849601959767, + "learning_rate": 5.394636829653738e-07, + "loss": 0.4706, + "step": 4726 + }, + { + "epoch": 0.8660681568340052, + "grad_norm": 0.32339419909309275, + "learning_rate": 5.380197196272529e-07, + "loss": 0.4419, + "step": 4727 + }, + { + "epoch": 0.8662513741297179, + "grad_norm": 0.3435968523036632, + "learning_rate": 5.365775814899771e-07, + "loss": 0.4658, + "step": 4728 + }, + { + "epoch": 0.8664345914254306, + "grad_norm": 0.41714361742399847, + "learning_rate": 5.351372691434603e-07, + "loss": 0.4458, + "step": 4729 + }, + { + "epoch": 0.8666178087211432, + "grad_norm": 0.3380184941975371, + "learning_rate": 5.336987831768714e-07, + "loss": 0.4659, + "step": 4730 + }, + { + "epoch": 0.866801026016856, + "grad_norm": 0.36595369071152395, + "learning_rate": 5.322621241786325e-07, + "loss": 0.4874, + "step": 4731 + }, + { + "epoch": 0.8669842433125687, + "grad_norm": 0.32017617118250824, + "learning_rate": 5.308272927364194e-07, + "loss": 0.4417, + "step": 4732 + }, + { + "epoch": 0.8671674606082814, + "grad_norm": 0.3844876584806139, + "learning_rate": 5.293942894371556e-07, + "loss": 0.4617, + "step": 4733 + }, + { + "epoch": 0.8673506779039941, + "grad_norm": 0.34264875726785166, + "learning_rate": 5.279631148670216e-07, + "loss": 0.4975, + "step": 4734 + }, + { + "epoch": 0.8675338951997068, + "grad_norm": 0.3903946140831761, + "learning_rate": 5.265337696114497e-07, + "loss": 0.474, + "step": 4735 + }, + { + "epoch": 0.8677171124954196, + "grad_norm": 0.39605878595595756, + "learning_rate": 5.251062542551211e-07, + "loss": 0.4196, + "step": 4736 + }, + { + "epoch": 0.8679003297911323, + "grad_norm": 0.4472024751314984, + "learning_rate": 5.236805693819685e-07, + "loss": 0.4737, + "step": 4737 + }, + { + "epoch": 0.868083547086845, + "grad_norm": 0.3467748168624409, + "learning_rate": 5.222567155751796e-07, + "loss": 0.4479, + "step": 4738 + }, + { + "epoch": 0.8682667643825577, + "grad_norm": 0.37063105808747765, + "learning_rate": 5.208346934171898e-07, + "loss": 0.4581, + "step": 4739 + }, + { + "epoch": 0.8684499816782705, + "grad_norm": 0.36670100634199615, + "learning_rate": 5.194145034896863e-07, + "loss": 0.4874, + "step": 4740 + }, + { + "epoch": 0.8686331989739832, + "grad_norm": 0.3823601114766665, + "learning_rate": 5.179961463736055e-07, + "loss": 0.4679, + "step": 4741 + }, + { + "epoch": 0.8688164162696959, + "grad_norm": 0.3734179838391516, + "learning_rate": 5.165796226491359e-07, + "loss": 0.4739, + "step": 4742 + }, + { + "epoch": 0.8689996335654085, + "grad_norm": 0.35927806436645127, + "learning_rate": 5.15164932895717e-07, + "loss": 0.4582, + "step": 4743 + }, + { + "epoch": 0.8691828508611212, + "grad_norm": 0.45571157098485937, + "learning_rate": 5.137520776920346e-07, + "loss": 0.44, + "step": 4744 + }, + { + "epoch": 0.869366068156834, + "grad_norm": 0.34616664862473023, + "learning_rate": 5.123410576160254e-07, + "loss": 0.4436, + "step": 4745 + }, + { + "epoch": 0.8695492854525467, + "grad_norm": 0.33715339955423174, + "learning_rate": 5.109318732448771e-07, + "loss": 0.4434, + "step": 4746 + }, + { + "epoch": 0.8697325027482594, + "grad_norm": 0.3659002972241651, + "learning_rate": 5.095245251550257e-07, + "loss": 0.4789, + "step": 4747 + }, + { + "epoch": 0.8699157200439721, + "grad_norm": 0.3378834071104717, + "learning_rate": 5.081190139221553e-07, + "loss": 0.4584, + "step": 4748 + }, + { + "epoch": 0.8700989373396849, + "grad_norm": 0.35105083206336474, + "learning_rate": 5.067153401211982e-07, + "loss": 0.4694, + "step": 4749 + }, + { + "epoch": 0.8702821546353976, + "grad_norm": 0.32889579815820624, + "learning_rate": 5.053135043263375e-07, + "loss": 0.4152, + "step": 4750 + }, + { + "epoch": 0.8704653719311103, + "grad_norm": 0.35785872889020776, + "learning_rate": 5.039135071110007e-07, + "loss": 0.446, + "step": 4751 + }, + { + "epoch": 0.870648589226823, + "grad_norm": 0.35561119928811563, + "learning_rate": 5.025153490478679e-07, + "loss": 0.4857, + "step": 4752 + }, + { + "epoch": 0.8708318065225358, + "grad_norm": 0.38463153410624384, + "learning_rate": 5.011190307088626e-07, + "loss": 0.4954, + "step": 4753 + }, + { + "epoch": 0.8710150238182485, + "grad_norm": 0.4290747855238213, + "learning_rate": 4.997245526651584e-07, + "loss": 0.4842, + "step": 4754 + }, + { + "epoch": 0.8711982411139612, + "grad_norm": 0.35862541092456407, + "learning_rate": 4.983319154871741e-07, + "loss": 0.455, + "step": 4755 + }, + { + "epoch": 0.8713814584096738, + "grad_norm": 0.3739303551648784, + "learning_rate": 4.969411197445784e-07, + "loss": 0.4578, + "step": 4756 + }, + { + "epoch": 0.8715646757053865, + "grad_norm": 0.39272725857499513, + "learning_rate": 4.955521660062828e-07, + "loss": 0.4506, + "step": 4757 + }, + { + "epoch": 0.8717478930010993, + "grad_norm": 0.404472418608521, + "learning_rate": 4.941650548404497e-07, + "loss": 0.4591, + "step": 4758 + }, + { + "epoch": 0.871931110296812, + "grad_norm": 0.3499079695554252, + "learning_rate": 4.927797868144835e-07, + "loss": 0.4778, + "step": 4759 + }, + { + "epoch": 0.8721143275925247, + "grad_norm": 0.3880106466488939, + "learning_rate": 4.913963624950391e-07, + "loss": 0.4835, + "step": 4760 + }, + { + "epoch": 0.8722975448882374, + "grad_norm": 0.3475585090670171, + "learning_rate": 4.900147824480123e-07, + "loss": 0.45, + "step": 4761 + }, + { + "epoch": 0.8724807621839502, + "grad_norm": 0.32425385031972004, + "learning_rate": 4.886350472385487e-07, + "loss": 0.4272, + "step": 4762 + }, + { + "epoch": 0.8726639794796629, + "grad_norm": 0.37037229617563727, + "learning_rate": 4.87257157431037e-07, + "loss": 0.4744, + "step": 4763 + }, + { + "epoch": 0.8728471967753756, + "grad_norm": 0.3395073814887526, + "learning_rate": 4.858811135891117e-07, + "loss": 0.4647, + "step": 4764 + }, + { + "epoch": 0.8730304140710883, + "grad_norm": 0.41192305317438427, + "learning_rate": 4.845069162756521e-07, + "loss": 0.4559, + "step": 4765 + }, + { + "epoch": 0.8732136313668011, + "grad_norm": 0.4352162602022674, + "learning_rate": 4.831345660527825e-07, + "loss": 0.4538, + "step": 4766 + }, + { + "epoch": 0.8733968486625138, + "grad_norm": 0.341815126962705, + "learning_rate": 4.817640634818704e-07, + "loss": 0.4688, + "step": 4767 + }, + { + "epoch": 0.8735800659582265, + "grad_norm": 0.3299801608239833, + "learning_rate": 4.803954091235297e-07, + "loss": 0.4627, + "step": 4768 + }, + { + "epoch": 0.8737632832539391, + "grad_norm": 0.3298055837117004, + "learning_rate": 4.790286035376168e-07, + "loss": 0.4226, + "step": 4769 + }, + { + "epoch": 0.8739465005496518, + "grad_norm": 0.3727028786005208, + "learning_rate": 4.776636472832302e-07, + "loss": 0.4532, + "step": 4770 + }, + { + "epoch": 0.8741297178453646, + "grad_norm": 0.30744926824515184, + "learning_rate": 4.763005409187155e-07, + "loss": 0.4518, + "step": 4771 + }, + { + "epoch": 0.8743129351410773, + "grad_norm": 0.39273173872581413, + "learning_rate": 4.7493928500165986e-07, + "loss": 0.4477, + "step": 4772 + }, + { + "epoch": 0.87449615243679, + "grad_norm": 0.376326794866468, + "learning_rate": 4.7357988008889354e-07, + "loss": 0.4551, + "step": 4773 + }, + { + "epoch": 0.8746793697325027, + "grad_norm": 0.37425009838690476, + "learning_rate": 4.7222232673648837e-07, + "loss": 0.4453, + "step": 4774 + }, + { + "epoch": 0.8748625870282155, + "grad_norm": 0.3727087090108478, + "learning_rate": 4.7086662549976015e-07, + "loss": 0.4289, + "step": 4775 + }, + { + "epoch": 0.8750458043239282, + "grad_norm": 0.4208165594640674, + "learning_rate": 4.6951277693326914e-07, + "loss": 0.4643, + "step": 4776 + }, + { + "epoch": 0.8752290216196409, + "grad_norm": 0.3442024412651745, + "learning_rate": 4.68160781590814e-07, + "loss": 0.4607, + "step": 4777 + }, + { + "epoch": 0.8754122389153536, + "grad_norm": 0.3596386760063062, + "learning_rate": 4.6681064002543574e-07, + "loss": 0.4778, + "step": 4778 + }, + { + "epoch": 0.8755954562110663, + "grad_norm": 0.31720810896246016, + "learning_rate": 4.654623527894192e-07, + "loss": 0.4657, + "step": 4779 + }, + { + "epoch": 0.8757786735067791, + "grad_norm": 0.4030061297949739, + "learning_rate": 4.6411592043429044e-07, + "loss": 0.4663, + "step": 4780 + }, + { + "epoch": 0.8759618908024918, + "grad_norm": 0.3769604502722791, + "learning_rate": 4.6277134351081556e-07, + "loss": 0.4516, + "step": 4781 + }, + { + "epoch": 0.8761451080982045, + "grad_norm": 0.3700845331491086, + "learning_rate": 4.614286225690012e-07, + "loss": 0.4583, + "step": 4782 + }, + { + "epoch": 0.8763283253939171, + "grad_norm": 0.39098536642671033, + "learning_rate": 4.600877581580959e-07, + "loss": 0.4757, + "step": 4783 + }, + { + "epoch": 0.8765115426896299, + "grad_norm": 0.41127924005263067, + "learning_rate": 4.587487508265914e-07, + "loss": 0.486, + "step": 4784 + }, + { + "epoch": 0.8766947599853426, + "grad_norm": 0.38323923045026226, + "learning_rate": 4.5741160112221293e-07, + "loss": 0.4605, + "step": 4785 + }, + { + "epoch": 0.8768779772810553, + "grad_norm": 0.3616073664821172, + "learning_rate": 4.560763095919313e-07, + "loss": 0.4524, + "step": 4786 + }, + { + "epoch": 0.877061194576768, + "grad_norm": 0.34056012201594105, + "learning_rate": 4.5474287678195785e-07, + "loss": 0.4148, + "step": 4787 + }, + { + "epoch": 0.8772444118724808, + "grad_norm": 0.4365847584117727, + "learning_rate": 4.5341130323773906e-07, + "loss": 0.4771, + "step": 4788 + }, + { + "epoch": 0.8774276291681935, + "grad_norm": 0.34944361429647225, + "learning_rate": 4.5208158950396375e-07, + "loss": 0.4624, + "step": 4789 + }, + { + "epoch": 0.8776108464639062, + "grad_norm": 0.3317899355072112, + "learning_rate": 4.507537361245606e-07, + "loss": 0.4715, + "step": 4790 + }, + { + "epoch": 0.8777940637596189, + "grad_norm": 0.3358288851963391, + "learning_rate": 4.4942774364269635e-07, + "loss": 0.4442, + "step": 4791 + }, + { + "epoch": 0.8779772810553316, + "grad_norm": 0.40184100533564304, + "learning_rate": 4.4810361260077593e-07, + "loss": 0.4594, + "step": 4792 + }, + { + "epoch": 0.8781604983510444, + "grad_norm": 0.4193781627758221, + "learning_rate": 4.467813435404428e-07, + "loss": 0.4505, + "step": 4793 + }, + { + "epoch": 0.8783437156467571, + "grad_norm": 0.3483074915921811, + "learning_rate": 4.454609370025798e-07, + "loss": 0.487, + "step": 4794 + }, + { + "epoch": 0.8785269329424698, + "grad_norm": 0.36982252646055697, + "learning_rate": 4.441423935273087e-07, + "loss": 0.4916, + "step": 4795 + }, + { + "epoch": 0.8787101502381824, + "grad_norm": 0.3448570936730333, + "learning_rate": 4.4282571365398706e-07, + "loss": 0.4776, + "step": 4796 + }, + { + "epoch": 0.8788933675338952, + "grad_norm": 0.3860983784082014, + "learning_rate": 4.415108979212096e-07, + "loss": 0.4641, + "step": 4797 + }, + { + "epoch": 0.8790765848296079, + "grad_norm": 0.36867590724818655, + "learning_rate": 4.4019794686681106e-07, + "loss": 0.5113, + "step": 4798 + }, + { + "epoch": 0.8792598021253206, + "grad_norm": 0.34675623543418466, + "learning_rate": 4.388868610278635e-07, + "loss": 0.4239, + "step": 4799 + }, + { + "epoch": 0.8794430194210333, + "grad_norm": 0.4152580056946591, + "learning_rate": 4.3757764094067344e-07, + "loss": 0.4611, + "step": 4800 + }, + { + "epoch": 0.8796262367167461, + "grad_norm": 0.3946927256297955, + "learning_rate": 4.362702871407848e-07, + "loss": 0.5147, + "step": 4801 + }, + { + "epoch": 0.8798094540124588, + "grad_norm": 0.3343148710014495, + "learning_rate": 4.349648001629797e-07, + "loss": 0.4637, + "step": 4802 + }, + { + "epoch": 0.8799926713081715, + "grad_norm": 0.3418012239660755, + "learning_rate": 4.336611805412766e-07, + "loss": 0.4822, + "step": 4803 + }, + { + "epoch": 0.8801758886038842, + "grad_norm": 0.34128843322141145, + "learning_rate": 4.3235942880892835e-07, + "loss": 0.4362, + "step": 4804 + }, + { + "epoch": 0.8803591058995969, + "grad_norm": 0.3672763225631914, + "learning_rate": 4.31059545498424e-07, + "loss": 0.4584, + "step": 4805 + }, + { + "epoch": 0.8805423231953097, + "grad_norm": 0.37966058347067694, + "learning_rate": 4.2976153114148987e-07, + "loss": 0.4717, + "step": 4806 + }, + { + "epoch": 0.8807255404910224, + "grad_norm": 0.35942734596263926, + "learning_rate": 4.2846538626908576e-07, + "loss": 0.4757, + "step": 4807 + }, + { + "epoch": 0.880908757786735, + "grad_norm": 0.3336729142478857, + "learning_rate": 4.271711114114091e-07, + "loss": 0.4689, + "step": 4808 + }, + { + "epoch": 0.8810919750824477, + "grad_norm": 0.36969627499792773, + "learning_rate": 4.258787070978898e-07, + "loss": 0.4403, + "step": 4809 + }, + { + "epoch": 0.8812751923781605, + "grad_norm": 0.34600574780847104, + "learning_rate": 4.2458817385719555e-07, + "loss": 0.418, + "step": 4810 + }, + { + "epoch": 0.8814584096738732, + "grad_norm": 0.3660457660928218, + "learning_rate": 4.232995122172245e-07, + "loss": 0.4557, + "step": 4811 + }, + { + "epoch": 0.8816416269695859, + "grad_norm": 0.32264603544309023, + "learning_rate": 4.2201272270511515e-07, + "loss": 0.4592, + "step": 4812 + }, + { + "epoch": 0.8818248442652986, + "grad_norm": 0.3404768602369994, + "learning_rate": 4.207278058472336e-07, + "loss": 0.4343, + "step": 4813 + }, + { + "epoch": 0.8820080615610114, + "grad_norm": 0.35760723524834126, + "learning_rate": 4.1944476216918495e-07, + "loss": 0.4982, + "step": 4814 + }, + { + "epoch": 0.8821912788567241, + "grad_norm": 0.32694545635571576, + "learning_rate": 4.1816359219580494e-07, + "loss": 0.4745, + "step": 4815 + }, + { + "epoch": 0.8823744961524368, + "grad_norm": 0.3903420507202521, + "learning_rate": 4.16884296451166e-07, + "loss": 0.4427, + "step": 4816 + }, + { + "epoch": 0.8825577134481495, + "grad_norm": 0.34731884542459673, + "learning_rate": 4.156068754585696e-07, + "loss": 0.4534, + "step": 4817 + }, + { + "epoch": 0.8827409307438622, + "grad_norm": 0.39836088103002565, + "learning_rate": 4.143313297405549e-07, + "loss": 0.4526, + "step": 4818 + }, + { + "epoch": 0.882924148039575, + "grad_norm": 0.47161866526153085, + "learning_rate": 4.130576598188907e-07, + "loss": 0.4695, + "step": 4819 + }, + { + "epoch": 0.8831073653352877, + "grad_norm": 0.38945841046337376, + "learning_rate": 4.117858662145802e-07, + "loss": 0.4636, + "step": 4820 + }, + { + "epoch": 0.8832905826310004, + "grad_norm": 0.37045955392899627, + "learning_rate": 4.105159494478583e-07, + "loss": 0.4804, + "step": 4821 + }, + { + "epoch": 0.883473799926713, + "grad_norm": 0.34619149838705326, + "learning_rate": 4.092479100381913e-07, + "loss": 0.4839, + "step": 4822 + }, + { + "epoch": 0.8836570172224258, + "grad_norm": 0.32676624083960315, + "learning_rate": 4.0798174850427976e-07, + "loss": 0.443, + "step": 4823 + }, + { + "epoch": 0.8838402345181385, + "grad_norm": 0.3862498656274975, + "learning_rate": 4.0671746536405544e-07, + "loss": 0.4735, + "step": 4824 + }, + { + "epoch": 0.8840234518138512, + "grad_norm": 0.30920901806133055, + "learning_rate": 4.054550611346808e-07, + "loss": 0.4607, + "step": 4825 + }, + { + "epoch": 0.8842066691095639, + "grad_norm": 0.41394152746300106, + "learning_rate": 4.041945363325489e-07, + "loss": 0.4585, + "step": 4826 + }, + { + "epoch": 0.8843898864052766, + "grad_norm": 0.3205896143381676, + "learning_rate": 4.029358914732862e-07, + "loss": 0.4576, + "step": 4827 + }, + { + "epoch": 0.8845731037009894, + "grad_norm": 0.37803624220470905, + "learning_rate": 4.0167912707175014e-07, + "loss": 0.475, + "step": 4828 + }, + { + "epoch": 0.8847563209967021, + "grad_norm": 0.35494836287627074, + "learning_rate": 4.004242436420269e-07, + "loss": 0.4249, + "step": 4829 + }, + { + "epoch": 0.8849395382924148, + "grad_norm": 0.34571671298659484, + "learning_rate": 3.991712416974347e-07, + "loss": 0.4918, + "step": 4830 + }, + { + "epoch": 0.8851227555881275, + "grad_norm": 0.342585336743226, + "learning_rate": 3.9792012175052076e-07, + "loss": 0.4745, + "step": 4831 + }, + { + "epoch": 0.8853059728838403, + "grad_norm": 0.4208362433264992, + "learning_rate": 3.9667088431306623e-07, + "loss": 0.4976, + "step": 4832 + }, + { + "epoch": 0.885489190179553, + "grad_norm": 0.3408535994427791, + "learning_rate": 3.9542352989607734e-07, + "loss": 0.4664, + "step": 4833 + }, + { + "epoch": 0.8856724074752657, + "grad_norm": 0.3637487104821648, + "learning_rate": 3.9417805900979154e-07, + "loss": 0.4873, + "step": 4834 + }, + { + "epoch": 0.8858556247709783, + "grad_norm": 0.3172577709939467, + "learning_rate": 3.929344721636774e-07, + "loss": 0.4319, + "step": 4835 + }, + { + "epoch": 0.8860388420666911, + "grad_norm": 0.35731796191973786, + "learning_rate": 3.916927698664341e-07, + "loss": 0.4579, + "step": 4836 + }, + { + "epoch": 0.8862220593624038, + "grad_norm": 0.31638865716709486, + "learning_rate": 3.904529526259837e-07, + "loss": 0.4413, + "step": 4837 + }, + { + "epoch": 0.8864052766581165, + "grad_norm": 0.3752269942491934, + "learning_rate": 3.8921502094948325e-07, + "loss": 0.4428, + "step": 4838 + }, + { + "epoch": 0.8865884939538292, + "grad_norm": 0.388634970734894, + "learning_rate": 3.8797897534331553e-07, + "loss": 0.4989, + "step": 4839 + }, + { + "epoch": 0.8867717112495419, + "grad_norm": 0.35430989732140517, + "learning_rate": 3.86744816313096e-07, + "loss": 0.4841, + "step": 4840 + }, + { + "epoch": 0.8869549285452547, + "grad_norm": 0.4037787319233198, + "learning_rate": 3.855125443636598e-07, + "loss": 0.4591, + "step": 4841 + }, + { + "epoch": 0.8871381458409674, + "grad_norm": 0.34530954560023164, + "learning_rate": 3.842821599990787e-07, + "loss": 0.4478, + "step": 4842 + }, + { + "epoch": 0.8873213631366801, + "grad_norm": 0.3822611567187025, + "learning_rate": 3.830536637226495e-07, + "loss": 0.4913, + "step": 4843 + }, + { + "epoch": 0.8875045804323928, + "grad_norm": 0.3790966864214009, + "learning_rate": 3.818270560368953e-07, + "loss": 0.4503, + "step": 4844 + }, + { + "epoch": 0.8876877977281056, + "grad_norm": 0.3609214112353151, + "learning_rate": 3.8060233744356634e-07, + "loss": 0.4526, + "step": 4845 + }, + { + "epoch": 0.8878710150238183, + "grad_norm": 0.38962620466211084, + "learning_rate": 3.79379508443643e-07, + "loss": 0.5017, + "step": 4846 + }, + { + "epoch": 0.888054232319531, + "grad_norm": 0.3236292022047194, + "learning_rate": 3.7815856953733133e-07, + "loss": 0.4847, + "step": 4847 + }, + { + "epoch": 0.8882374496152436, + "grad_norm": 0.3454054346717301, + "learning_rate": 3.76939521224064e-07, + "loss": 0.4729, + "step": 4848 + }, + { + "epoch": 0.8884206669109564, + "grad_norm": 0.39094294746186065, + "learning_rate": 3.757223640024982e-07, + "loss": 0.4545, + "step": 4849 + }, + { + "epoch": 0.8886038842066691, + "grad_norm": 0.37443688886913046, + "learning_rate": 3.745070983705218e-07, + "loss": 0.4944, + "step": 4850 + }, + { + "epoch": 0.8887871015023818, + "grad_norm": 0.37988411345917805, + "learning_rate": 3.732937248252472e-07, + "loss": 0.4511, + "step": 4851 + }, + { + "epoch": 0.8889703187980945, + "grad_norm": 0.36261796976127203, + "learning_rate": 3.720822438630106e-07, + "loss": 0.4533, + "step": 4852 + }, + { + "epoch": 0.8891535360938072, + "grad_norm": 0.42303113105179446, + "learning_rate": 3.7087265597937627e-07, + "loss": 0.4978, + "step": 4853 + }, + { + "epoch": 0.88933675338952, + "grad_norm": 0.3741272328204488, + "learning_rate": 3.696649616691339e-07, + "loss": 0.4654, + "step": 4854 + }, + { + "epoch": 0.8895199706852327, + "grad_norm": 0.4172627358691279, + "learning_rate": 3.684591614263e-07, + "loss": 0.472, + "step": 4855 + }, + { + "epoch": 0.8897031879809454, + "grad_norm": 0.42319286052058397, + "learning_rate": 3.6725525574411226e-07, + "loss": 0.4687, + "step": 4856 + }, + { + "epoch": 0.8898864052766581, + "grad_norm": 0.4723309908271977, + "learning_rate": 3.6605324511503617e-07, + "loss": 0.4969, + "step": 4857 + }, + { + "epoch": 0.8900696225723709, + "grad_norm": 0.3752229256717594, + "learning_rate": 3.648531300307634e-07, + "loss": 0.4568, + "step": 4858 + }, + { + "epoch": 0.8902528398680836, + "grad_norm": 0.38091634747093434, + "learning_rate": 3.6365491098220683e-07, + "loss": 0.4716, + "step": 4859 + }, + { + "epoch": 0.8904360571637963, + "grad_norm": 0.36029154867058755, + "learning_rate": 3.6245858845950557e-07, + "loss": 0.4827, + "step": 4860 + }, + { + "epoch": 0.890619274459509, + "grad_norm": 0.37161029931977824, + "learning_rate": 3.612641629520236e-07, + "loss": 0.4775, + "step": 4861 + }, + { + "epoch": 0.8908024917552217, + "grad_norm": 0.33321589030345006, + "learning_rate": 3.6007163494834795e-07, + "loss": 0.4689, + "step": 4862 + }, + { + "epoch": 0.8909857090509344, + "grad_norm": 0.348867845087718, + "learning_rate": 3.588810049362895e-07, + "loss": 0.4192, + "step": 4863 + }, + { + "epoch": 0.8911689263466471, + "grad_norm": 0.3555745602487136, + "learning_rate": 3.57692273402882e-07, + "loss": 0.4967, + "step": 4864 + }, + { + "epoch": 0.8913521436423598, + "grad_norm": 0.34640124866589594, + "learning_rate": 3.5650544083438435e-07, + "loss": 0.495, + "step": 4865 + }, + { + "epoch": 0.8915353609380725, + "grad_norm": 0.33479455916636325, + "learning_rate": 3.5532050771627856e-07, + "loss": 0.4517, + "step": 4866 + }, + { + "epoch": 0.8917185782337853, + "grad_norm": 0.3609802730979179, + "learning_rate": 3.5413747453326766e-07, + "loss": 0.4567, + "step": 4867 + }, + { + "epoch": 0.891901795529498, + "grad_norm": 0.37200960905168673, + "learning_rate": 3.5295634176927884e-07, + "loss": 0.4518, + "step": 4868 + }, + { + "epoch": 0.8920850128252107, + "grad_norm": 0.40911098686063996, + "learning_rate": 3.5177710990746185e-07, + "loss": 0.4813, + "step": 4869 + }, + { + "epoch": 0.8922682301209234, + "grad_norm": 0.3475862174413965, + "learning_rate": 3.505997794301902e-07, + "loss": 0.4568, + "step": 4870 + }, + { + "epoch": 0.8924514474166362, + "grad_norm": 0.3211170793371448, + "learning_rate": 3.49424350819057e-07, + "loss": 0.4648, + "step": 4871 + }, + { + "epoch": 0.8926346647123489, + "grad_norm": 0.3820057704565937, + "learning_rate": 3.4825082455487766e-07, + "loss": 0.4445, + "step": 4872 + }, + { + "epoch": 0.8928178820080616, + "grad_norm": 0.34484017095374153, + "learning_rate": 3.4707920111769324e-07, + "loss": 0.4573, + "step": 4873 + }, + { + "epoch": 0.8930010993037742, + "grad_norm": 0.49589294733887046, + "learning_rate": 3.4590948098676027e-07, + "loss": 0.4609, + "step": 4874 + }, + { + "epoch": 0.8931843165994869, + "grad_norm": 0.3318680679469057, + "learning_rate": 3.4474166464056327e-07, + "loss": 0.4461, + "step": 4875 + }, + { + "epoch": 0.8933675338951997, + "grad_norm": 0.4345104517070929, + "learning_rate": 3.435757525568023e-07, + "loss": 0.4741, + "step": 4876 + }, + { + "epoch": 0.8935507511909124, + "grad_norm": 0.37560256167506345, + "learning_rate": 3.4241174521240207e-07, + "loss": 0.4817, + "step": 4877 + }, + { + "epoch": 0.8937339684866251, + "grad_norm": 0.3429361436041367, + "learning_rate": 3.412496430835066e-07, + "loss": 0.4628, + "step": 4878 + }, + { + "epoch": 0.8939171857823378, + "grad_norm": 0.35920723682658684, + "learning_rate": 3.400894466454824e-07, + "loss": 0.4946, + "step": 4879 + }, + { + "epoch": 0.8941004030780506, + "grad_norm": 0.34913783744258275, + "learning_rate": 3.389311563729131e-07, + "loss": 0.4584, + "step": 4880 + }, + { + "epoch": 0.8942836203737633, + "grad_norm": 0.3607379716234879, + "learning_rate": 3.377747727396058e-07, + "loss": 0.4581, + "step": 4881 + }, + { + "epoch": 0.894466837669476, + "grad_norm": 0.3249396315401186, + "learning_rate": 3.366202962185855e-07, + "loss": 0.4227, + "step": 4882 + }, + { + "epoch": 0.8946500549651887, + "grad_norm": 0.9585570157817012, + "learning_rate": 3.3546772728209944e-07, + "loss": 0.4702, + "step": 4883 + }, + { + "epoch": 0.8948332722609015, + "grad_norm": 0.33438126766914245, + "learning_rate": 3.34317066401611e-07, + "loss": 0.461, + "step": 4884 + }, + { + "epoch": 0.8950164895566142, + "grad_norm": 0.3824373992314347, + "learning_rate": 3.331683140478076e-07, + "loss": 0.4824, + "step": 4885 + }, + { + "epoch": 0.8951997068523269, + "grad_norm": 0.40915051499084326, + "learning_rate": 3.3202147069059164e-07, + "loss": 0.4949, + "step": 4886 + }, + { + "epoch": 0.8953829241480395, + "grad_norm": 0.41656330798473534, + "learning_rate": 3.308765367990879e-07, + "loss": 0.4726, + "step": 4887 + }, + { + "epoch": 0.8955661414437522, + "grad_norm": 0.34810133904135593, + "learning_rate": 3.2973351284163847e-07, + "loss": 0.4557, + "step": 4888 + }, + { + "epoch": 0.895749358739465, + "grad_norm": 0.36877548663436777, + "learning_rate": 3.2859239928580313e-07, + "loss": 0.4991, + "step": 4889 + }, + { + "epoch": 0.8959325760351777, + "grad_norm": 0.37909584725872897, + "learning_rate": 3.274531965983624e-07, + "loss": 0.4862, + "step": 4890 + }, + { + "epoch": 0.8961157933308904, + "grad_norm": 0.37925915099093715, + "learning_rate": 3.2631590524531466e-07, + "loss": 0.4633, + "step": 4891 + }, + { + "epoch": 0.8962990106266031, + "grad_norm": 0.35308293184942124, + "learning_rate": 3.251805256918761e-07, + "loss": 0.4647, + "step": 4892 + }, + { + "epoch": 0.8964822279223159, + "grad_norm": 0.36542400624489646, + "learning_rate": 3.2404705840247963e-07, + "loss": 0.4623, + "step": 4893 + }, + { + "epoch": 0.8966654452180286, + "grad_norm": 0.4378360510171086, + "learning_rate": 3.229155038407772e-07, + "loss": 0.4748, + "step": 4894 + }, + { + "epoch": 0.8968486625137413, + "grad_norm": 0.3346352173394287, + "learning_rate": 3.2178586246964017e-07, + "loss": 0.4801, + "step": 4895 + }, + { + "epoch": 0.897031879809454, + "grad_norm": 0.351346282405727, + "learning_rate": 3.206581347511539e-07, + "loss": 0.5004, + "step": 4896 + }, + { + "epoch": 0.8972150971051668, + "grad_norm": 0.45421522016972793, + "learning_rate": 3.1953232114662226e-07, + "loss": 0.4802, + "step": 4897 + }, + { + "epoch": 0.8973983144008795, + "grad_norm": 0.34716498956767444, + "learning_rate": 3.184084221165662e-07, + "loss": 0.4529, + "step": 4898 + }, + { + "epoch": 0.8975815316965922, + "grad_norm": 0.4260128703732617, + "learning_rate": 3.172864381207252e-07, + "loss": 0.4593, + "step": 4899 + }, + { + "epoch": 0.8977647489923049, + "grad_norm": 0.2939127973558809, + "learning_rate": 3.161663696180528e-07, + "loss": 0.4332, + "step": 4900 + }, + { + "epoch": 0.8979479662880175, + "grad_norm": 0.35356862311039927, + "learning_rate": 3.150482170667191e-07, + "loss": 0.4814, + "step": 4901 + }, + { + "epoch": 0.8981311835837303, + "grad_norm": 0.38680759640757195, + "learning_rate": 3.1393198092411203e-07, + "loss": 0.4607, + "step": 4902 + }, + { + "epoch": 0.898314400879443, + "grad_norm": 0.3868502737073052, + "learning_rate": 3.128176616468359e-07, + "loss": 0.4681, + "step": 4903 + }, + { + "epoch": 0.8984976181751557, + "grad_norm": 0.4067157939025431, + "learning_rate": 3.1170525969070887e-07, + "loss": 0.4874, + "step": 4904 + }, + { + "epoch": 0.8986808354708684, + "grad_norm": 0.38211113949419306, + "learning_rate": 3.105947755107658e-07, + "loss": 0.4815, + "step": 4905 + }, + { + "epoch": 0.8988640527665812, + "grad_norm": 0.35304839855864534, + "learning_rate": 3.0948620956125776e-07, + "loss": 0.4927, + "step": 4906 + }, + { + "epoch": 0.8990472700622939, + "grad_norm": 0.341979957021192, + "learning_rate": 3.0837956229565146e-07, + "loss": 0.4392, + "step": 4907 + }, + { + "epoch": 0.8992304873580066, + "grad_norm": 0.4065961587687938, + "learning_rate": 3.072748341666254e-07, + "loss": 0.4509, + "step": 4908 + }, + { + "epoch": 0.8994137046537193, + "grad_norm": 0.3828604497041965, + "learning_rate": 3.0617202562607697e-07, + "loss": 0.413, + "step": 4909 + }, + { + "epoch": 0.899596921949432, + "grad_norm": 0.34332134194985975, + "learning_rate": 3.0507113712511747e-07, + "loss": 0.4279, + "step": 4910 + }, + { + "epoch": 0.8997801392451448, + "grad_norm": 0.36002957873086067, + "learning_rate": 3.039721691140712e-07, + "loss": 0.4839, + "step": 4911 + }, + { + "epoch": 0.8999633565408575, + "grad_norm": 0.33542914499568, + "learning_rate": 3.02875122042478e-07, + "loss": 0.4206, + "step": 4912 + }, + { + "epoch": 0.9001465738365702, + "grad_norm": 0.3381133626689856, + "learning_rate": 3.017799963590917e-07, + "loss": 0.4822, + "step": 4913 + }, + { + "epoch": 0.9003297911322828, + "grad_norm": 0.5578966824983292, + "learning_rate": 3.0068679251188116e-07, + "loss": 0.4699, + "step": 4914 + }, + { + "epoch": 0.9005130084279956, + "grad_norm": 0.35496332159524224, + "learning_rate": 2.995955109480275e-07, + "loss": 0.4328, + "step": 4915 + }, + { + "epoch": 0.9006962257237083, + "grad_norm": 0.3339523811709404, + "learning_rate": 2.98506152113926e-07, + "loss": 0.4461, + "step": 4916 + }, + { + "epoch": 0.900879443019421, + "grad_norm": 0.35217428631241154, + "learning_rate": 2.9741871645518616e-07, + "loss": 0.4356, + "step": 4917 + }, + { + "epoch": 0.9010626603151337, + "grad_norm": 0.38091196373635966, + "learning_rate": 2.963332044166306e-07, + "loss": 0.467, + "step": 4918 + }, + { + "epoch": 0.9012458776108465, + "grad_norm": 0.3272784506047193, + "learning_rate": 2.9524961644229465e-07, + "loss": 0.4737, + "step": 4919 + }, + { + "epoch": 0.9014290949065592, + "grad_norm": 0.3403394173577311, + "learning_rate": 2.94167952975426e-07, + "loss": 0.4448, + "step": 4920 + }, + { + "epoch": 0.9016123122022719, + "grad_norm": 0.7186807488432906, + "learning_rate": 2.9308821445848567e-07, + "loss": 0.4616, + "step": 4921 + }, + { + "epoch": 0.9017955294979846, + "grad_norm": 0.3483530685369334, + "learning_rate": 2.920104013331493e-07, + "loss": 0.4645, + "step": 4922 + }, + { + "epoch": 0.9019787467936973, + "grad_norm": 0.3081393762018205, + "learning_rate": 2.909345140403019e-07, + "loss": 0.473, + "step": 4923 + }, + { + "epoch": 0.9021619640894101, + "grad_norm": 0.4025947080240943, + "learning_rate": 2.8986055302004045e-07, + "loss": 0.4756, + "step": 4924 + }, + { + "epoch": 0.9023451813851228, + "grad_norm": 0.40263858263747304, + "learning_rate": 2.887885187116779e-07, + "loss": 0.4814, + "step": 4925 + }, + { + "epoch": 0.9025283986808355, + "grad_norm": 0.3310030739455688, + "learning_rate": 2.8771841155373513e-07, + "loss": 0.474, + "step": 4926 + }, + { + "epoch": 0.9027116159765481, + "grad_norm": 0.3354060061636858, + "learning_rate": 2.866502319839465e-07, + "loss": 0.4517, + "step": 4927 + }, + { + "epoch": 0.9028948332722609, + "grad_norm": 0.372528256679799, + "learning_rate": 2.855839804392568e-07, + "loss": 0.5142, + "step": 4928 + }, + { + "epoch": 0.9030780505679736, + "grad_norm": 0.3365831382289672, + "learning_rate": 2.8451965735582455e-07, + "loss": 0.4323, + "step": 4929 + }, + { + "epoch": 0.9032612678636863, + "grad_norm": 0.3385594222248257, + "learning_rate": 2.834572631690158e-07, + "loss": 0.4777, + "step": 4930 + }, + { + "epoch": 0.903444485159399, + "grad_norm": 0.32658799669276706, + "learning_rate": 2.8239679831341126e-07, + "loss": 0.4596, + "step": 4931 + }, + { + "epoch": 0.9036277024551118, + "grad_norm": 0.36702488522666377, + "learning_rate": 2.813382632227996e-07, + "loss": 0.4504, + "step": 4932 + }, + { + "epoch": 0.9038109197508245, + "grad_norm": 0.37208252431065797, + "learning_rate": 2.802816583301815e-07, + "loss": 0.4906, + "step": 4933 + }, + { + "epoch": 0.9039941370465372, + "grad_norm": 0.3739481848320804, + "learning_rate": 2.792269840677675e-07, + "loss": 0.4695, + "step": 4934 + }, + { + "epoch": 0.9041773543422499, + "grad_norm": 0.3630361210797529, + "learning_rate": 2.781742408669791e-07, + "loss": 0.4712, + "step": 4935 + }, + { + "epoch": 0.9043605716379626, + "grad_norm": 0.3779355465449995, + "learning_rate": 2.771234291584468e-07, + "loss": 0.4801, + "step": 4936 + }, + { + "epoch": 0.9045437889336754, + "grad_norm": 0.4409045039425335, + "learning_rate": 2.7607454937201295e-07, + "loss": 0.4846, + "step": 4937 + }, + { + "epoch": 0.9047270062293881, + "grad_norm": 0.3971717083693932, + "learning_rate": 2.7502760193672704e-07, + "loss": 0.49, + "step": 4938 + }, + { + "epoch": 0.9049102235251008, + "grad_norm": 0.36233330924393153, + "learning_rate": 2.739825872808505e-07, + "loss": 0.4922, + "step": 4939 + }, + { + "epoch": 0.9050934408208134, + "grad_norm": 0.35326440600726605, + "learning_rate": 2.729395058318518e-07, + "loss": 0.4681, + "step": 4940 + }, + { + "epoch": 0.9052766581165262, + "grad_norm": 0.3742666329408186, + "learning_rate": 2.718983580164114e-07, + "loss": 0.4246, + "step": 4941 + }, + { + "epoch": 0.9054598754122389, + "grad_norm": 0.31951665385306743, + "learning_rate": 2.7085914426041583e-07, + "loss": 0.4509, + "step": 4942 + }, + { + "epoch": 0.9056430927079516, + "grad_norm": 0.33975152526881497, + "learning_rate": 2.698218649889639e-07, + "loss": 0.4769, + "step": 4943 + }, + { + "epoch": 0.9058263100036643, + "grad_norm": 0.41742124442249395, + "learning_rate": 2.6878652062635957e-07, + "loss": 0.5013, + "step": 4944 + }, + { + "epoch": 0.9060095272993771, + "grad_norm": 0.33213220542154853, + "learning_rate": 2.677531115961163e-07, + "loss": 0.4374, + "step": 4945 + }, + { + "epoch": 0.9061927445950898, + "grad_norm": 0.3337974421764124, + "learning_rate": 2.667216383209581e-07, + "loss": 0.4504, + "step": 4946 + }, + { + "epoch": 0.9063759618908025, + "grad_norm": 0.3607510563290672, + "learning_rate": 2.656921012228153e-07, + "loss": 0.4442, + "step": 4947 + }, + { + "epoch": 0.9065591791865152, + "grad_norm": 0.37281076534193025, + "learning_rate": 2.64664500722826e-07, + "loss": 0.4624, + "step": 4948 + }, + { + "epoch": 0.9067423964822279, + "grad_norm": 0.3352794538752936, + "learning_rate": 2.636388372413362e-07, + "loss": 0.4724, + "step": 4949 + }, + { + "epoch": 0.9069256137779407, + "grad_norm": 0.33618199210104194, + "learning_rate": 2.626151111979003e-07, + "loss": 0.4389, + "step": 4950 + }, + { + "epoch": 0.9071088310736534, + "grad_norm": 0.3108497882664409, + "learning_rate": 2.6159332301128104e-07, + "loss": 0.4339, + "step": 4951 + }, + { + "epoch": 0.907292048369366, + "grad_norm": 0.33961572104286003, + "learning_rate": 2.605734730994458e-07, + "loss": 0.444, + "step": 4952 + }, + { + "epoch": 0.9074752656650787, + "grad_norm": 0.388456727312225, + "learning_rate": 2.595555618795703e-07, + "loss": 0.4256, + "step": 4953 + }, + { + "epoch": 0.9076584829607915, + "grad_norm": 0.35528741783968276, + "learning_rate": 2.585395897680376e-07, + "loss": 0.4481, + "step": 4954 + }, + { + "epoch": 0.9078417002565042, + "grad_norm": 0.35227484068282305, + "learning_rate": 2.575255571804391e-07, + "loss": 0.4624, + "step": 4955 + }, + { + "epoch": 0.9080249175522169, + "grad_norm": 0.3434351345497852, + "learning_rate": 2.5651346453157033e-07, + "loss": 0.4679, + "step": 4956 + }, + { + "epoch": 0.9082081348479296, + "grad_norm": 0.35506322987234856, + "learning_rate": 2.5550331223543225e-07, + "loss": 0.516, + "step": 4957 + }, + { + "epoch": 0.9083913521436423, + "grad_norm": 0.3657441811308741, + "learning_rate": 2.54495100705236e-07, + "loss": 0.4701, + "step": 4958 + }, + { + "epoch": 0.9085745694393551, + "grad_norm": 0.38263732785390736, + "learning_rate": 2.534888303533972e-07, + "loss": 0.5025, + "step": 4959 + }, + { + "epoch": 0.9087577867350678, + "grad_norm": 0.3791519380336447, + "learning_rate": 2.524845015915345e-07, + "loss": 0.4607, + "step": 4960 + }, + { + "epoch": 0.9089410040307805, + "grad_norm": 0.3591365364403032, + "learning_rate": 2.514821148304769e-07, + "loss": 0.4691, + "step": 4961 + }, + { + "epoch": 0.9091242213264932, + "grad_norm": 0.3250739846769537, + "learning_rate": 2.5048167048025654e-07, + "loss": 0.4226, + "step": 4962 + }, + { + "epoch": 0.909307438622206, + "grad_norm": 0.44294552795890796, + "learning_rate": 2.49483168950112e-07, + "loss": 0.468, + "step": 4963 + }, + { + "epoch": 0.9094906559179187, + "grad_norm": 0.3798759612827671, + "learning_rate": 2.4848661064848457e-07, + "loss": 0.4775, + "step": 4964 + }, + { + "epoch": 0.9096738732136314, + "grad_norm": 0.36415582806252733, + "learning_rate": 2.4749199598302407e-07, + "loss": 0.4615, + "step": 4965 + }, + { + "epoch": 0.909857090509344, + "grad_norm": 0.33150822853754874, + "learning_rate": 2.4649932536058407e-07, + "loss": 0.4556, + "step": 4966 + }, + { + "epoch": 0.9100403078050568, + "grad_norm": 0.3571307471461616, + "learning_rate": 2.455085991872225e-07, + "loss": 0.4638, + "step": 4967 + }, + { + "epoch": 0.9102235251007695, + "grad_norm": 0.35370135640303424, + "learning_rate": 2.4451981786820125e-07, + "loss": 0.4219, + "step": 4968 + }, + { + "epoch": 0.9104067423964822, + "grad_norm": 0.3500808697242027, + "learning_rate": 2.435329818079879e-07, + "loss": 0.4508, + "step": 4969 + }, + { + "epoch": 0.9105899596921949, + "grad_norm": 0.3803030866937755, + "learning_rate": 2.4254809141025514e-07, + "loss": 0.4566, + "step": 4970 + }, + { + "epoch": 0.9107731769879076, + "grad_norm": 0.4175219114445916, + "learning_rate": 2.4156514707787683e-07, + "loss": 0.4689, + "step": 4971 + }, + { + "epoch": 0.9109563942836204, + "grad_norm": 0.3947862943272951, + "learning_rate": 2.4058414921293293e-07, + "loss": 0.4743, + "step": 4972 + }, + { + "epoch": 0.9111396115793331, + "grad_norm": 0.371315623148494, + "learning_rate": 2.39605098216707e-07, + "loss": 0.4899, + "step": 4973 + }, + { + "epoch": 0.9113228288750458, + "grad_norm": 0.3599195767751902, + "learning_rate": 2.386279944896858e-07, + "loss": 0.4335, + "step": 4974 + }, + { + "epoch": 0.9115060461707585, + "grad_norm": 0.37787364249791316, + "learning_rate": 2.3765283843156018e-07, + "loss": 0.4647, + "step": 4975 + }, + { + "epoch": 0.9116892634664713, + "grad_norm": 0.4434485262054723, + "learning_rate": 2.3667963044122278e-07, + "loss": 0.4894, + "step": 4976 + }, + { + "epoch": 0.911872480762184, + "grad_norm": 0.34575147394332395, + "learning_rate": 2.3570837091677124e-07, + "loss": 0.4444, + "step": 4977 + }, + { + "epoch": 0.9120556980578967, + "grad_norm": 0.36294758932153603, + "learning_rate": 2.3473906025550497e-07, + "loss": 0.483, + "step": 4978 + }, + { + "epoch": 0.9122389153536093, + "grad_norm": 0.3425817566606844, + "learning_rate": 2.3377169885392737e-07, + "loss": 0.4282, + "step": 4979 + }, + { + "epoch": 0.9124221326493221, + "grad_norm": 0.3671103730502629, + "learning_rate": 2.3280628710774189e-07, + "loss": 0.446, + "step": 4980 + }, + { + "epoch": 0.9126053499450348, + "grad_norm": 0.7652588185039066, + "learning_rate": 2.318428254118582e-07, + "loss": 0.4768, + "step": 4981 + }, + { + "epoch": 0.9127885672407475, + "grad_norm": 0.33741484196406774, + "learning_rate": 2.308813141603844e-07, + "loss": 0.4806, + "step": 4982 + }, + { + "epoch": 0.9129717845364602, + "grad_norm": 0.3759089142457896, + "learning_rate": 2.2992175374663417e-07, + "loss": 0.4632, + "step": 4983 + }, + { + "epoch": 0.9131550018321729, + "grad_norm": 0.3924723854820959, + "learning_rate": 2.2896414456312077e-07, + "loss": 0.4487, + "step": 4984 + }, + { + "epoch": 0.9133382191278857, + "grad_norm": 0.41602868539879684, + "learning_rate": 2.2800848700156088e-07, + "loss": 0.4736, + "step": 4985 + }, + { + "epoch": 0.9135214364235984, + "grad_norm": 0.3902806002675323, + "learning_rate": 2.2705478145287174e-07, + "loss": 0.5048, + "step": 4986 + }, + { + "epoch": 0.9137046537193111, + "grad_norm": 0.3143375670454848, + "learning_rate": 2.2610302830717302e-07, + "loss": 0.4759, + "step": 4987 + }, + { + "epoch": 0.9138878710150238, + "grad_norm": 0.3474712545379862, + "learning_rate": 2.2515322795378436e-07, + "loss": 0.4766, + "step": 4988 + }, + { + "epoch": 0.9140710883107366, + "grad_norm": 1.248372328610258, + "learning_rate": 2.2420538078122832e-07, + "loss": 0.4781, + "step": 4989 + }, + { + "epoch": 0.9142543056064493, + "grad_norm": 0.3649223569747327, + "learning_rate": 2.2325948717722756e-07, + "loss": 0.4791, + "step": 4990 + }, + { + "epoch": 0.914437522902162, + "grad_norm": 0.3475327593359362, + "learning_rate": 2.2231554752870588e-07, + "loss": 0.4567, + "step": 4991 + }, + { + "epoch": 0.9146207401978746, + "grad_norm": 0.37399469693739446, + "learning_rate": 2.2137356222178664e-07, + "loss": 0.4723, + "step": 4992 + }, + { + "epoch": 0.9148039574935874, + "grad_norm": 0.3066478829926953, + "learning_rate": 2.204335316417966e-07, + "loss": 0.4396, + "step": 4993 + }, + { + "epoch": 0.9149871747893001, + "grad_norm": 0.35458166777623096, + "learning_rate": 2.194954561732593e-07, + "loss": 0.4671, + "step": 4994 + }, + { + "epoch": 0.9151703920850128, + "grad_norm": 0.43618141764486884, + "learning_rate": 2.1855933619990167e-07, + "loss": 0.4655, + "step": 4995 + }, + { + "epoch": 0.9153536093807255, + "grad_norm": 0.3590879427768603, + "learning_rate": 2.1762517210464907e-07, + "loss": 0.4867, + "step": 4996 + }, + { + "epoch": 0.9155368266764382, + "grad_norm": 0.3741014318403361, + "learning_rate": 2.1669296426962638e-07, + "loss": 0.472, + "step": 4997 + }, + { + "epoch": 0.915720043972151, + "grad_norm": 0.3838728310327138, + "learning_rate": 2.157627130761597e-07, + "loss": 0.425, + "step": 4998 + }, + { + "epoch": 0.9159032612678637, + "grad_norm": 0.37795448261761533, + "learning_rate": 2.1483441890477464e-07, + "loss": 0.499, + "step": 4999 + }, + { + "epoch": 0.9160864785635764, + "grad_norm": 0.38619666356652016, + "learning_rate": 2.139080821351952e-07, + "loss": 0.4581, + "step": 5000 + } + ], + "logging_steps": 1, + "max_steps": 5458, + "num_input_tokens_seen": 0, + "num_train_epochs": 1, + "save_steps": 1000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 889063246725120.0, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}