Models_Ouput2 / last-checkpoint /trainer_state.json
youssefedweqd's picture
Model save
ae530ba verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.7152317880794703,
"eval_steps": 100,
"global_step": 16400,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.008278145695364239,
"grad_norm": 2.812814950942993,
"learning_rate": 2.7041942604856512e-06,
"loss": 1.4882,
"step": 50
},
{
"epoch": 0.016556291390728478,
"grad_norm": 2.6125919818878174,
"learning_rate": 5.463576158940398e-06,
"loss": 1.2862,
"step": 100
},
{
"epoch": 0.016556291390728478,
"eval_loss": 1.2020893096923828,
"eval_runtime": 1896.4557,
"eval_samples_per_second": 3.185,
"eval_steps_per_second": 3.185,
"step": 100
},
{
"epoch": 0.024834437086092714,
"grad_norm": 2.393519878387451,
"learning_rate": 8.222958057395145e-06,
"loss": 1.1353,
"step": 150
},
{
"epoch": 0.033112582781456956,
"grad_norm": 2.5604305267333984,
"learning_rate": 1.0982339955849891e-05,
"loss": 1.0539,
"step": 200
},
{
"epoch": 0.033112582781456956,
"eval_loss": 1.0201358795166016,
"eval_runtime": 1896.3026,
"eval_samples_per_second": 3.185,
"eval_steps_per_second": 3.185,
"step": 200
},
{
"epoch": 0.041390728476821195,
"grad_norm": 3.0053958892822266,
"learning_rate": 1.3741721854304637e-05,
"loss": 0.9816,
"step": 250
},
{
"epoch": 0.04966887417218543,
"grad_norm": 3.028010368347168,
"learning_rate": 1.6501103752759385e-05,
"loss": 0.9372,
"step": 300
},
{
"epoch": 0.057947019867549666,
"grad_norm": 2.484466552734375,
"learning_rate": 1.926048565121413e-05,
"loss": 0.9137,
"step": 350
},
{
"epoch": 0.06622516556291391,
"grad_norm": 3.078425884246826,
"learning_rate": 2.2019867549668874e-05,
"loss": 0.8786,
"step": 400
},
{
"epoch": 0.07450331125827815,
"grad_norm": 2.6672778129577637,
"learning_rate": 2.477924944812362e-05,
"loss": 0.8348,
"step": 450
},
{
"epoch": 0.08278145695364239,
"grad_norm": 2.8199753761291504,
"learning_rate": 2.753863134657837e-05,
"loss": 0.8417,
"step": 500
},
{
"epoch": 0.09105960264900662,
"grad_norm": 2.961965322494507,
"learning_rate": 3.0298013245033112e-05,
"loss": 0.8209,
"step": 550
},
{
"epoch": 0.09933774834437085,
"grad_norm": 2.5217514038085938,
"learning_rate": 3.305739514348786e-05,
"loss": 0.8235,
"step": 600
},
{
"epoch": 0.1076158940397351,
"grad_norm": 2.4740138053894043,
"learning_rate": 3.581677704194261e-05,
"loss": 0.7918,
"step": 650
},
{
"epoch": 0.11589403973509933,
"grad_norm": 2.5529448986053467,
"learning_rate": 3.8576158940397354e-05,
"loss": 0.7749,
"step": 700
},
{
"epoch": 0.12417218543046357,
"grad_norm": 2.151698589324951,
"learning_rate": 4.13355408388521e-05,
"loss": 0.7838,
"step": 750
},
{
"epoch": 0.13245033112582782,
"grad_norm": 1.9026315212249756,
"learning_rate": 4.4094922737306846e-05,
"loss": 0.7668,
"step": 800
},
{
"epoch": 0.14072847682119205,
"grad_norm": 1.9555529356002808,
"learning_rate": 4.685430463576159e-05,
"loss": 0.7684,
"step": 850
},
{
"epoch": 0.1490066225165563,
"grad_norm": 2.308894157409668,
"learning_rate": 4.961368653421634e-05,
"loss": 0.7661,
"step": 900
},
{
"epoch": 0.15728476821192053,
"grad_norm": 2.3362715244293213,
"learning_rate": 5.237306843267108e-05,
"loss": 0.736,
"step": 950
},
{
"epoch": 0.16556291390728478,
"grad_norm": 1.8228410482406616,
"learning_rate": 5.513245033112583e-05,
"loss": 0.7213,
"step": 1000
},
{
"epoch": 0.173841059602649,
"grad_norm": 1.8289754390716553,
"learning_rate": 5.789183222958058e-05,
"loss": 0.7335,
"step": 1050
},
{
"epoch": 0.18211920529801323,
"grad_norm": 1.4989681243896484,
"learning_rate": 6.065121412803533e-05,
"loss": 0.7326,
"step": 1100
},
{
"epoch": 0.19039735099337748,
"grad_norm": 1.5326098203659058,
"learning_rate": 6.341059602649006e-05,
"loss": 0.7311,
"step": 1150
},
{
"epoch": 0.1986754966887417,
"grad_norm": 1.4897147417068481,
"learning_rate": 6.616997792494481e-05,
"loss": 0.6918,
"step": 1200
},
{
"epoch": 0.20695364238410596,
"grad_norm": 1.634765863418579,
"learning_rate": 6.892935982339957e-05,
"loss": 0.7051,
"step": 1250
},
{
"epoch": 0.2152317880794702,
"grad_norm": 1.4463587999343872,
"learning_rate": 7.168874172185431e-05,
"loss": 0.6955,
"step": 1300
},
{
"epoch": 0.22350993377483444,
"grad_norm": 1.632133960723877,
"learning_rate": 7.444812362030905e-05,
"loss": 0.6901,
"step": 1350
},
{
"epoch": 0.23178807947019867,
"grad_norm": 1.4062328338623047,
"learning_rate": 7.72075055187638e-05,
"loss": 0.6833,
"step": 1400
},
{
"epoch": 0.24006622516556292,
"grad_norm": 1.2914466857910156,
"learning_rate": 7.996688741721855e-05,
"loss": 0.6663,
"step": 1450
},
{
"epoch": 0.24834437086092714,
"grad_norm": 1.4995919466018677,
"learning_rate": 8.272626931567329e-05,
"loss": 0.6959,
"step": 1500
},
{
"epoch": 0.25662251655629137,
"grad_norm": 1.1299749612808228,
"learning_rate": 8.548565121412803e-05,
"loss": 0.6685,
"step": 1550
},
{
"epoch": 0.26490066225165565,
"grad_norm": 1.329004168510437,
"learning_rate": 8.824503311258279e-05,
"loss": 0.6678,
"step": 1600
},
{
"epoch": 0.2731788079470199,
"grad_norm": 1.5191948413848877,
"learning_rate": 9.100441501103754e-05,
"loss": 0.6731,
"step": 1650
},
{
"epoch": 0.2814569536423841,
"grad_norm": 1.739169716835022,
"learning_rate": 9.376379690949227e-05,
"loss": 0.6691,
"step": 1700
},
{
"epoch": 0.2897350993377483,
"grad_norm": 1.2906118631362915,
"learning_rate": 9.652317880794703e-05,
"loss": 0.6718,
"step": 1750
},
{
"epoch": 0.2980132450331126,
"grad_norm": 1.289502501487732,
"learning_rate": 9.928256070640178e-05,
"loss": 0.6581,
"step": 1800
},
{
"epoch": 0.30629139072847683,
"grad_norm": 1.3923128843307495,
"learning_rate": 9.999872989402833e-05,
"loss": 0.6589,
"step": 1850
},
{
"epoch": 0.31456953642384106,
"grad_norm": 1.1048816442489624,
"learning_rate": 9.999297790520483e-05,
"loss": 0.6341,
"step": 1900
},
{
"epoch": 0.3228476821192053,
"grad_norm": 1.3568603992462158,
"learning_rate": 9.998258777484084e-05,
"loss": 0.6318,
"step": 1950
},
{
"epoch": 0.33112582781456956,
"grad_norm": 0.923786997795105,
"learning_rate": 9.996756046688961e-05,
"loss": 0.6318,
"step": 2000
},
{
"epoch": 0.3394039735099338,
"grad_norm": 1.102367877960205,
"learning_rate": 9.994789737552259e-05,
"loss": 0.6193,
"step": 2050
},
{
"epoch": 0.347682119205298,
"grad_norm": 1.0738896131515503,
"learning_rate": 9.992360032500001e-05,
"loss": 0.6184,
"step": 2100
},
{
"epoch": 0.35596026490066224,
"grad_norm": 1.279288649559021,
"learning_rate": 9.98946715695016e-05,
"loss": 0.626,
"step": 2150
},
{
"epoch": 0.36423841059602646,
"grad_norm": 1.2009036540985107,
"learning_rate": 9.986111379291759e-05,
"loss": 0.6305,
"step": 2200
},
{
"epoch": 0.37251655629139074,
"grad_norm": 0.8177038431167603,
"learning_rate": 9.982293010859955e-05,
"loss": 0.6266,
"step": 2250
},
{
"epoch": 0.38079470198675497,
"grad_norm": 1.2464983463287354,
"learning_rate": 9.978012405907165e-05,
"loss": 0.6148,
"step": 2300
},
{
"epoch": 0.3890728476821192,
"grad_norm": 1.2841860055923462,
"learning_rate": 9.973269961570195e-05,
"loss": 0.5946,
"step": 2350
},
{
"epoch": 0.3973509933774834,
"grad_norm": 1.2200431823730469,
"learning_rate": 9.968066117833401e-05,
"loss": 0.6166,
"step": 2400
},
{
"epoch": 0.4056291390728477,
"grad_norm": 1.128247857093811,
"learning_rate": 9.962401357487863e-05,
"loss": 0.5992,
"step": 2450
},
{
"epoch": 0.4139072847682119,
"grad_norm": 1.0683091878890991,
"learning_rate": 9.956276206086597e-05,
"loss": 0.6048,
"step": 2500
},
{
"epoch": 0.42218543046357615,
"grad_norm": 1.1819758415222168,
"learning_rate": 9.949691231895791e-05,
"loss": 0.5944,
"step": 2550
},
{
"epoch": 0.4304635761589404,
"grad_norm": 1.0043411254882812,
"learning_rate": 9.942647045842095e-05,
"loss": 0.5962,
"step": 2600
},
{
"epoch": 0.43874172185430466,
"grad_norm": 1.0588668584823608,
"learning_rate": 9.93514430145593e-05,
"loss": 0.6067,
"step": 2650
},
{
"epoch": 0.4470198675496689,
"grad_norm": 0.9364084601402283,
"learning_rate": 9.927183694810862e-05,
"loss": 0.5928,
"step": 2700
},
{
"epoch": 0.4552980132450331,
"grad_norm": 1.155172348022461,
"learning_rate": 9.918765964459022e-05,
"loss": 0.5987,
"step": 2750
},
{
"epoch": 0.46357615894039733,
"grad_norm": 1.1639224290847778,
"learning_rate": 9.909891891362587e-05,
"loss": 0.5745,
"step": 2800
},
{
"epoch": 0.4718543046357616,
"grad_norm": 0.9658174514770508,
"learning_rate": 9.900562298821323e-05,
"loss": 0.5825,
"step": 2850
},
{
"epoch": 0.48013245033112584,
"grad_norm": 1.118033766746521,
"learning_rate": 9.890778052396205e-05,
"loss": 0.5806,
"step": 2900
},
{
"epoch": 0.48841059602649006,
"grad_norm": 0.9781912565231323,
"learning_rate": 9.880540059829115e-05,
"loss": 0.5712,
"step": 2950
},
{
"epoch": 0.4966887417218543,
"grad_norm": 1.2145684957504272,
"learning_rate": 9.869849270958622e-05,
"loss": 0.5855,
"step": 3000
},
{
"epoch": 0.5049668874172185,
"grad_norm": 0.999279260635376,
"learning_rate": 9.858706677631862e-05,
"loss": 0.5843,
"step": 3050
},
{
"epoch": 0.5132450331125827,
"grad_norm": 1.098258137702942,
"learning_rate": 9.847113313612517e-05,
"loss": 0.5605,
"step": 3100
},
{
"epoch": 0.5215231788079471,
"grad_norm": 0.627949059009552,
"learning_rate": 9.835070254484912e-05,
"loss": 0.5538,
"step": 3150
},
{
"epoch": 0.5298013245033113,
"grad_norm": 1.0991902351379395,
"learning_rate": 9.822578617554219e-05,
"loss": 0.5555,
"step": 3200
},
{
"epoch": 0.5380794701986755,
"grad_norm": 0.9670843482017517,
"learning_rate": 9.8096395617428e-05,
"loss": 0.5647,
"step": 3250
},
{
"epoch": 0.5463576158940397,
"grad_norm": 0.9838133454322815,
"learning_rate": 9.796254287482693e-05,
"loss": 0.5561,
"step": 3300
},
{
"epoch": 0.554635761589404,
"grad_norm": 1.1465744972229004,
"learning_rate": 9.782424036604234e-05,
"loss": 0.559,
"step": 3350
},
{
"epoch": 0.5629139072847682,
"grad_norm": 1.1423758268356323,
"learning_rate": 9.768150092220849e-05,
"loss": 0.5517,
"step": 3400
},
{
"epoch": 0.5711920529801324,
"grad_norm": 1.1365066766738892,
"learning_rate": 9.753433778610008e-05,
"loss": 0.5464,
"step": 3450
},
{
"epoch": 0.5794701986754967,
"grad_norm": 0.81045001745224,
"learning_rate": 9.738276461090371e-05,
"loss": 0.5493,
"step": 3500
},
{
"epoch": 0.5877483443708609,
"grad_norm": 1.0236687660217285,
"learning_rate": 9.72267954589511e-05,
"loss": 0.567,
"step": 3550
},
{
"epoch": 0.5960264900662252,
"grad_norm": 0.9495602250099182,
"learning_rate": 9.706644480041455e-05,
"loss": 0.5474,
"step": 3600
},
{
"epoch": 0.6043046357615894,
"grad_norm": 0.960738480091095,
"learning_rate": 9.690172751196437e-05,
"loss": 0.5238,
"step": 3650
},
{
"epoch": 0.6125827814569537,
"grad_norm": 1.0488675832748413,
"learning_rate": 9.67326588753887e-05,
"loss": 0.521,
"step": 3700
},
{
"epoch": 0.6208609271523179,
"grad_norm": 0.8753538727760315,
"learning_rate": 9.65592545761758e-05,
"loss": 0.5232,
"step": 3750
},
{
"epoch": 0.6291390728476821,
"grad_norm": 1.0551217794418335,
"learning_rate": 9.638153070205871e-05,
"loss": 0.5432,
"step": 3800
},
{
"epoch": 0.6374172185430463,
"grad_norm": 1.158676028251648,
"learning_rate": 9.619950374152278e-05,
"loss": 0.5416,
"step": 3850
},
{
"epoch": 0.6456953642384106,
"grad_norm": 1.0036752223968506,
"learning_rate": 9.601319058227589e-05,
"loss": 0.5496,
"step": 3900
},
{
"epoch": 0.6539735099337748,
"grad_norm": 0.8905594348907471,
"learning_rate": 9.58226085096817e-05,
"loss": 0.5335,
"step": 3950
},
{
"epoch": 0.6622516556291391,
"grad_norm": 0.9868190884590149,
"learning_rate": 9.562777520515598e-05,
"loss": 0.5094,
"step": 4000
},
{
"epoch": 0.6705298013245033,
"grad_norm": 0.9672690629959106,
"learning_rate": 9.542870874452618e-05,
"loss": 0.5061,
"step": 4050
},
{
"epoch": 0.6788079470198676,
"grad_norm": 1.044123888015747,
"learning_rate": 9.52254275963545e-05,
"loss": 0.5253,
"step": 4100
},
{
"epoch": 0.6870860927152318,
"grad_norm": 1.0346958637237549,
"learning_rate": 9.501795062022434e-05,
"loss": 0.5149,
"step": 4150
},
{
"epoch": 0.695364238410596,
"grad_norm": 1.0799248218536377,
"learning_rate": 9.48062970649907e-05,
"loss": 0.5207,
"step": 4200
},
{
"epoch": 0.7036423841059603,
"grad_norm": 0.9847925901412964,
"learning_rate": 9.459048656699427e-05,
"loss": 0.531,
"step": 4250
},
{
"epoch": 0.7119205298013245,
"grad_norm": 1.134179949760437,
"learning_rate": 9.43705391482397e-05,
"loss": 0.5202,
"step": 4300
},
{
"epoch": 0.7201986754966887,
"grad_norm": 0.9750307202339172,
"learning_rate": 9.414647521453798e-05,
"loss": 0.5183,
"step": 4350
},
{
"epoch": 0.7284768211920529,
"grad_norm": 1.372010350227356,
"learning_rate": 9.391831555361341e-05,
"loss": 0.5203,
"step": 4400
},
{
"epoch": 0.7367549668874173,
"grad_norm": 0.9671643376350403,
"learning_rate": 9.36860813331748e-05,
"loss": 0.5313,
"step": 4450
},
{
"epoch": 0.7450331125827815,
"grad_norm": 1.270264983177185,
"learning_rate": 9.344979409895178e-05,
"loss": 0.5236,
"step": 4500
},
{
"epoch": 0.7533112582781457,
"grad_norm": 1.1816293001174927,
"learning_rate": 9.320947577269581e-05,
"loss": 0.518,
"step": 4550
},
{
"epoch": 0.7615894039735099,
"grad_norm": 0.8809527158737183,
"learning_rate": 9.29651486501464e-05,
"loss": 0.5086,
"step": 4600
},
{
"epoch": 0.7698675496688742,
"grad_norm": 0.9570929408073425,
"learning_rate": 9.271683539896257e-05,
"loss": 0.5195,
"step": 4650
},
{
"epoch": 0.7781456953642384,
"grad_norm": 1.147157907485962,
"learning_rate": 9.246455905661983e-05,
"loss": 0.509,
"step": 4700
},
{
"epoch": 0.7864238410596026,
"grad_norm": 0.9548070430755615,
"learning_rate": 9.220834302827295e-05,
"loss": 0.5078,
"step": 4750
},
{
"epoch": 0.7947019867549668,
"grad_norm": 1.0823866128921509,
"learning_rate": 9.194821108458438e-05,
"loss": 0.5088,
"step": 4800
},
{
"epoch": 0.8029801324503312,
"grad_norm": 1.0077965259552002,
"learning_rate": 9.168418735951902e-05,
"loss": 0.4994,
"step": 4850
},
{
"epoch": 0.8112582781456954,
"grad_norm": 1.1390560865402222,
"learning_rate": 9.141629634810516e-05,
"loss": 0.5098,
"step": 4900
},
{
"epoch": 0.8195364238410596,
"grad_norm": 1.1819497346878052,
"learning_rate": 9.114456290416186e-05,
"loss": 0.5012,
"step": 4950
},
{
"epoch": 0.8278145695364238,
"grad_norm": 0.8828374147415161,
"learning_rate": 9.08690122379932e-05,
"loss": 0.4895,
"step": 5000
},
{
"epoch": 0.8360927152317881,
"grad_norm": 1.263590693473816,
"learning_rate": 9.058966991404933e-05,
"loss": 0.5088,
"step": 5050
},
{
"epoch": 0.8443708609271523,
"grad_norm": 1.0225087404251099,
"learning_rate": 9.03065618485547e-05,
"loss": 0.5029,
"step": 5100
},
{
"epoch": 0.8526490066225165,
"grad_norm": 1.0702545642852783,
"learning_rate": 9.001971430710368e-05,
"loss": 0.5042,
"step": 5150
},
{
"epoch": 0.8609271523178808,
"grad_norm": 1.1163524389266968,
"learning_rate": 8.972915390222376e-05,
"loss": 0.4973,
"step": 5200
},
{
"epoch": 0.8692052980132451,
"grad_norm": 1.0248407125473022,
"learning_rate": 8.943490759090648e-05,
"loss": 0.4907,
"step": 5250
},
{
"epoch": 0.8774834437086093,
"grad_norm": 1.125511646270752,
"learning_rate": 8.913700267210657e-05,
"loss": 0.4982,
"step": 5300
},
{
"epoch": 0.8857615894039735,
"grad_norm": 1.0324009656906128,
"learning_rate": 8.883546678420917e-05,
"loss": 0.4973,
"step": 5350
},
{
"epoch": 0.8940397350993378,
"grad_norm": 0.9293155670166016,
"learning_rate": 8.853032790246575e-05,
"loss": 0.4938,
"step": 5400
},
{
"epoch": 0.902317880794702,
"grad_norm": 1.0467238426208496,
"learning_rate": 8.822161433639864e-05,
"loss": 0.5071,
"step": 5450
},
{
"epoch": 0.9105960264900662,
"grad_norm": 1.26283597946167,
"learning_rate": 8.790935472717452e-05,
"loss": 0.4943,
"step": 5500
},
{
"epoch": 0.9188741721854304,
"grad_norm": 1.0565682649612427,
"learning_rate": 8.75935780449473e-05,
"loss": 0.4884,
"step": 5550
},
{
"epoch": 0.9271523178807947,
"grad_norm": 1.0465178489685059,
"learning_rate": 8.727431358617042e-05,
"loss": 0.48,
"step": 5600
},
{
"epoch": 0.9354304635761589,
"grad_norm": 0.8586106300354004,
"learning_rate": 8.695159097087872e-05,
"loss": 0.4759,
"step": 5650
},
{
"epoch": 0.9437086092715232,
"grad_norm": 1.185562014579773,
"learning_rate": 8.662544013994054e-05,
"loss": 0.4788,
"step": 5700
},
{
"epoch": 0.9519867549668874,
"grad_norm": 0.8274655938148499,
"learning_rate": 8.62958913522798e-05,
"loss": 0.4831,
"step": 5750
},
{
"epoch": 0.9602649006622517,
"grad_norm": 1.0315494537353516,
"learning_rate": 8.596297518206889e-05,
"loss": 0.4755,
"step": 5800
},
{
"epoch": 0.9685430463576159,
"grad_norm": 1.0181632041931152,
"learning_rate": 8.562672251589188e-05,
"loss": 0.4784,
"step": 5850
},
{
"epoch": 0.9768211920529801,
"grad_norm": 1.0092475414276123,
"learning_rate": 8.528716454987927e-05,
"loss": 0.4638,
"step": 5900
},
{
"epoch": 0.9850993377483444,
"grad_norm": 1.1267504692077637,
"learning_rate": 8.494433278681347e-05,
"loss": 0.4765,
"step": 5950
},
{
"epoch": 0.9933774834437086,
"grad_norm": 1.1000663042068481,
"learning_rate": 8.459825903320628e-05,
"loss": 0.495,
"step": 6000
},
{
"epoch": 1.0,
"eval_loss": 0.4767566919326782,
"eval_runtime": 1888.1726,
"eval_samples_per_second": 3.199,
"eval_steps_per_second": 3.199,
"step": 6040
},
{
"epoch": 1.0016556291390728,
"grad_norm": 1.1419239044189453,
"learning_rate": 8.424897539634801e-05,
"loss": 0.4541,
"step": 6050
},
{
"epoch": 1.009933774834437,
"grad_norm": 1.235561728477478,
"learning_rate": 8.389651428132857e-05,
"loss": 0.4453,
"step": 6100
},
{
"epoch": 1.0182119205298013,
"grad_norm": 1.079288363456726,
"learning_rate": 8.354090838803115e-05,
"loss": 0.4274,
"step": 6150
},
{
"epoch": 1.0264900662251655,
"grad_norm": 1.0246189832687378,
"learning_rate": 8.318219070809851e-05,
"loss": 0.4297,
"step": 6200
},
{
"epoch": 1.0347682119205297,
"grad_norm": 1.0732612609863281,
"learning_rate": 8.282039452187206e-05,
"loss": 0.4369,
"step": 6250
},
{
"epoch": 1.0430463576158941,
"grad_norm": 1.0800081491470337,
"learning_rate": 8.245555339530427e-05,
"loss": 0.4245,
"step": 6300
},
{
"epoch": 1.0513245033112584,
"grad_norm": 1.2054654359817505,
"learning_rate": 8.208770117684455e-05,
"loss": 0.4366,
"step": 6350
},
{
"epoch": 1.0596026490066226,
"grad_norm": 0.8937723636627197,
"learning_rate": 8.171687199429901e-05,
"loss": 0.4247,
"step": 6400
},
{
"epoch": 1.0678807947019868,
"grad_norm": 1.1639046669006348,
"learning_rate": 8.13431002516641e-05,
"loss": 0.4334,
"step": 6450
},
{
"epoch": 1.076158940397351,
"grad_norm": 0.9183699488639832,
"learning_rate": 8.096642062593489e-05,
"loss": 0.426,
"step": 6500
},
{
"epoch": 1.0844370860927153,
"grad_norm": 1.05479097366333,
"learning_rate": 8.058686806388772e-05,
"loss": 0.4317,
"step": 6550
},
{
"epoch": 1.0927152317880795,
"grad_norm": 1.0889151096343994,
"learning_rate": 8.020447777883813e-05,
"loss": 0.4214,
"step": 6600
},
{
"epoch": 1.1009933774834437,
"grad_norm": 0.9915281534194946,
"learning_rate": 7.981928524737386e-05,
"loss": 0.437,
"step": 6650
},
{
"epoch": 1.109271523178808,
"grad_norm": 1.159812569618225,
"learning_rate": 7.943132620606341e-05,
"loss": 0.433,
"step": 6700
},
{
"epoch": 1.1175496688741722,
"grad_norm": 1.2697975635528564,
"learning_rate": 7.904063664814065e-05,
"loss": 0.4099,
"step": 6750
},
{
"epoch": 1.1258278145695364,
"grad_norm": 1.1016411781311035,
"learning_rate": 7.86472528201655e-05,
"loss": 0.4172,
"step": 6800
},
{
"epoch": 1.1341059602649006,
"grad_norm": 1.155731439590454,
"learning_rate": 7.825121121866106e-05,
"loss": 0.4262,
"step": 6850
},
{
"epoch": 1.1423841059602649,
"grad_norm": 1.0718499422073364,
"learning_rate": 7.785254858672768e-05,
"loss": 0.4245,
"step": 6900
},
{
"epoch": 1.150662251655629,
"grad_norm": 1.3144177198410034,
"learning_rate": 7.745130191063405e-05,
"loss": 0.4213,
"step": 6950
},
{
"epoch": 1.1589403973509933,
"grad_norm": 1.1578547954559326,
"learning_rate": 7.704750841638581e-05,
"loss": 0.4179,
"step": 7000
},
{
"epoch": 1.1672185430463575,
"grad_norm": 1.2637394666671753,
"learning_rate": 7.664120556627181e-05,
"loss": 0.4392,
"step": 7050
},
{
"epoch": 1.1754966887417218,
"grad_norm": 0.9308638572692871,
"learning_rate": 7.623243105538858e-05,
"loss": 0.4306,
"step": 7100
},
{
"epoch": 1.1837748344370862,
"grad_norm": 1.062219500541687,
"learning_rate": 7.582122280814305e-05,
"loss": 0.4332,
"step": 7150
},
{
"epoch": 1.1920529801324504,
"grad_norm": 1.0162498950958252,
"learning_rate": 7.540761897473421e-05,
"loss": 0.4105,
"step": 7200
},
{
"epoch": 1.2003311258278146,
"grad_norm": 1.3984551429748535,
"learning_rate": 7.499165792761355e-05,
"loss": 0.4216,
"step": 7250
},
{
"epoch": 1.2086092715231789,
"grad_norm": 1.1185648441314697,
"learning_rate": 7.457337825792515e-05,
"loss": 0.4115,
"step": 7300
},
{
"epoch": 1.216887417218543,
"grad_norm": 1.1665916442871094,
"learning_rate": 7.415281877192525e-05,
"loss": 0.4203,
"step": 7350
},
{
"epoch": 1.2251655629139073,
"grad_norm": 1.1253561973571777,
"learning_rate": 7.373001848738202e-05,
"loss": 0.4035,
"step": 7400
},
{
"epoch": 1.2334437086092715,
"grad_norm": 1.0157604217529297,
"learning_rate": 7.330501662995566e-05,
"loss": 0.4215,
"step": 7450
},
{
"epoch": 1.2417218543046358,
"grad_norm": 0.9860717058181763,
"learning_rate": 7.287785262955919e-05,
"loss": 0.4305,
"step": 7500
},
{
"epoch": 1.25,
"grad_norm": 0.9860566258430481,
"learning_rate": 7.244856611670025e-05,
"loss": 0.4188,
"step": 7550
},
{
"epoch": 1.2582781456953642,
"grad_norm": 1.1415915489196777,
"learning_rate": 7.201719691880446e-05,
"loss": 0.4273,
"step": 7600
},
{
"epoch": 1.2665562913907285,
"grad_norm": 1.2406401634216309,
"learning_rate": 7.158378505652033e-05,
"loss": 0.4078,
"step": 7650
},
{
"epoch": 1.2748344370860927,
"grad_norm": 1.1644037961959839,
"learning_rate": 7.11483707400063e-05,
"loss": 0.4271,
"step": 7700
},
{
"epoch": 1.283112582781457,
"grad_norm": 1.1102714538574219,
"learning_rate": 7.07109943652002e-05,
"loss": 0.4047,
"step": 7750
},
{
"epoch": 1.2913907284768211,
"grad_norm": 0.9731109142303467,
"learning_rate": 7.027169651007156e-05,
"loss": 0.4123,
"step": 7800
},
{
"epoch": 1.2996688741721854,
"grad_norm": 1.0506606101989746,
"learning_rate": 6.983051793085688e-05,
"loss": 0.4143,
"step": 7850
},
{
"epoch": 1.3079470198675498,
"grad_norm": 1.263192057609558,
"learning_rate": 6.938749955827842e-05,
"loss": 0.4151,
"step": 7900
},
{
"epoch": 1.3162251655629138,
"grad_norm": 0.9878462553024292,
"learning_rate": 6.894268249374689e-05,
"loss": 0.4149,
"step": 7950
},
{
"epoch": 1.3245033112582782,
"grad_norm": 1.1122633218765259,
"learning_rate": 6.84961080055482e-05,
"loss": 0.3948,
"step": 8000
},
{
"epoch": 1.3327814569536423,
"grad_norm": 1.1024750471115112,
"learning_rate": 6.804781752501475e-05,
"loss": 0.4124,
"step": 8050
},
{
"epoch": 1.3410596026490067,
"grad_norm": 1.1353663206100464,
"learning_rate": 6.759785264268154e-05,
"loss": 0.4062,
"step": 8100
},
{
"epoch": 1.349337748344371,
"grad_norm": 1.1376144886016846,
"learning_rate": 6.714625510442773e-05,
"loss": 0.4244,
"step": 8150
},
{
"epoch": 1.3576158940397351,
"grad_norm": 1.0264923572540283,
"learning_rate": 6.669306680760351e-05,
"loss": 0.4061,
"step": 8200
},
{
"epoch": 1.3658940397350994,
"grad_norm": 0.8679251670837402,
"learning_rate": 6.623832979714302e-05,
"loss": 0.4018,
"step": 8250
},
{
"epoch": 1.3741721854304636,
"grad_norm": 1.014840841293335,
"learning_rate": 6.57820862616637e-05,
"loss": 0.4058,
"step": 8300
},
{
"epoch": 1.3824503311258278,
"grad_norm": 0.9928984045982361,
"learning_rate": 6.53243785295521e-05,
"loss": 0.4022,
"step": 8350
},
{
"epoch": 1.390728476821192,
"grad_norm": 1.077910304069519,
"learning_rate": 6.48652490650369e-05,
"loss": 0.4092,
"step": 8400
},
{
"epoch": 1.3990066225165563,
"grad_norm": 1.1118338108062744,
"learning_rate": 6.440474046424923e-05,
"loss": 0.4047,
"step": 8450
},
{
"epoch": 1.4072847682119205,
"grad_norm": 1.0150924921035767,
"learning_rate": 6.394289545127073e-05,
"loss": 0.4052,
"step": 8500
},
{
"epoch": 1.4155629139072847,
"grad_norm": 1.129887342453003,
"learning_rate": 6.34797568741699e-05,
"loss": 0.4124,
"step": 8550
},
{
"epoch": 1.423841059602649,
"grad_norm": 1.0517767667770386,
"learning_rate": 6.30153677010267e-05,
"loss": 0.3971,
"step": 8600
},
{
"epoch": 1.4321192052980132,
"grad_norm": 1.107248067855835,
"learning_rate": 6.254977101594625e-05,
"loss": 0.3871,
"step": 8650
},
{
"epoch": 1.4403973509933774,
"grad_norm": 1.0460799932479858,
"learning_rate": 6.208301001506162e-05,
"loss": 0.3998,
"step": 8700
},
{
"epoch": 1.4486754966887418,
"grad_norm": 1.0304828882217407,
"learning_rate": 6.16151280025263e-05,
"loss": 0.4116,
"step": 8750
},
{
"epoch": 1.4569536423841059,
"grad_norm": 1.321883201599121,
"learning_rate": 6.114616838649656e-05,
"loss": 0.3988,
"step": 8800
},
{
"epoch": 1.4652317880794703,
"grad_norm": 1.1075403690338135,
"learning_rate": 6.067617467510429e-05,
"loss": 0.4081,
"step": 8850
},
{
"epoch": 1.4735099337748343,
"grad_norm": 1.1279836893081665,
"learning_rate": 6.020519047242046e-05,
"loss": 0.4025,
"step": 8900
},
{
"epoch": 1.4817880794701987,
"grad_norm": 1.2322368621826172,
"learning_rate": 5.973325947440972e-05,
"loss": 0.398,
"step": 8950
},
{
"epoch": 1.490066225165563,
"grad_norm": 0.6588311195373535,
"learning_rate": 5.926042546487647e-05,
"loss": 0.4045,
"step": 9000
},
{
"epoch": 1.4983443708609272,
"grad_norm": 0.9893741607666016,
"learning_rate": 5.878673231140279e-05,
"loss": 0.4069,
"step": 9050
},
{
"epoch": 1.5066225165562914,
"grad_norm": 1.0072598457336426,
"learning_rate": 5.831222396127858e-05,
"loss": 0.4099,
"step": 9100
},
{
"epoch": 1.5149006622516556,
"grad_norm": 0.9812730550765991,
"learning_rate": 5.783694443742429e-05,
"loss": 0.3965,
"step": 9150
},
{
"epoch": 1.5231788079470199,
"grad_norm": 1.0679937601089478,
"learning_rate": 5.7360937834306693e-05,
"loss": 0.3856,
"step": 9200
},
{
"epoch": 1.531456953642384,
"grad_norm": 1.1114174127578735,
"learning_rate": 5.688424831384795e-05,
"loss": 0.3869,
"step": 9250
},
{
"epoch": 1.5397350993377483,
"grad_norm": 0.962031900882721,
"learning_rate": 5.640692010132851e-05,
"loss": 0.3917,
"step": 9300
},
{
"epoch": 1.5480132450331126,
"grad_norm": 0.6730946898460388,
"learning_rate": 5.5928997481283976e-05,
"loss": 0.3911,
"step": 9350
},
{
"epoch": 1.5562913907284768,
"grad_norm": 1.235037922859192,
"learning_rate": 5.545052479339662e-05,
"loss": 0.3775,
"step": 9400
},
{
"epoch": 1.564569536423841,
"grad_norm": 0.990074634552002,
"learning_rate": 5.497154642838179e-05,
"loss": 0.396,
"step": 9450
},
{
"epoch": 1.5728476821192054,
"grad_norm": 1.0766572952270508,
"learning_rate": 5.449210682386942e-05,
"loss": 0.3711,
"step": 9500
},
{
"epoch": 1.5811258278145695,
"grad_norm": 0.9285266995429993,
"learning_rate": 5.401225046028131e-05,
"loss": 0.383,
"step": 9550
},
{
"epoch": 1.589403973509934,
"grad_norm": 1.1176483631134033,
"learning_rate": 5.3532021856704504e-05,
"loss": 0.3889,
"step": 9600
},
{
"epoch": 1.597682119205298,
"grad_norm": 1.178252935409546,
"learning_rate": 5.3051465566760895e-05,
"loss": 0.3879,
"step": 9650
},
{
"epoch": 1.6059602649006623,
"grad_norm": 1.2217224836349487,
"learning_rate": 5.2570626174473756e-05,
"loss": 0.384,
"step": 9700
},
{
"epoch": 1.6142384105960264,
"grad_norm": 1.11509370803833,
"learning_rate": 5.208954829013145e-05,
"loss": 0.379,
"step": 9750
},
{
"epoch": 1.6225165562913908,
"grad_norm": 0.9148808121681213,
"learning_rate": 5.1608276546148616e-05,
"loss": 0.3813,
"step": 9800
},
{
"epoch": 1.6307947019867548,
"grad_norm": 0.9564769268035889,
"learning_rate": 5.112685559292542e-05,
"loss": 0.3975,
"step": 9850
},
{
"epoch": 1.6390728476821192,
"grad_norm": 1.0157302618026733,
"learning_rate": 5.064533009470499e-05,
"loss": 0.3754,
"step": 9900
},
{
"epoch": 1.6473509933774835,
"grad_norm": 1.2600700855255127,
"learning_rate": 5.016374472542978e-05,
"loss": 0.3705,
"step": 9950
},
{
"epoch": 1.6556291390728477,
"grad_norm": 1.298558235168457,
"learning_rate": 4.968214416459678e-05,
"loss": 0.3898,
"step": 10000
},
{
"epoch": 1.663907284768212,
"grad_norm": 1.1298922300338745,
"learning_rate": 4.9200573093112384e-05,
"loss": 0.3898,
"step": 10050
},
{
"epoch": 1.6721854304635762,
"grad_norm": 0.9608431458473206,
"learning_rate": 4.871907618914714e-05,
"loss": 0.3705,
"step": 10100
},
{
"epoch": 1.6804635761589404,
"grad_norm": 1.0675232410430908,
"learning_rate": 4.823769812399059e-05,
"loss": 0.3881,
"step": 10150
},
{
"epoch": 1.6887417218543046,
"grad_norm": 1.101462960243225,
"learning_rate": 4.775648355790691e-05,
"loss": 0.3765,
"step": 10200
},
{
"epoch": 1.6970198675496688,
"grad_norm": 1.147665023803711,
"learning_rate": 4.7275477135991535e-05,
"loss": 0.3836,
"step": 10250
},
{
"epoch": 1.705298013245033,
"grad_norm": 1.0882872343063354,
"learning_rate": 4.679472348402913e-05,
"loss": 0.3841,
"step": 10300
},
{
"epoch": 1.7135761589403975,
"grad_norm": 1.0927859544754028,
"learning_rate": 4.631426720435339e-05,
"loss": 0.3833,
"step": 10350
},
{
"epoch": 1.7218543046357615,
"grad_norm": 1.2883074283599854,
"learning_rate": 4.583415287170908e-05,
"loss": 0.3785,
"step": 10400
},
{
"epoch": 1.730132450331126,
"grad_norm": 1.11234450340271,
"learning_rate": 4.535442502911653e-05,
"loss": 0.3805,
"step": 10450
},
{
"epoch": 1.73841059602649,
"grad_norm": 1.2244069576263428,
"learning_rate": 4.487512818373906e-05,
"loss": 0.3715,
"step": 10500
},
{
"epoch": 1.7466887417218544,
"grad_norm": 1.1873053312301636,
"learning_rate": 4.439630680275393e-05,
"loss": 0.3939,
"step": 10550
},
{
"epoch": 1.7549668874172184,
"grad_norm": 1.0485515594482422,
"learning_rate": 4.391800530922675e-05,
"loss": 0.3776,
"step": 10600
},
{
"epoch": 1.7632450331125828,
"grad_norm": 0.9849908351898193,
"learning_rate": 4.344026807799012e-05,
"loss": 0.3903,
"step": 10650
},
{
"epoch": 1.771523178807947,
"grad_norm": 1.3031227588653564,
"learning_rate": 4.296313943152673e-05,
"loss": 0.3715,
"step": 10700
},
{
"epoch": 1.7798013245033113,
"grad_norm": 1.0970308780670166,
"learning_rate": 4.2486663635857286e-05,
"loss": 0.3729,
"step": 10750
},
{
"epoch": 1.7880794701986755,
"grad_norm": 0.9399189949035645,
"learning_rate": 4.201088489643372e-05,
"loss": 0.363,
"step": 10800
},
{
"epoch": 1.7963576158940397,
"grad_norm": 1.198516845703125,
"learning_rate": 4.153584735403795e-05,
"loss": 0.3842,
"step": 10850
},
{
"epoch": 1.804635761589404,
"grad_norm": 0.9462034702301025,
"learning_rate": 4.106159508068668e-05,
"loss": 0.369,
"step": 10900
},
{
"epoch": 1.8129139072847682,
"grad_norm": 1.0740764141082764,
"learning_rate": 4.058817207554266e-05,
"loss": 0.354,
"step": 10950
},
{
"epoch": 1.8211920529801324,
"grad_norm": 1.0017762184143066,
"learning_rate": 4.011562226083254e-05,
"loss": 0.3705,
"step": 11000
},
{
"epoch": 1.8294701986754967,
"grad_norm": 1.1367080211639404,
"learning_rate": 3.964398947777196e-05,
"loss": 0.3889,
"step": 11050
},
{
"epoch": 1.8377483443708609,
"grad_norm": 1.0371805429458618,
"learning_rate": 3.9173317482498176e-05,
"loss": 0.3652,
"step": 11100
},
{
"epoch": 1.846026490066225,
"grad_norm": 1.145815134048462,
"learning_rate": 3.8703649942010535e-05,
"loss": 0.3701,
"step": 11150
},
{
"epoch": 1.8543046357615895,
"grad_norm": 1.0999510288238525,
"learning_rate": 3.8235030430119215e-05,
"loss": 0.3573,
"step": 11200
},
{
"epoch": 1.8625827814569536,
"grad_norm": 1.1765943765640259,
"learning_rate": 3.7767502423402645e-05,
"loss": 0.3663,
"step": 11250
},
{
"epoch": 1.870860927152318,
"grad_norm": 1.0484038591384888,
"learning_rate": 3.730110929717393e-05,
"loss": 0.3665,
"step": 11300
},
{
"epoch": 1.879139072847682,
"grad_norm": 0.9852702021598816,
"learning_rate": 3.6835894321456655e-05,
"loss": 0.3674,
"step": 11350
},
{
"epoch": 1.8874172185430464,
"grad_norm": 0.5963965654373169,
"learning_rate": 3.6371900656970446e-05,
"loss": 0.3527,
"step": 11400
},
{
"epoch": 1.8956953642384105,
"grad_norm": 0.9792616367340088,
"learning_rate": 3.59091713511268e-05,
"loss": 0.3624,
"step": 11450
},
{
"epoch": 1.903973509933775,
"grad_norm": 1.276879072189331,
"learning_rate": 3.5447749334035205e-05,
"loss": 0.3653,
"step": 11500
},
{
"epoch": 1.9122516556291391,
"grad_norm": 1.0514193773269653,
"learning_rate": 3.498767741452028e-05,
"loss": 0.3635,
"step": 11550
},
{
"epoch": 1.9205298013245033,
"grad_norm": 1.130402684211731,
"learning_rate": 3.452899827615026e-05,
"loss": 0.3629,
"step": 11600
},
{
"epoch": 1.9288079470198676,
"grad_norm": 0.9695820212364197,
"learning_rate": 3.407175447327685e-05,
"loss": 0.3649,
"step": 11650
},
{
"epoch": 1.9370860927152318,
"grad_norm": 1.0880156755447388,
"learning_rate": 3.361598842708727e-05,
"loss": 0.3618,
"step": 11700
},
{
"epoch": 1.945364238410596,
"grad_norm": 1.1198087930679321,
"learning_rate": 3.3161742421668654e-05,
"loss": 0.3672,
"step": 11750
},
{
"epoch": 1.9536423841059603,
"grad_norm": 0.8677820563316345,
"learning_rate": 3.2709058600084964e-05,
"loss": 0.3551,
"step": 11800
},
{
"epoch": 1.9619205298013245,
"grad_norm": 0.8376750349998474,
"learning_rate": 3.225797896046724e-05,
"loss": 0.3709,
"step": 11850
},
{
"epoch": 1.9701986754966887,
"grad_norm": 1.0590113401412964,
"learning_rate": 3.180854535211721e-05,
"loss": 0.3617,
"step": 11900
},
{
"epoch": 1.978476821192053,
"grad_norm": 1.0090305805206299,
"learning_rate": 3.136079947162456e-05,
"loss": 0.3584,
"step": 11950
},
{
"epoch": 1.9867549668874172,
"grad_norm": 0.9464877843856812,
"learning_rate": 3.091478285899862e-05,
"loss": 0.3612,
"step": 12000
},
{
"epoch": 1.9950331125827816,
"grad_norm": 1.2669893503189087,
"learning_rate": 3.0470536893814385e-05,
"loss": 0.3633,
"step": 12050
},
{
"epoch": 2.0,
"eval_loss": 0.38875386118888855,
"eval_runtime": 1884.7691,
"eval_samples_per_second": 3.205,
"eval_steps_per_second": 3.205,
"step": 12080
},
{
"epoch": 2.0033112582781456,
"grad_norm": 1.1019831895828247,
"learning_rate": 3.0028102791373535e-05,
"loss": 0.3221,
"step": 12100
},
{
"epoch": 2.01158940397351,
"grad_norm": 1.095613956451416,
"learning_rate": 2.9587521598880573e-05,
"loss": 0.281,
"step": 12150
},
{
"epoch": 2.019867549668874,
"grad_norm": 1.1338448524475098,
"learning_rate": 2.914883419163475e-05,
"loss": 0.2824,
"step": 12200
},
{
"epoch": 2.0281456953642385,
"grad_norm": 1.1247187852859497,
"learning_rate": 2.871208126923771e-05,
"loss": 0.2986,
"step": 12250
},
{
"epoch": 2.0364238410596025,
"grad_norm": 1.0773439407348633,
"learning_rate": 2.827730335181765e-05,
"loss": 0.2852,
"step": 12300
},
{
"epoch": 2.044701986754967,
"grad_norm": 1.343947410583496,
"learning_rate": 2.7844540776269924e-05,
"loss": 0.2905,
"step": 12350
},
{
"epoch": 2.052980132450331,
"grad_norm": 0.9737703800201416,
"learning_rate": 2.7413833692514844e-05,
"loss": 0.2791,
"step": 12400
},
{
"epoch": 2.0612582781456954,
"grad_norm": 1.0275969505310059,
"learning_rate": 2.698522205977273e-05,
"loss": 0.2887,
"step": 12450
},
{
"epoch": 2.0695364238410594,
"grad_norm": 1.1896377801895142,
"learning_rate": 2.655874564285656e-05,
"loss": 0.2845,
"step": 12500
},
{
"epoch": 2.077814569536424,
"grad_norm": 1.0280028581619263,
"learning_rate": 2.613444400848287e-05,
"loss": 0.284,
"step": 12550
},
{
"epoch": 2.0860927152317883,
"grad_norm": 1.2673025131225586,
"learning_rate": 2.571235652160091e-05,
"loss": 0.2852,
"step": 12600
},
{
"epoch": 2.0943708609271523,
"grad_norm": 1.076798915863037,
"learning_rate": 2.529252234174041e-05,
"loss": 0.2893,
"step": 12650
},
{
"epoch": 2.1026490066225167,
"grad_norm": 0.7327485084533691,
"learning_rate": 2.4874980419378647e-05,
"loss": 0.2837,
"step": 12700
},
{
"epoch": 2.1109271523178808,
"grad_norm": 1.2723850011825562,
"learning_rate": 2.445976949232676e-05,
"loss": 0.2859,
"step": 12750
},
{
"epoch": 2.119205298013245,
"grad_norm": 1.1690808534622192,
"learning_rate": 2.4046928082135733e-05,
"loss": 0.2841,
"step": 12800
},
{
"epoch": 2.127483443708609,
"grad_norm": 1.2063932418823242,
"learning_rate": 2.3636494490522624e-05,
"loss": 0.2772,
"step": 12850
},
{
"epoch": 2.1357615894039736,
"grad_norm": 1.2601985931396484,
"learning_rate": 2.3228506795817072e-05,
"loss": 0.2829,
"step": 12900
},
{
"epoch": 2.1440397350993377,
"grad_norm": 1.1223324537277222,
"learning_rate": 2.282300284942846e-05,
"loss": 0.293,
"step": 12950
},
{
"epoch": 2.152317880794702,
"grad_norm": 1.2425414323806763,
"learning_rate": 2.2420020272334337e-05,
"loss": 0.2946,
"step": 13000
},
{
"epoch": 2.160596026490066,
"grad_norm": 1.176163673400879,
"learning_rate": 2.2019596451590047e-05,
"loss": 0.2888,
"step": 13050
},
{
"epoch": 2.1688741721854305,
"grad_norm": 1.073018193244934,
"learning_rate": 2.162176853686006e-05,
"loss": 0.2921,
"step": 13100
},
{
"epoch": 2.1771523178807946,
"grad_norm": 1.3194624185562134,
"learning_rate": 2.1226573436971487e-05,
"loss": 0.3014,
"step": 13150
},
{
"epoch": 2.185430463576159,
"grad_norm": 1.3240703344345093,
"learning_rate": 2.0834047816489772e-05,
"loss": 0.2828,
"step": 13200
},
{
"epoch": 2.193708609271523,
"grad_norm": 1.529966950416565,
"learning_rate": 2.0444228092317057e-05,
"loss": 0.2946,
"step": 13250
},
{
"epoch": 2.2019867549668874,
"grad_norm": 1.0397886037826538,
"learning_rate": 2.005715043031369e-05,
"loss": 0.2905,
"step": 13300
},
{
"epoch": 2.2102649006622515,
"grad_norm": 0.8956925868988037,
"learning_rate": 1.967285074194283e-05,
"loss": 0.2847,
"step": 13350
},
{
"epoch": 2.218543046357616,
"grad_norm": 1.2051565647125244,
"learning_rate": 1.9291364680938688e-05,
"loss": 0.2709,
"step": 13400
},
{
"epoch": 2.22682119205298,
"grad_norm": 1.2182625532150269,
"learning_rate": 1.891272763999884e-05,
"loss": 0.2754,
"step": 13450
},
{
"epoch": 2.2350993377483444,
"grad_norm": 1.2709006071090698,
"learning_rate": 1.8536974747500556e-05,
"loss": 0.278,
"step": 13500
},
{
"epoch": 2.243377483443709,
"grad_norm": 1.142155647277832,
"learning_rate": 1.8164140864241723e-05,
"loss": 0.2863,
"step": 13550
},
{
"epoch": 2.251655629139073,
"grad_norm": 1.218385100364685,
"learning_rate": 1.7794260580206673e-05,
"loss": 0.282,
"step": 13600
},
{
"epoch": 2.2599337748344372,
"grad_norm": 0.6981252431869507,
"learning_rate": 1.742736821135702e-05,
"loss": 0.2756,
"step": 13650
},
{
"epoch": 2.2682119205298013,
"grad_norm": 1.193818211555481,
"learning_rate": 1.7063497796447935e-05,
"loss": 0.2711,
"step": 13700
},
{
"epoch": 2.2764900662251657,
"grad_norm": 1.1021385192871094,
"learning_rate": 1.670268309387029e-05,
"loss": 0.2723,
"step": 13750
},
{
"epoch": 2.2847682119205297,
"grad_norm": 1.4959111213684082,
"learning_rate": 1.634495757851855e-05,
"loss": 0.2821,
"step": 13800
},
{
"epoch": 2.293046357615894,
"grad_norm": 1.3577196598052979,
"learning_rate": 1.599035443868518e-05,
"loss": 0.2701,
"step": 13850
},
{
"epoch": 2.301324503311258,
"grad_norm": 1.391587495803833,
"learning_rate": 1.5638906572981604e-05,
"loss": 0.2809,
"step": 13900
},
{
"epoch": 2.3096026490066226,
"grad_norm": 1.1502324342727661,
"learning_rate": 1.529064658728598e-05,
"loss": 0.2847,
"step": 13950
},
{
"epoch": 2.3178807947019866,
"grad_norm": 1.315737009048462,
"learning_rate": 1.4945606791718092e-05,
"loss": 0.2772,
"step": 14000
},
{
"epoch": 2.326158940397351,
"grad_norm": 1.5041948556900024,
"learning_rate": 1.4603819197641883e-05,
"loss": 0.2798,
"step": 14050
},
{
"epoch": 2.334437086092715,
"grad_norm": 1.1226636171340942,
"learning_rate": 1.4265315514695488e-05,
"loss": 0.2712,
"step": 14100
},
{
"epoch": 2.3427152317880795,
"grad_norm": 0.982099175453186,
"learning_rate": 1.3930127147849314e-05,
"loss": 0.2805,
"step": 14150
},
{
"epoch": 2.3509933774834435,
"grad_norm": 1.165469765663147,
"learning_rate": 1.3598285194492521e-05,
"loss": 0.2882,
"step": 14200
},
{
"epoch": 2.359271523178808,
"grad_norm": 1.305770754814148,
"learning_rate": 1.326982044154787e-05,
"loss": 0.2837,
"step": 14250
},
{
"epoch": 2.3675496688741724,
"grad_norm": 1.5525685548782349,
"learning_rate": 1.2944763362615413e-05,
"loss": 0.2742,
"step": 14300
},
{
"epoch": 2.3758278145695364,
"grad_norm": 1.0929417610168457,
"learning_rate": 1.2623144115145342e-05,
"loss": 0.2698,
"step": 14350
},
{
"epoch": 2.384105960264901,
"grad_norm": 1.1220225095748901,
"learning_rate": 1.2304992537640092e-05,
"loss": 0.2764,
"step": 14400
},
{
"epoch": 2.392384105960265,
"grad_norm": 1.229873776435852,
"learning_rate": 1.1990338146885977e-05,
"loss": 0.2729,
"step": 14450
},
{
"epoch": 2.4006622516556293,
"grad_norm": 1.4030100107192993,
"learning_rate": 1.1679210135214858e-05,
"loss": 0.2764,
"step": 14500
},
{
"epoch": 2.4089403973509933,
"grad_norm": 1.1974605321884155,
"learning_rate": 1.1371637367795735e-05,
"loss": 0.2718,
"step": 14550
},
{
"epoch": 2.4172185430463577,
"grad_norm": 1.3544508218765259,
"learning_rate": 1.1067648379956714e-05,
"loss": 0.288,
"step": 14600
},
{
"epoch": 2.4254966887417218,
"grad_norm": 1.3159441947937012,
"learning_rate": 1.0767271374537724e-05,
"loss": 0.27,
"step": 14650
},
{
"epoch": 2.433774834437086,
"grad_norm": 1.531108021736145,
"learning_rate": 1.0470534219273903e-05,
"loss": 0.2764,
"step": 14700
},
{
"epoch": 2.44205298013245,
"grad_norm": 1.3860596418380737,
"learning_rate": 1.0177464444210133e-05,
"loss": 0.2685,
"step": 14750
},
{
"epoch": 2.4503311258278146,
"grad_norm": 1.0219587087631226,
"learning_rate": 9.888089239146963e-06,
"loss": 0.2882,
"step": 14800
},
{
"epoch": 2.4586092715231787,
"grad_norm": 1.3057914972305298,
"learning_rate": 9.602435451118047e-06,
"loss": 0.288,
"step": 14850
},
{
"epoch": 2.466887417218543,
"grad_norm": 1.2784461975097656,
"learning_rate": 9.320529581899335e-06,
"loss": 0.2797,
"step": 14900
},
{
"epoch": 2.475165562913907,
"grad_norm": 1.4044731855392456,
"learning_rate": 9.042397785550405e-06,
"loss": 0.2812,
"step": 14950
},
{
"epoch": 2.4834437086092715,
"grad_norm": 1.240498661994934,
"learning_rate": 8.768065865987995e-06,
"loss": 0.2719,
"step": 15000
},
{
"epoch": 2.491721854304636,
"grad_norm": 1.1060854196548462,
"learning_rate": 8.49755927459196e-06,
"loss": 0.2862,
"step": 15050
},
{
"epoch": 2.5,
"grad_norm": 1.4075932502746582,
"learning_rate": 8.230903107844078e-06,
"loss": 0.2622,
"step": 15100
},
{
"epoch": 2.508278145695364,
"grad_norm": 1.277173399925232,
"learning_rate": 7.968122104999676e-06,
"loss": 0.2735,
"step": 15150
},
{
"epoch": 2.5165562913907285,
"grad_norm": 1.4734828472137451,
"learning_rate": 7.70924064579236e-06,
"loss": 0.2857,
"step": 15200
},
{
"epoch": 2.524834437086093,
"grad_norm": 1.183908224105835,
"learning_rate": 7.454282748172281e-06,
"loss": 0.2647,
"step": 15250
},
{
"epoch": 2.533112582781457,
"grad_norm": 1.2471554279327393,
"learning_rate": 7.2032720660777706e-06,
"loss": 0.272,
"step": 15300
},
{
"epoch": 2.5413907284768213,
"grad_norm": 1.2211905717849731,
"learning_rate": 6.95623188724081e-06,
"loss": 0.2815,
"step": 15350
},
{
"epoch": 2.5496688741721854,
"grad_norm": 1.453582525253296,
"learning_rate": 6.713185131026567e-06,
"loss": 0.2808,
"step": 15400
},
{
"epoch": 2.55794701986755,
"grad_norm": 1.27803635597229,
"learning_rate": 6.474154346306999e-06,
"loss": 0.2719,
"step": 15450
},
{
"epoch": 2.566225165562914,
"grad_norm": 1.2783340215682983,
"learning_rate": 6.239161709368774e-06,
"loss": 0.2743,
"step": 15500
},
{
"epoch": 2.5745033112582782,
"grad_norm": 1.371692180633545,
"learning_rate": 6.00822902185601e-06,
"loss": 0.2817,
"step": 15550
},
{
"epoch": 2.5827814569536423,
"grad_norm": 1.6115282773971558,
"learning_rate": 5.781377708747493e-06,
"loss": 0.2665,
"step": 15600
},
{
"epoch": 2.5910596026490067,
"grad_norm": 1.3451199531555176,
"learning_rate": 5.558628816368972e-06,
"loss": 0.2767,
"step": 15650
},
{
"epoch": 2.5993377483443707,
"grad_norm": 1.2596265077590942,
"learning_rate": 5.340003010440603e-06,
"loss": 0.2721,
"step": 15700
},
{
"epoch": 2.607615894039735,
"grad_norm": 1.3860641717910767,
"learning_rate": 5.125520574159654e-06,
"loss": 0.2797,
"step": 15750
},
{
"epoch": 2.6158940397350996,
"grad_norm": 1.2152941226959229,
"learning_rate": 4.915201406318676e-06,
"loss": 0.2669,
"step": 15800
},
{
"epoch": 2.6241721854304636,
"grad_norm": 1.2556190490722656,
"learning_rate": 4.7090650194594465e-06,
"loss": 0.2756,
"step": 15850
},
{
"epoch": 2.6324503311258276,
"grad_norm": 1.2963857650756836,
"learning_rate": 4.50713053806262e-06,
"loss": 0.2737,
"step": 15900
},
{
"epoch": 2.640728476821192,
"grad_norm": 1.0009130239486694,
"learning_rate": 4.309416696773455e-06,
"loss": 0.27,
"step": 15950
},
{
"epoch": 2.6490066225165565,
"grad_norm": 1.1450060606002808,
"learning_rate": 4.1159418386636895e-06,
"loss": 0.2659,
"step": 16000
},
{
"epoch": 2.6572847682119205,
"grad_norm": 1.2123562097549438,
"learning_rate": 3.926723913529773e-06,
"loss": 0.2739,
"step": 16050
},
{
"epoch": 2.6655629139072845,
"grad_norm": 1.3714101314544678,
"learning_rate": 3.7417804762274968e-06,
"loss": 0.2801,
"step": 16100
},
{
"epoch": 2.673841059602649,
"grad_norm": 1.3810080289840698,
"learning_rate": 3.5611286850433967e-06,
"loss": 0.2646,
"step": 16150
},
{
"epoch": 2.6821192052980134,
"grad_norm": 1.0282268524169922,
"learning_rate": 3.3847853001028495e-06,
"loss": 0.2699,
"step": 16200
},
{
"epoch": 2.6903973509933774,
"grad_norm": 1.1786317825317383,
"learning_rate": 3.2127666818151046e-06,
"loss": 0.2717,
"step": 16250
},
{
"epoch": 2.698675496688742,
"grad_norm": 1.0810590982437134,
"learning_rate": 3.045088789355488e-06,
"loss": 0.277,
"step": 16300
},
{
"epoch": 2.706953642384106,
"grad_norm": 1.194557785987854,
"learning_rate": 2.8817671791847634e-06,
"loss": 0.2652,
"step": 16350
},
{
"epoch": 2.7152317880794703,
"grad_norm": 1.1923502683639526,
"learning_rate": 2.7228170036058153e-06,
"loss": 0.2617,
"step": 16400
}
],
"logging_steps": 50,
"max_steps": 18120,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 100,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.2874605314388787e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}