Arko007's picture
V7 MARATHON: Uploading fine-tuned CodeLlama-7B model
5e5a4fe verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 8.559201141226819,
"eval_steps": 500,
"global_step": 6000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.07132667617689016,
"grad_norm": 8.374612808227539,
"learning_rate": 2.9400000000000002e-06,
"loss": 1.4876,
"step": 50
},
{
"epoch": 0.14265335235378032,
"grad_norm": 1.39237642288208,
"learning_rate": 2.9772093023255814e-06,
"loss": 1.2975,
"step": 100
},
{
"epoch": 0.21398002853067047,
"grad_norm": 1.192126989364624,
"learning_rate": 2.953953488372093e-06,
"loss": 1.2197,
"step": 150
},
{
"epoch": 0.28530670470756064,
"grad_norm": 1.450819730758667,
"learning_rate": 2.9306976744186045e-06,
"loss": 1.1853,
"step": 200
},
{
"epoch": 0.3566333808844508,
"grad_norm": 1.5788187980651855,
"learning_rate": 2.9074418604651165e-06,
"loss": 1.1815,
"step": 250
},
{
"epoch": 0.42796005706134094,
"grad_norm": 1.7196673154830933,
"learning_rate": 2.884186046511628e-06,
"loss": 1.1645,
"step": 300
},
{
"epoch": 0.4992867332382311,
"grad_norm": 1.5300871133804321,
"learning_rate": 2.8609302325581396e-06,
"loss": 1.1449,
"step": 350
},
{
"epoch": 0.5706134094151213,
"grad_norm": 1.5986438989639282,
"learning_rate": 2.837674418604651e-06,
"loss": 1.142,
"step": 400
},
{
"epoch": 0.6419400855920114,
"grad_norm": 1.7197072505950928,
"learning_rate": 2.8144186046511628e-06,
"loss": 1.135,
"step": 450
},
{
"epoch": 0.7132667617689016,
"grad_norm": 1.5155837535858154,
"learning_rate": 2.7911627906976743e-06,
"loss": 1.118,
"step": 500
},
{
"epoch": 0.7845934379457917,
"grad_norm": 1.4595504999160767,
"learning_rate": 2.767906976744186e-06,
"loss": 1.1237,
"step": 550
},
{
"epoch": 0.8559201141226819,
"grad_norm": 1.3575159311294556,
"learning_rate": 2.7446511627906975e-06,
"loss": 1.1367,
"step": 600
},
{
"epoch": 0.927246790299572,
"grad_norm": 1.3612666130065918,
"learning_rate": 2.7213953488372094e-06,
"loss": 1.1206,
"step": 650
},
{
"epoch": 0.9985734664764622,
"grad_norm": 1.3593813180923462,
"learning_rate": 2.698139534883721e-06,
"loss": 1.1385,
"step": 700
},
{
"epoch": 1.0699001426533523,
"grad_norm": 0.8913443088531494,
"learning_rate": 2.6748837209302326e-06,
"loss": 1.108,
"step": 750
},
{
"epoch": 1.1412268188302426,
"grad_norm": 0.9751181602478027,
"learning_rate": 2.651627906976744e-06,
"loss": 1.0941,
"step": 800
},
{
"epoch": 1.2125534950071326,
"grad_norm": 0.9726682901382446,
"learning_rate": 2.6283720930232557e-06,
"loss": 1.1093,
"step": 850
},
{
"epoch": 1.2838801711840229,
"grad_norm": 0.87285315990448,
"learning_rate": 2.6051162790697673e-06,
"loss": 1.135,
"step": 900
},
{
"epoch": 1.355206847360913,
"grad_norm": 0.8699735403060913,
"learning_rate": 2.581860465116279e-06,
"loss": 1.104,
"step": 950
},
{
"epoch": 1.4265335235378032,
"grad_norm": 1.0313692092895508,
"learning_rate": 2.558604651162791e-06,
"loss": 1.1197,
"step": 1000
},
{
"epoch": 1.4978601997146934,
"grad_norm": 0.9118053317070007,
"learning_rate": 2.5353488372093024e-06,
"loss": 1.096,
"step": 1050
},
{
"epoch": 1.5691868758915835,
"grad_norm": 0.9036098122596741,
"learning_rate": 2.512093023255814e-06,
"loss": 1.0938,
"step": 1100
},
{
"epoch": 1.6405135520684735,
"grad_norm": 1.0469361543655396,
"learning_rate": 2.4888372093023255e-06,
"loss": 1.093,
"step": 1150
},
{
"epoch": 1.7118402282453637,
"grad_norm": 0.9310563206672668,
"learning_rate": 2.4655813953488375e-06,
"loss": 1.1114,
"step": 1200
},
{
"epoch": 1.783166904422254,
"grad_norm": 0.8712385296821594,
"learning_rate": 2.442325581395349e-06,
"loss": 1.0992,
"step": 1250
},
{
"epoch": 1.854493580599144,
"grad_norm": 0.9832571148872375,
"learning_rate": 2.4190697674418606e-06,
"loss": 1.0896,
"step": 1300
},
{
"epoch": 1.925820256776034,
"grad_norm": 0.9631858468055725,
"learning_rate": 2.395813953488372e-06,
"loss": 1.0729,
"step": 1350
},
{
"epoch": 1.9971469329529246,
"grad_norm": 0.9085518717765808,
"learning_rate": 2.3725581395348838e-06,
"loss": 1.1032,
"step": 1400
},
{
"epoch": 2.0684736091298146,
"grad_norm": 0.7245323061943054,
"learning_rate": 2.3493023255813958e-06,
"loss": 1.1042,
"step": 1450
},
{
"epoch": 2.1398002853067046,
"grad_norm": 0.8104898929595947,
"learning_rate": 2.3260465116279073e-06,
"loss": 1.0829,
"step": 1500
},
{
"epoch": 2.2111269614835947,
"grad_norm": 0.7655696272850037,
"learning_rate": 2.302790697674419e-06,
"loss": 1.0701,
"step": 1550
},
{
"epoch": 2.282453637660485,
"grad_norm": 0.6674259305000305,
"learning_rate": 2.2795348837209304e-06,
"loss": 1.0831,
"step": 1600
},
{
"epoch": 2.353780313837375,
"grad_norm": 0.7704442739486694,
"learning_rate": 2.256279069767442e-06,
"loss": 1.0916,
"step": 1650
},
{
"epoch": 2.425106990014265,
"grad_norm": 0.9494684934616089,
"learning_rate": 2.2330232558139536e-06,
"loss": 1.0756,
"step": 1700
},
{
"epoch": 2.4964336661911553,
"grad_norm": 0.8056609630584717,
"learning_rate": 2.209767441860465e-06,
"loss": 1.0921,
"step": 1750
},
{
"epoch": 2.5677603423680457,
"grad_norm": 0.736752986907959,
"learning_rate": 2.1865116279069767e-06,
"loss": 1.1026,
"step": 1800
},
{
"epoch": 2.6390870185449358,
"grad_norm": 0.676659882068634,
"learning_rate": 2.1632558139534887e-06,
"loss": 1.127,
"step": 1850
},
{
"epoch": 2.710413694721826,
"grad_norm": 0.7117391228675842,
"learning_rate": 2.1400000000000003e-06,
"loss": 1.064,
"step": 1900
},
{
"epoch": 2.7817403708987163,
"grad_norm": 0.649662971496582,
"learning_rate": 2.116744186046512e-06,
"loss": 1.0887,
"step": 1950
},
{
"epoch": 2.8530670470756063,
"grad_norm": 0.7031757831573486,
"learning_rate": 2.0934883720930234e-06,
"loss": 1.084,
"step": 2000
},
{
"epoch": 2.9243937232524964,
"grad_norm": 0.8348050117492676,
"learning_rate": 2.070232558139535e-06,
"loss": 1.0606,
"step": 2050
},
{
"epoch": 2.995720399429387,
"grad_norm": 0.7338863015174866,
"learning_rate": 2.0469767441860465e-06,
"loss": 1.0748,
"step": 2100
},
{
"epoch": 3.067047075606277,
"grad_norm": 0.8033592700958252,
"learning_rate": 2.023720930232558e-06,
"loss": 1.0861,
"step": 2150
},
{
"epoch": 3.138373751783167,
"grad_norm": 0.7755851745605469,
"learning_rate": 2.0004651162790697e-06,
"loss": 1.0839,
"step": 2200
},
{
"epoch": 3.209700427960057,
"grad_norm": 0.7161579728126526,
"learning_rate": 1.9772093023255812e-06,
"loss": 1.0688,
"step": 2250
},
{
"epoch": 3.281027104136947,
"grad_norm": 0.854751706123352,
"learning_rate": 1.953953488372093e-06,
"loss": 1.0622,
"step": 2300
},
{
"epoch": 3.3523537803138375,
"grad_norm": 0.7753428220748901,
"learning_rate": 1.9306976744186048e-06,
"loss": 1.0677,
"step": 2350
},
{
"epoch": 3.4236804564907275,
"grad_norm": 0.7227015495300293,
"learning_rate": 1.9074418604651163e-06,
"loss": 1.0706,
"step": 2400
},
{
"epoch": 3.4950071326676175,
"grad_norm": 0.7708733677864075,
"learning_rate": 1.884186046511628e-06,
"loss": 1.0783,
"step": 2450
},
{
"epoch": 3.566333808844508,
"grad_norm": 0.7089836001396179,
"learning_rate": 1.8609302325581395e-06,
"loss": 1.0783,
"step": 2500
},
{
"epoch": 3.637660485021398,
"grad_norm": 0.7261233329772949,
"learning_rate": 1.837674418604651e-06,
"loss": 1.0997,
"step": 2550
},
{
"epoch": 3.708987161198288,
"grad_norm": 0.792870819568634,
"learning_rate": 1.8144186046511628e-06,
"loss": 1.0904,
"step": 2600
},
{
"epoch": 3.7803138373751786,
"grad_norm": 0.7912273406982422,
"learning_rate": 1.7911627906976744e-06,
"loss": 1.0609,
"step": 2650
},
{
"epoch": 3.8516405135520686,
"grad_norm": 0.8693161010742188,
"learning_rate": 1.7679069767441861e-06,
"loss": 1.0507,
"step": 2700
},
{
"epoch": 3.9229671897289586,
"grad_norm": 0.8193188905715942,
"learning_rate": 1.7446511627906977e-06,
"loss": 1.0802,
"step": 2750
},
{
"epoch": 3.9942938659058487,
"grad_norm": 0.8383765816688538,
"learning_rate": 1.7213953488372095e-06,
"loss": 1.0747,
"step": 2800
},
{
"epoch": 4.065620542082739,
"grad_norm": 0.6324388980865479,
"learning_rate": 1.698139534883721e-06,
"loss": 1.0643,
"step": 2850
},
{
"epoch": 4.136947218259629,
"grad_norm": 0.6263940930366516,
"learning_rate": 1.6748837209302326e-06,
"loss": 1.0707,
"step": 2900
},
{
"epoch": 4.20827389443652,
"grad_norm": 0.7516465187072754,
"learning_rate": 1.6516279069767442e-06,
"loss": 1.0723,
"step": 2950
},
{
"epoch": 4.279600570613409,
"grad_norm": 0.6298231482505798,
"learning_rate": 1.6283720930232558e-06,
"loss": 1.0902,
"step": 3000
},
{
"epoch": 4.3509272467903,
"grad_norm": 0.7128252387046814,
"learning_rate": 1.6051162790697673e-06,
"loss": 1.08,
"step": 3050
},
{
"epoch": 4.422253922967189,
"grad_norm": 0.8887324333190918,
"learning_rate": 1.5818604651162793e-06,
"loss": 1.064,
"step": 3100
},
{
"epoch": 4.49358059914408,
"grad_norm": 0.5942298769950867,
"learning_rate": 1.5586046511627909e-06,
"loss": 1.0748,
"step": 3150
},
{
"epoch": 4.56490727532097,
"grad_norm": 0.6519397497177124,
"learning_rate": 1.5353488372093024e-06,
"loss": 1.049,
"step": 3200
},
{
"epoch": 4.63623395149786,
"grad_norm": 0.8501943945884705,
"learning_rate": 1.512093023255814e-06,
"loss": 1.0689,
"step": 3250
},
{
"epoch": 4.70756062767475,
"grad_norm": 0.8405245542526245,
"learning_rate": 1.4888372093023256e-06,
"loss": 1.089,
"step": 3300
},
{
"epoch": 4.778887303851641,
"grad_norm": 0.8166652321815491,
"learning_rate": 1.4655813953488371e-06,
"loss": 1.0583,
"step": 3350
},
{
"epoch": 4.85021398002853,
"grad_norm": 0.6082524061203003,
"learning_rate": 1.442325581395349e-06,
"loss": 1.0721,
"step": 3400
},
{
"epoch": 4.921540656205421,
"grad_norm": 0.7854369282722473,
"learning_rate": 1.4190697674418605e-06,
"loss": 1.0536,
"step": 3450
},
{
"epoch": 4.9928673323823105,
"grad_norm": 0.662228524684906,
"learning_rate": 1.395813953488372e-06,
"loss": 1.0562,
"step": 3500
},
{
"epoch": 5.064194008559201,
"grad_norm": 0.5415034294128418,
"learning_rate": 1.3725581395348836e-06,
"loss": 1.0698,
"step": 3550
},
{
"epoch": 5.1355206847360915,
"grad_norm": 0.5144311785697937,
"learning_rate": 1.3493023255813954e-06,
"loss": 1.0739,
"step": 3600
},
{
"epoch": 5.206847360912981,
"grad_norm": 0.6178659796714783,
"learning_rate": 1.326046511627907e-06,
"loss": 1.0592,
"step": 3650
},
{
"epoch": 5.2781740370898715,
"grad_norm": 0.625819981098175,
"learning_rate": 1.3027906976744185e-06,
"loss": 1.0627,
"step": 3700
},
{
"epoch": 5.349500713266762,
"grad_norm": 0.605794370174408,
"learning_rate": 1.2795348837209303e-06,
"loss": 1.0903,
"step": 3750
},
{
"epoch": 5.420827389443652,
"grad_norm": 0.5412514209747314,
"learning_rate": 1.2562790697674418e-06,
"loss": 1.038,
"step": 3800
},
{
"epoch": 5.492154065620542,
"grad_norm": 0.5697770118713379,
"learning_rate": 1.2330232558139536e-06,
"loss": 1.0812,
"step": 3850
},
{
"epoch": 5.563480741797433,
"grad_norm": 0.6216000318527222,
"learning_rate": 1.2097674418604652e-06,
"loss": 1.0546,
"step": 3900
},
{
"epoch": 5.634807417974322,
"grad_norm": 0.6266453266143799,
"learning_rate": 1.1865116279069768e-06,
"loss": 1.0684,
"step": 3950
},
{
"epoch": 5.706134094151213,
"grad_norm": 0.5270595550537109,
"learning_rate": 1.1632558139534885e-06,
"loss": 1.0808,
"step": 4000
},
{
"epoch": 5.777460770328103,
"grad_norm": 0.5590298771858215,
"learning_rate": 1.14e-06,
"loss": 1.0685,
"step": 4050
},
{
"epoch": 5.848787446504993,
"grad_norm": 0.5806601643562317,
"learning_rate": 1.1167441860465117e-06,
"loss": 1.0524,
"step": 4100
},
{
"epoch": 5.920114122681883,
"grad_norm": 0.595583975315094,
"learning_rate": 1.0934883720930232e-06,
"loss": 1.0502,
"step": 4150
},
{
"epoch": 5.991440798858774,
"grad_norm": 0.5921752452850342,
"learning_rate": 1.070232558139535e-06,
"loss": 1.0579,
"step": 4200
},
{
"epoch": 6.062767475035663,
"grad_norm": 0.49118900299072266,
"learning_rate": 1.0469767441860466e-06,
"loss": 1.0414,
"step": 4250
},
{
"epoch": 6.134094151212554,
"grad_norm": 0.5080968141555786,
"learning_rate": 1.0237209302325581e-06,
"loss": 1.0669,
"step": 4300
},
{
"epoch": 6.205420827389443,
"grad_norm": 0.5190466642379761,
"learning_rate": 1.0004651162790697e-06,
"loss": 1.0815,
"step": 4350
},
{
"epoch": 6.276747503566334,
"grad_norm": 0.571312427520752,
"learning_rate": 9.772093023255815e-07,
"loss": 1.0586,
"step": 4400
},
{
"epoch": 6.348074179743224,
"grad_norm": 0.5268988013267517,
"learning_rate": 9.53953488372093e-07,
"loss": 1.0493,
"step": 4450
},
{
"epoch": 6.419400855920114,
"grad_norm": 0.5871033072471619,
"learning_rate": 9.306976744186046e-07,
"loss": 1.0639,
"step": 4500
},
{
"epoch": 6.490727532097004,
"grad_norm": 0.49360474944114685,
"learning_rate": 9.074418604651163e-07,
"loss": 1.0544,
"step": 4550
},
{
"epoch": 6.562054208273894,
"grad_norm": 0.5026892423629761,
"learning_rate": 8.841860465116279e-07,
"loss": 1.0788,
"step": 4600
},
{
"epoch": 6.633380884450784,
"grad_norm": 0.6054779291152954,
"learning_rate": 8.609302325581396e-07,
"loss": 1.0689,
"step": 4650
},
{
"epoch": 6.704707560627675,
"grad_norm": 0.5217434167861938,
"learning_rate": 8.376744186046512e-07,
"loss": 1.051,
"step": 4700
},
{
"epoch": 6.7760342368045645,
"grad_norm": 0.47993195056915283,
"learning_rate": 8.144186046511627e-07,
"loss": 1.0578,
"step": 4750
},
{
"epoch": 6.847360912981455,
"grad_norm": 0.5778261423110962,
"learning_rate": 7.911627906976745e-07,
"loss": 1.0706,
"step": 4800
},
{
"epoch": 6.9186875891583455,
"grad_norm": 0.5452645421028137,
"learning_rate": 7.679069767441861e-07,
"loss": 1.0663,
"step": 4850
},
{
"epoch": 6.990014265335235,
"grad_norm": 0.5666407346725464,
"learning_rate": 7.446511627906977e-07,
"loss": 1.0599,
"step": 4900
},
{
"epoch": 7.0613409415121255,
"grad_norm": 0.5317495465278625,
"learning_rate": 7.213953488372093e-07,
"loss": 1.0537,
"step": 4950
},
{
"epoch": 7.132667617689016,
"grad_norm": 0.5068672299385071,
"learning_rate": 6.981395348837209e-07,
"loss": 1.0779,
"step": 5000
},
{
"epoch": 7.203994293865906,
"grad_norm": 0.5887683629989624,
"learning_rate": 6.748837209302326e-07,
"loss": 1.057,
"step": 5050
},
{
"epoch": 7.275320970042796,
"grad_norm": 0.4751645028591156,
"learning_rate": 6.516279069767442e-07,
"loss": 1.0531,
"step": 5100
},
{
"epoch": 7.346647646219687,
"grad_norm": 0.4856860935688019,
"learning_rate": 6.283720930232559e-07,
"loss": 1.0575,
"step": 5150
},
{
"epoch": 7.417974322396576,
"grad_norm": 0.5266442894935608,
"learning_rate": 6.051162790697675e-07,
"loss": 1.0607,
"step": 5200
},
{
"epoch": 7.489300998573467,
"grad_norm": 0.5558798909187317,
"learning_rate": 5.818604651162791e-07,
"loss": 1.0653,
"step": 5250
},
{
"epoch": 7.560627674750357,
"grad_norm": 0.5203211307525635,
"learning_rate": 5.586046511627907e-07,
"loss": 1.0583,
"step": 5300
},
{
"epoch": 7.631954350927247,
"grad_norm": 0.4732428789138794,
"learning_rate": 5.353488372093024e-07,
"loss": 1.0534,
"step": 5350
},
{
"epoch": 7.703281027104137,
"grad_norm": 0.6876837015151978,
"learning_rate": 5.120930232558139e-07,
"loss": 1.0562,
"step": 5400
},
{
"epoch": 7.774607703281027,
"grad_norm": 0.5170673727989197,
"learning_rate": 4.888372093023256e-07,
"loss": 1.066,
"step": 5450
},
{
"epoch": 7.845934379457917,
"grad_norm": 0.5236818790435791,
"learning_rate": 4.6558139534883723e-07,
"loss": 1.0644,
"step": 5500
},
{
"epoch": 7.917261055634808,
"grad_norm": 0.47422999143600464,
"learning_rate": 4.423255813953489e-07,
"loss": 1.0594,
"step": 5550
},
{
"epoch": 7.988587731811697,
"grad_norm": 0.5537230968475342,
"learning_rate": 4.1906976744186046e-07,
"loss": 1.0578,
"step": 5600
},
{
"epoch": 8.059914407988588,
"grad_norm": 0.5376424789428711,
"learning_rate": 3.9581395348837213e-07,
"loss": 1.0523,
"step": 5650
},
{
"epoch": 8.131241084165477,
"grad_norm": 0.4635964334011078,
"learning_rate": 3.725581395348837e-07,
"loss": 1.0559,
"step": 5700
},
{
"epoch": 8.202567760342369,
"grad_norm": 0.46095508337020874,
"learning_rate": 3.493023255813953e-07,
"loss": 1.0825,
"step": 5750
},
{
"epoch": 8.273894436519258,
"grad_norm": 0.48418450355529785,
"learning_rate": 3.26046511627907e-07,
"loss": 1.0541,
"step": 5800
},
{
"epoch": 8.345221112696148,
"grad_norm": 0.45845749974250793,
"learning_rate": 3.027906976744186e-07,
"loss": 1.0731,
"step": 5850
},
{
"epoch": 8.41654778887304,
"grad_norm": 0.5616029500961304,
"learning_rate": 2.795348837209302e-07,
"loss": 1.055,
"step": 5900
},
{
"epoch": 8.487874465049929,
"grad_norm": 0.5239433646202087,
"learning_rate": 2.5627906976744184e-07,
"loss": 1.0522,
"step": 5950
},
{
"epoch": 8.559201141226819,
"grad_norm": 0.46216270327568054,
"learning_rate": 2.3302325581395349e-07,
"loss": 1.0674,
"step": 6000
}
],
"logging_steps": 50,
"max_steps": 6500,
"num_input_tokens_seen": 0,
"num_train_epochs": 10,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.4510459760115139e+18,
"train_batch_size": 48,
"trial_name": null,
"trial_params": null
}