Kudod's picture
Training in progress, step 60000, checkpoint
4704419 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.989190088142358,
"eval_steps": 10000,
"global_step": 60000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.04157658406785299,
"grad_norm": 0.8059402108192444,
"learning_rate": 0.00025,
"loss": 7.9387,
"step": 500
},
{
"epoch": 0.08315316813570597,
"grad_norm": 0.18919387459754944,
"learning_rate": 0.0005,
"loss": 7.951,
"step": 1000
},
{
"epoch": 0.12472975220355896,
"grad_norm": 0.4990023076534271,
"learning_rate": 0.000495772027735498,
"loss": 7.9528,
"step": 1500
},
{
"epoch": 0.16630633627141195,
"grad_norm": 0.5437057018280029,
"learning_rate": 0.0004915440554709961,
"loss": 7.9422,
"step": 2000
},
{
"epoch": 0.20788292033926492,
"grad_norm": 0.35262131690979004,
"learning_rate": 0.00048731608320649417,
"loss": 7.9496,
"step": 2500
},
{
"epoch": 0.24945950440711792,
"grad_norm": 1.5567623376846313,
"learning_rate": 0.00048308811094199225,
"loss": 7.9473,
"step": 3000
},
{
"epoch": 0.2910360884749709,
"grad_norm": 1.070528507232666,
"learning_rate": 0.0004788601386774903,
"loss": 7.9368,
"step": 3500
},
{
"epoch": 0.3326126725428239,
"grad_norm": 0.2238803505897522,
"learning_rate": 0.0004746321664129883,
"loss": 7.9446,
"step": 4000
},
{
"epoch": 0.37418925661067687,
"grad_norm": 0.6900185346603394,
"learning_rate": 0.0004704041941484864,
"loss": 7.9435,
"step": 4500
},
{
"epoch": 0.41576584067852984,
"grad_norm": 0.4404136538505554,
"learning_rate": 0.00046617622188398444,
"loss": 7.9488,
"step": 5000
},
{
"epoch": 0.4573424247463828,
"grad_norm": 0.3816683888435364,
"learning_rate": 0.00046195670556401153,
"loss": 7.9463,
"step": 5500
},
{
"epoch": 0.49891900881423584,
"grad_norm": 0.41241884231567383,
"learning_rate": 0.00045772873329950957,
"loss": 7.9452,
"step": 6000
},
{
"epoch": 0.5404955928820888,
"grad_norm": 0.8631793260574341,
"learning_rate": 0.00045350076103500765,
"loss": 7.94,
"step": 6500
},
{
"epoch": 0.5820721769499418,
"grad_norm": 0.44780728220939636,
"learning_rate": 0.00044927278877050563,
"loss": 7.9472,
"step": 7000
},
{
"epoch": 0.6236487610177948,
"grad_norm": 0.2136773020029068,
"learning_rate": 0.0004450617283950618,
"loss": 7.9409,
"step": 7500
},
{
"epoch": 0.6652253450856478,
"grad_norm": 0.2820763885974884,
"learning_rate": 0.0004408337561305598,
"loss": 7.9483,
"step": 8000
},
{
"epoch": 0.7068019291535007,
"grad_norm": 0.694266140460968,
"learning_rate": 0.00043660578386605785,
"loss": 7.9427,
"step": 8500
},
{
"epoch": 0.7483785132213537,
"grad_norm": 1.1062744855880737,
"learning_rate": 0.00043237781160155594,
"loss": 7.9448,
"step": 9000
},
{
"epoch": 0.7899550972892068,
"grad_norm": 0.6461020708084106,
"learning_rate": 0.00042814983933705397,
"loss": 7.9392,
"step": 9500
},
{
"epoch": 0.8315316813570597,
"grad_norm": 0.7867754101753235,
"learning_rate": 0.00042392186707255206,
"loss": 7.9435,
"step": 10000
},
{
"epoch": 0.8315316813570597,
"eval_loss": 9.379227638244629,
"eval_runtime": 2903.9955,
"eval_samples_per_second": 132.515,
"eval_steps_per_second": 4.141,
"step": 10000
},
{
"epoch": 0.8731082654249127,
"grad_norm": 0.5909550786018372,
"learning_rate": 0.00041970235075257904,
"loss": 7.948,
"step": 10500
},
{
"epoch": 0.9146848494927656,
"grad_norm": 0.24700529873371124,
"learning_rate": 0.00041547437848807713,
"loss": 7.9407,
"step": 11000
},
{
"epoch": 0.9562614335606187,
"grad_norm": 0.3503686487674713,
"learning_rate": 0.00041124640622357516,
"loss": 7.9396,
"step": 11500
},
{
"epoch": 0.9978380176284717,
"grad_norm": 0.4738515019416809,
"learning_rate": 0.00040701843395907325,
"loss": 7.9312,
"step": 12000
},
{
"epoch": 1.0394146016963246,
"grad_norm": 0.32388758659362793,
"learning_rate": 0.0004027904616945713,
"loss": 7.9352,
"step": 12500
},
{
"epoch": 1.0809911857641776,
"grad_norm": 1.2330721616744995,
"learning_rate": 0.0003985624894300693,
"loss": 7.9493,
"step": 13000
},
{
"epoch": 1.1225677698320307,
"grad_norm": 0.21379995346069336,
"learning_rate": 0.0003943345171655674,
"loss": 7.9334,
"step": 13500
},
{
"epoch": 1.1641443538998835,
"grad_norm": 0.3356311023235321,
"learning_rate": 0.00039010654490106544,
"loss": 7.9524,
"step": 14000
},
{
"epoch": 1.2057209379677365,
"grad_norm": 0.69386225938797,
"learning_rate": 0.00038587857263656353,
"loss": 7.9395,
"step": 14500
},
{
"epoch": 1.2472975220355895,
"grad_norm": 0.3345106542110443,
"learning_rate": 0.00038165060037206156,
"loss": 7.9449,
"step": 15000
},
{
"epoch": 1.2888741061034426,
"grad_norm": 0.8065163493156433,
"learning_rate": 0.00037743108405208865,
"loss": 7.945,
"step": 15500
},
{
"epoch": 1.3304506901712956,
"grad_norm": 0.5001415014266968,
"learning_rate": 0.0003732115677321157,
"loss": 7.9435,
"step": 16000
},
{
"epoch": 1.3720272742391484,
"grad_norm": 1.0568392276763916,
"learning_rate": 0.0003689835954676137,
"loss": 7.9435,
"step": 16500
},
{
"epoch": 1.4136038583070014,
"grad_norm": 0.36474671959877014,
"learning_rate": 0.0003647556232031118,
"loss": 7.9396,
"step": 17000
},
{
"epoch": 1.4551804423748544,
"grad_norm": 0.3686589300632477,
"learning_rate": 0.00036053610688313885,
"loss": 7.9323,
"step": 17500
},
{
"epoch": 1.4967570264427075,
"grad_norm": 0.8043748736381531,
"learning_rate": 0.0003563165905631659,
"loss": 7.9493,
"step": 18000
},
{
"epoch": 1.5383336105105605,
"grad_norm": 0.17725110054016113,
"learning_rate": 0.000352097074243193,
"loss": 7.9321,
"step": 18500
},
{
"epoch": 1.5799101945784133,
"grad_norm": 0.6144613027572632,
"learning_rate": 0.00034786910197869106,
"loss": 7.9382,
"step": 19000
},
{
"epoch": 1.6214867786462666,
"grad_norm": 0.3076309859752655,
"learning_rate": 0.0003436411297141891,
"loss": 7.9424,
"step": 19500
},
{
"epoch": 1.6630633627141194,
"grad_norm": 0.45199695229530334,
"learning_rate": 0.00033941315744968713,
"loss": 7.9344,
"step": 20000
},
{
"epoch": 1.6630633627141194,
"eval_loss": 9.162654876708984,
"eval_runtime": 2896.236,
"eval_samples_per_second": 132.87,
"eval_steps_per_second": 4.152,
"step": 20000
},
{
"epoch": 1.7046399467819724,
"grad_norm": 0.6658689975738525,
"learning_rate": 0.0003351851851851852,
"loss": 7.9352,
"step": 20500
},
{
"epoch": 1.7462165308498254,
"grad_norm": 0.5613082051277161,
"learning_rate": 0.00033095721292068325,
"loss": 7.9292,
"step": 21000
},
{
"epoch": 1.7877931149176782,
"grad_norm": 0.38716328144073486,
"learning_rate": 0.00032672924065618134,
"loss": 7.9327,
"step": 21500
},
{
"epoch": 1.8293696989855315,
"grad_norm": 0.7330873608589172,
"learning_rate": 0.00032250126839167937,
"loss": 7.928,
"step": 22000
},
{
"epoch": 1.8709462830533843,
"grad_norm": 0.28898411989212036,
"learning_rate": 0.0003182986639607644,
"loss": 7.9521,
"step": 22500
},
{
"epoch": 1.9125228671212373,
"grad_norm": 0.7989892959594727,
"learning_rate": 0.00031407069169626245,
"loss": 7.9357,
"step": 23000
},
{
"epoch": 1.9540994511890903,
"grad_norm": 2.788367509841919,
"learning_rate": 0.00030985117537628954,
"loss": 7.9401,
"step": 23500
},
{
"epoch": 1.9956760352569431,
"grad_norm": 0.5419213175773621,
"learning_rate": 0.00030562320311178757,
"loss": 7.9396,
"step": 24000
},
{
"epoch": 2.0372526193247964,
"grad_norm": 0.6582406759262085,
"learning_rate": 0.00030139523084728566,
"loss": 7.9346,
"step": 24500
},
{
"epoch": 2.078829203392649,
"grad_norm": 0.5304723381996155,
"learning_rate": 0.0002971672585827837,
"loss": 7.9344,
"step": 25000
},
{
"epoch": 2.1204057874605025,
"grad_norm": 0.5516705513000488,
"learning_rate": 0.0002929392863182818,
"loss": 7.9274,
"step": 25500
},
{
"epoch": 2.1619823715283553,
"grad_norm": 1.1628555059432983,
"learning_rate": 0.00028871976999830876,
"loss": 7.9465,
"step": 26000
},
{
"epoch": 2.203558955596208,
"grad_norm": 0.3907375931739807,
"learning_rate": 0.00028449179773380685,
"loss": 7.9391,
"step": 26500
},
{
"epoch": 2.2451355396640613,
"grad_norm": 1.1125935316085815,
"learning_rate": 0.0002802638254693049,
"loss": 7.9292,
"step": 27000
},
{
"epoch": 2.286712123731914,
"grad_norm": 0.8965011835098267,
"learning_rate": 0.000276035853204803,
"loss": 7.9489,
"step": 27500
},
{
"epoch": 2.328288707799767,
"grad_norm": 0.9403777122497559,
"learning_rate": 0.000271807880940301,
"loss": 7.9436,
"step": 28000
},
{
"epoch": 2.36986529186762,
"grad_norm": 0.8826812505722046,
"learning_rate": 0.00026757990867579904,
"loss": 7.9383,
"step": 28500
},
{
"epoch": 2.411441875935473,
"grad_norm": 1.86482572555542,
"learning_rate": 0.00026335193641129713,
"loss": 7.9397,
"step": 29000
},
{
"epoch": 2.4530184600033262,
"grad_norm": 0.2199900895357132,
"learning_rate": 0.00025912396414679516,
"loss": 7.9301,
"step": 29500
},
{
"epoch": 2.494595044071179,
"grad_norm": 4.152510166168213,
"learning_rate": 0.00025489599188229325,
"loss": 7.9276,
"step": 30000
},
{
"epoch": 2.494595044071179,
"eval_loss": 9.38607406616211,
"eval_runtime": 19770.2885,
"eval_samples_per_second": 19.465,
"eval_steps_per_second": 0.608,
"step": 30000
},
{
"epoch": 2.536171628139032,
"grad_norm": 1.1997425556182861,
"learning_rate": 0.0002506680196177913,
"loss": 7.9276,
"step": 30500
},
{
"epoch": 2.577748212206885,
"grad_norm": 1.2270859479904175,
"learning_rate": 0.00024644004735328937,
"loss": 7.924,
"step": 31000
},
{
"epoch": 2.619324796274738,
"grad_norm": 0.8749690651893616,
"learning_rate": 0.00024221207508878743,
"loss": 7.9268,
"step": 31500
},
{
"epoch": 2.660901380342591,
"grad_norm": 1.2462141513824463,
"learning_rate": 0.00023799255876881447,
"loss": 7.9462,
"step": 32000
},
{
"epoch": 2.702477964410444,
"grad_norm": 0.6699239015579224,
"learning_rate": 0.00023376458650431253,
"loss": 7.9261,
"step": 32500
},
{
"epoch": 2.7440545484782968,
"grad_norm": 0.8309171795845032,
"learning_rate": 0.0002295450701843396,
"loss": 7.9316,
"step": 33000
},
{
"epoch": 2.78563113254615,
"grad_norm": 1.003183364868164,
"learning_rate": 0.00022531709791983765,
"loss": 7.9262,
"step": 33500
},
{
"epoch": 2.827207716614003,
"grad_norm": 0.7003466486930847,
"learning_rate": 0.00022108912565533572,
"loss": 7.9384,
"step": 34000
},
{
"epoch": 2.868784300681856,
"grad_norm": 0.6416522264480591,
"learning_rate": 0.00021686115339083375,
"loss": 7.9367,
"step": 34500
},
{
"epoch": 2.910360884749709,
"grad_norm": 0.7745890021324158,
"learning_rate": 0.0002126331811263318,
"loss": 7.9493,
"step": 35000
},
{
"epoch": 2.9519374688175617,
"grad_norm": 2.5929224491119385,
"learning_rate": 0.00020840520886182987,
"loss": 7.939,
"step": 35500
},
{
"epoch": 2.993514052885415,
"grad_norm": 0.6639001965522766,
"learning_rate": 0.00020417723659732793,
"loss": 7.932,
"step": 36000
},
{
"epoch": 3.0350906369532678,
"grad_norm": 4.156096935272217,
"learning_rate": 0.000199949264332826,
"loss": 7.9312,
"step": 36500
},
{
"epoch": 3.076667221021121,
"grad_norm": 0.6016917824745178,
"learning_rate": 0.00019572129206832403,
"loss": 7.9321,
"step": 37000
},
{
"epoch": 3.118243805088974,
"grad_norm": 0.6136893630027771,
"learning_rate": 0.0001914933198038221,
"loss": 7.934,
"step": 37500
},
{
"epoch": 3.159820389156827,
"grad_norm": 1.5245190858840942,
"learning_rate": 0.00018726534753932015,
"loss": 7.9371,
"step": 38000
},
{
"epoch": 3.20139697322468,
"grad_norm": 1.779859185218811,
"learning_rate": 0.0001830373752748182,
"loss": 7.9355,
"step": 38500
},
{
"epoch": 3.2429735572925327,
"grad_norm": 0.7921754717826843,
"learning_rate": 0.00017880940301031627,
"loss": 7.9297,
"step": 39000
},
{
"epoch": 3.284550141360386,
"grad_norm": 0.40482184290885925,
"learning_rate": 0.0001745814307458143,
"loss": 7.943,
"step": 39500
},
{
"epoch": 3.3261267254282387,
"grad_norm": 0.950188398361206,
"learning_rate": 0.00017036191442584137,
"loss": 7.9307,
"step": 40000
},
{
"epoch": 3.3261267254282387,
"eval_loss": 8.992598533630371,
"eval_runtime": 19781.8489,
"eval_samples_per_second": 19.453,
"eval_steps_per_second": 0.608,
"step": 40000
},
{
"epoch": 3.367703309496092,
"grad_norm": 3.409175157546997,
"learning_rate": 0.00016614239810586843,
"loss": 7.9407,
"step": 40500
},
{
"epoch": 3.409279893563945,
"grad_norm": 0.41312387585639954,
"learning_rate": 0.0001619228817858955,
"loss": 7.9386,
"step": 41000
},
{
"epoch": 3.4508564776317976,
"grad_norm": 0.1976231038570404,
"learning_rate": 0.00015769490952139353,
"loss": 7.9346,
"step": 41500
},
{
"epoch": 3.492433061699651,
"grad_norm": 0.5483559966087341,
"learning_rate": 0.0001534669372568916,
"loss": 7.9234,
"step": 42000
},
{
"epoch": 3.5340096457675036,
"grad_norm": 0.7013275027275085,
"learning_rate": 0.00014923896499238965,
"loss": 7.9295,
"step": 42500
},
{
"epoch": 3.5755862298353565,
"grad_norm": 1.0993248224258423,
"learning_rate": 0.0001450109927278877,
"loss": 7.9313,
"step": 43000
},
{
"epoch": 3.6171628139032097,
"grad_norm": 0.9102666974067688,
"learning_rate": 0.00014079147640791475,
"loss": 7.9169,
"step": 43500
},
{
"epoch": 3.658739397971063,
"grad_norm": 0.3310043215751648,
"learning_rate": 0.0001365635041434128,
"loss": 7.9331,
"step": 44000
},
{
"epoch": 3.7003159820389158,
"grad_norm": 1.6578195095062256,
"learning_rate": 0.00013233553187891087,
"loss": 7.9408,
"step": 44500
},
{
"epoch": 3.7418925661067686,
"grad_norm": 1.0133479833602905,
"learning_rate": 0.00012810755961440893,
"loss": 7.9278,
"step": 45000
},
{
"epoch": 3.7834691501746214,
"grad_norm": 1.0534777641296387,
"learning_rate": 0.000123879587349907,
"loss": 7.9222,
"step": 45500
},
{
"epoch": 3.8250457342424746,
"grad_norm": 1.0995186567306519,
"learning_rate": 0.00011965161508540504,
"loss": 7.9222,
"step": 46000
},
{
"epoch": 3.866622318310328,
"grad_norm": 1.2467023134231567,
"learning_rate": 0.00011542364282090308,
"loss": 7.9374,
"step": 46500
},
{
"epoch": 3.9081989023781807,
"grad_norm": 0.6253928542137146,
"learning_rate": 0.00011119567055640114,
"loss": 7.9241,
"step": 47000
},
{
"epoch": 3.9497754864460335,
"grad_norm": 0.3281259536743164,
"learning_rate": 0.0001069676982918992,
"loss": 7.9298,
"step": 47500
},
{
"epoch": 3.9913520705138867,
"grad_norm": 0.6171555519104004,
"learning_rate": 0.00010273972602739725,
"loss": 7.9352,
"step": 48000
},
{
"epoch": 4.03292865458174,
"grad_norm": 0.41333335638046265,
"learning_rate": 9.851175376289531e-05,
"loss": 7.9356,
"step": 48500
},
{
"epoch": 4.074505238649593,
"grad_norm": 0.2815527021884918,
"learning_rate": 9.428378149839336e-05,
"loss": 7.9241,
"step": 49000
},
{
"epoch": 4.116081822717446,
"grad_norm": 0.8225882649421692,
"learning_rate": 9.005580923389142e-05,
"loss": 7.9214,
"step": 49500
},
{
"epoch": 4.157658406785298,
"grad_norm": 0.44963425397872925,
"learning_rate": 8.582783696938948e-05,
"loss": 7.9237,
"step": 50000
},
{
"epoch": 4.157658406785298,
"eval_loss": 8.952984809875488,
"eval_runtime": 2779.282,
"eval_samples_per_second": 138.462,
"eval_steps_per_second": 4.327,
"step": 50000
},
{
"epoch": 4.199234990853151,
"grad_norm": 0.5252857208251953,
"learning_rate": 8.159986470488753e-05,
"loss": 7.9331,
"step": 50500
},
{
"epoch": 4.240811574921005,
"grad_norm": 0.4898480176925659,
"learning_rate": 7.737189244038559e-05,
"loss": 7.9296,
"step": 51000
},
{
"epoch": 4.282388158988858,
"grad_norm": 0.597152054309845,
"learning_rate": 7.314392017588364e-05,
"loss": 7.9371,
"step": 51500
},
{
"epoch": 4.3239647430567105,
"grad_norm": 0.3579633831977844,
"learning_rate": 6.89159479113817e-05,
"loss": 7.92,
"step": 52000
},
{
"epoch": 4.365541327124563,
"grad_norm": 0.8714535236358643,
"learning_rate": 6.468797564687976e-05,
"loss": 7.9259,
"step": 52500
},
{
"epoch": 4.407117911192416,
"grad_norm": 0.4833739101886749,
"learning_rate": 6.046000338237781e-05,
"loss": 7.93,
"step": 53000
},
{
"epoch": 4.44869449526027,
"grad_norm": 0.7422254681587219,
"learning_rate": 5.623203111787587e-05,
"loss": 7.9443,
"step": 53500
},
{
"epoch": 4.490271079328123,
"grad_norm": 0.3269445598125458,
"learning_rate": 5.201251479790293e-05,
"loss": 7.9211,
"step": 54000
},
{
"epoch": 4.5318476633959754,
"grad_norm": 0.3518795073032379,
"learning_rate": 4.779299847792998e-05,
"loss": 7.9327,
"step": 54500
},
{
"epoch": 4.573424247463828,
"grad_norm": 0.29840198159217834,
"learning_rate": 4.356502621342804e-05,
"loss": 7.9341,
"step": 55000
},
{
"epoch": 4.615000831531681,
"grad_norm": 0.6424156427383423,
"learning_rate": 3.9337053948926096e-05,
"loss": 7.943,
"step": 55500
},
{
"epoch": 4.656577415599534,
"grad_norm": 0.2546742260456085,
"learning_rate": 3.511753762895315e-05,
"loss": 7.9313,
"step": 56000
},
{
"epoch": 4.698153999667388,
"grad_norm": 0.5346771478652954,
"learning_rate": 3.0889565364451214e-05,
"loss": 7.9275,
"step": 56500
},
{
"epoch": 4.73973058373524,
"grad_norm": 0.7033748030662537,
"learning_rate": 2.6661593099949264e-05,
"loss": 7.9358,
"step": 57000
},
{
"epoch": 4.781307167803093,
"grad_norm": 0.28617945313453674,
"learning_rate": 2.243362083544732e-05,
"loss": 7.9299,
"step": 57500
},
{
"epoch": 4.822883751870946,
"grad_norm": 0.5841323733329773,
"learning_rate": 1.8205648570945376e-05,
"loss": 7.9234,
"step": 58000
},
{
"epoch": 4.8644603359388,
"grad_norm": 0.4925735890865326,
"learning_rate": 1.397767630644343e-05,
"loss": 7.9181,
"step": 58500
},
{
"epoch": 4.9060369200066525,
"grad_norm": 0.7338876128196716,
"learning_rate": 9.749704041941485e-06,
"loss": 7.9225,
"step": 59000
},
{
"epoch": 4.947613504074505,
"grad_norm": 0.4279155731201172,
"learning_rate": 5.52173177743954e-06,
"loss": 7.9266,
"step": 59500
},
{
"epoch": 4.989190088142358,
"grad_norm": 0.9879505038261414,
"learning_rate": 1.2937595129375953e-06,
"loss": 7.9321,
"step": 60000
},
{
"epoch": 4.989190088142358,
"eval_loss": 8.944025993347168,
"eval_runtime": 2914.4582,
"eval_samples_per_second": 132.04,
"eval_steps_per_second": 4.126,
"step": 60000
}
],
"logging_steps": 500,
"max_steps": 60130,
"num_input_tokens_seen": 0,
"num_train_epochs": 5,
"save_steps": 10000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 6.726369615519345e+17,
"train_batch_size": 32,
"trial_name": null,
"trial_params": null
}