{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 3.0, "eval_steps": 500, "global_step": 1629, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01841620626151013, "grad_norm": 3.54826512795615, "learning_rate": 5.521472392638038e-07, "loss": 0.4164, "step": 10 }, { "epoch": 0.03683241252302026, "grad_norm": 2.150976325911803, "learning_rate": 1.165644171779141e-06, "loss": 0.3752, "step": 20 }, { "epoch": 0.055248618784530384, "grad_norm": 1.301734345190514, "learning_rate": 1.7791411042944787e-06, "loss": 0.3446, "step": 30 }, { "epoch": 0.07366482504604052, "grad_norm": 1.0129957906382288, "learning_rate": 2.392638036809816e-06, "loss": 0.3248, "step": 40 }, { "epoch": 0.09208103130755065, "grad_norm": 1.0713399309035538, "learning_rate": 3.0061349693251535e-06, "loss": 0.3026, "step": 50 }, { "epoch": 0.11049723756906077, "grad_norm": 1.0507425813581446, "learning_rate": 3.6196319018404913e-06, "loss": 0.304, "step": 60 }, { "epoch": 0.1289134438305709, "grad_norm": 0.8381317525552241, "learning_rate": 4.233128834355829e-06, "loss": 0.291, "step": 70 }, { "epoch": 0.14732965009208104, "grad_norm": 0.9207152395423905, "learning_rate": 4.846625766871166e-06, "loss": 0.2962, "step": 80 }, { "epoch": 0.16574585635359115, "grad_norm": 0.9703285578071639, "learning_rate": 5.460122699386503e-06, "loss": 0.2877, "step": 90 }, { "epoch": 0.1841620626151013, "grad_norm": 0.9941438616319775, "learning_rate": 6.073619631901841e-06, "loss": 0.2894, "step": 100 }, { "epoch": 0.20257826887661143, "grad_norm": 1.0580118824781048, "learning_rate": 6.687116564417178e-06, "loss": 0.2865, "step": 110 }, { "epoch": 0.22099447513812154, "grad_norm": 1.0527321061745536, "learning_rate": 7.300613496932516e-06, "loss": 0.2744, "step": 120 }, { "epoch": 0.23941068139963168, "grad_norm": 0.9446407682756467, "learning_rate": 7.914110429447854e-06, "loss": 0.2844, "step": 130 }, { "epoch": 0.2578268876611418, "grad_norm": 1.0199363774215442, "learning_rate": 8.527607361963191e-06, "loss": 0.2793, "step": 140 }, { "epoch": 0.27624309392265195, "grad_norm": 0.9261191382922327, "learning_rate": 9.14110429447853e-06, "loss": 0.2822, "step": 150 }, { "epoch": 0.2946593001841621, "grad_norm": 0.9028861540282759, "learning_rate": 9.754601226993867e-06, "loss": 0.2758, "step": 160 }, { "epoch": 0.31307550644567217, "grad_norm": 0.8298039066694322, "learning_rate": 9.999586697215748e-06, "loss": 0.2867, "step": 170 }, { "epoch": 0.3314917127071823, "grad_norm": 0.9401043712912456, "learning_rate": 9.997061205416203e-06, "loss": 0.2935, "step": 180 }, { "epoch": 0.34990791896869244, "grad_norm": 0.8170942120374912, "learning_rate": 9.992240992810445e-06, "loss": 0.276, "step": 190 }, { "epoch": 0.3683241252302026, "grad_norm": 0.7970942467849401, "learning_rate": 9.985128272907917e-06, "loss": 0.2797, "step": 200 }, { "epoch": 0.3867403314917127, "grad_norm": 0.8675972760036051, "learning_rate": 9.975726311969664e-06, "loss": 0.2782, "step": 210 }, { "epoch": 0.40515653775322286, "grad_norm": 0.8765243656942432, "learning_rate": 9.964039427508418e-06, "loss": 0.2815, "step": 220 }, { "epoch": 0.42357274401473294, "grad_norm": 0.8649203720855488, "learning_rate": 9.950072986305938e-06, "loss": 0.2757, "step": 230 }, { "epoch": 0.4419889502762431, "grad_norm": 0.8513089510736624, "learning_rate": 9.933833401948514e-06, "loss": 0.2728, "step": 240 }, { "epoch": 0.4604051565377532, "grad_norm": 0.775184166522136, "learning_rate": 9.915328131881745e-06, "loss": 0.2741, "step": 250 }, { "epoch": 0.47882136279926335, "grad_norm": 0.7978646282069273, "learning_rate": 9.894565673985986e-06, "loss": 0.2725, "step": 260 }, { "epoch": 0.4972375690607735, "grad_norm": 0.6613832505637378, "learning_rate": 9.871555562673996e-06, "loss": 0.2675, "step": 270 }, { "epoch": 0.5156537753222836, "grad_norm": 0.8757001574454977, "learning_rate": 9.846308364512607e-06, "loss": 0.2661, "step": 280 }, { "epoch": 0.5340699815837937, "grad_norm": 0.7547104807299371, "learning_rate": 9.8188356733704e-06, "loss": 0.2742, "step": 290 }, { "epoch": 0.5524861878453039, "grad_norm": 0.7251338284529852, "learning_rate": 9.789150105093647e-06, "loss": 0.2683, "step": 300 }, { "epoch": 0.570902394106814, "grad_norm": 0.7543457855276412, "learning_rate": 9.75726529171293e-06, "loss": 0.2756, "step": 310 }, { "epoch": 0.5893186003683242, "grad_norm": 0.7747270052423587, "learning_rate": 9.72319587518312e-06, "loss": 0.2687, "step": 320 }, { "epoch": 0.6077348066298343, "grad_norm": 0.8708864438161477, "learning_rate": 9.68695750065959e-06, "loss": 0.2687, "step": 330 }, { "epoch": 0.6261510128913443, "grad_norm": 0.7181496119078857, "learning_rate": 9.648566809313738e-06, "loss": 0.2754, "step": 340 }, { "epoch": 0.6445672191528545, "grad_norm": 0.7212363207801823, "learning_rate": 9.608041430691126e-06, "loss": 0.2699, "step": 350 }, { "epoch": 0.6629834254143646, "grad_norm": 0.707979023449836, "learning_rate": 9.565399974615744e-06, "loss": 0.2688, "step": 360 }, { "epoch": 0.6813996316758748, "grad_norm": 0.8334790893435787, "learning_rate": 9.52066202264412e-06, "loss": 0.2776, "step": 370 }, { "epoch": 0.6998158379373849, "grad_norm": 0.7596815946675103, "learning_rate": 9.473848119073188e-06, "loss": 0.2786, "step": 380 }, { "epoch": 0.7182320441988951, "grad_norm": 0.7331861291943189, "learning_rate": 9.42497976150607e-06, "loss": 0.2622, "step": 390 }, { "epoch": 0.7366482504604052, "grad_norm": 0.6706665335188781, "learning_rate": 9.374079390980058e-06, "loss": 0.2695, "step": 400 }, { "epoch": 0.7550644567219152, "grad_norm": 0.7515563648690087, "learning_rate": 9.321170381661383e-06, "loss": 0.2671, "step": 410 }, { "epoch": 0.7734806629834254, "grad_norm": 0.6978960493224456, "learning_rate": 9.266277030111474e-06, "loss": 0.2721, "step": 420 }, { "epoch": 0.7918968692449355, "grad_norm": 0.6317044519904149, "learning_rate": 9.209424544129621e-06, "loss": 0.2611, "step": 430 }, { "epoch": 0.8103130755064457, "grad_norm": 0.6933884011578646, "learning_rate": 9.150639031177211e-06, "loss": 0.2679, "step": 440 }, { "epoch": 0.8287292817679558, "grad_norm": 0.6662516396211534, "learning_rate": 9.08994748638881e-06, "loss": 0.2718, "step": 450 }, { "epoch": 0.8471454880294659, "grad_norm": 0.7537377588910371, "learning_rate": 9.02737778017562e-06, "loss": 0.2625, "step": 460 }, { "epoch": 0.8655616942909761, "grad_norm": 0.6490665535396116, "learning_rate": 8.962958645426989e-06, "loss": 0.2601, "step": 470 }, { "epoch": 0.8839779005524862, "grad_norm": 0.6773727676397064, "learning_rate": 8.896719664315866e-06, "loss": 0.2651, "step": 480 }, { "epoch": 0.9023941068139963, "grad_norm": 0.6798276722239568, "learning_rate": 8.828691254714259e-06, "loss": 0.2625, "step": 490 }, { "epoch": 0.9208103130755064, "grad_norm": 0.6979929906196801, "learning_rate": 8.758904656224904e-06, "loss": 0.2749, "step": 500 }, { "epoch": 0.9392265193370166, "grad_norm": 0.7308459913669413, "learning_rate": 8.687391915835617e-06, "loss": 0.2666, "step": 510 }, { "epoch": 0.9576427255985267, "grad_norm": 0.8055261702848038, "learning_rate": 8.614185873202852e-06, "loss": 0.2705, "step": 520 }, { "epoch": 0.9760589318600368, "grad_norm": 0.7313961658432152, "learning_rate": 8.539320145571277e-06, "loss": 0.2619, "step": 530 }, { "epoch": 0.994475138121547, "grad_norm": 0.6331517103040231, "learning_rate": 8.462829112336266e-06, "loss": 0.2671, "step": 540 }, { "epoch": 1.0128913443830572, "grad_norm": 0.7085468234966358, "learning_rate": 8.384747899256386e-06, "loss": 0.2388, "step": 550 }, { "epoch": 1.0313075506445673, "grad_norm": 0.6748955679522807, "learning_rate": 8.30511236232316e-06, "loss": 0.2226, "step": 560 }, { "epoch": 1.0497237569060773, "grad_norm": 0.7100102091480333, "learning_rate": 8.223959071295492e-06, "loss": 0.2297, "step": 570 }, { "epoch": 1.0681399631675874, "grad_norm": 0.6668769883519982, "learning_rate": 8.141325292906325e-06, "loss": 0.2263, "step": 580 }, { "epoch": 1.0865561694290977, "grad_norm": 0.6720758025127164, "learning_rate": 8.057248973749216e-06, "loss": 0.2288, "step": 590 }, { "epoch": 1.1049723756906078, "grad_norm": 0.6411173465966967, "learning_rate": 7.971768722852741e-06, "loss": 0.2253, "step": 600 }, { "epoch": 1.1233885819521179, "grad_norm": 0.6724293241376955, "learning_rate": 7.884923793950684e-06, "loss": 0.2326, "step": 610 }, { "epoch": 1.141804788213628, "grad_norm": 0.6371467332633137, "learning_rate": 7.796754067456168e-06, "loss": 0.2235, "step": 620 }, { "epoch": 1.160220994475138, "grad_norm": 0.6085432732125634, "learning_rate": 7.707300032148004e-06, "loss": 0.2289, "step": 630 }, { "epoch": 1.1786372007366483, "grad_norm": 0.7730530541898509, "learning_rate": 7.616602766577683e-06, "loss": 0.2263, "step": 640 }, { "epoch": 1.1970534069981584, "grad_norm": 0.7420431722106332, "learning_rate": 7.524703920205521e-06, "loss": 0.2331, "step": 650 }, { "epoch": 1.2154696132596685, "grad_norm": 0.6604167477110746, "learning_rate": 7.43164569427464e-06, "loss": 0.2278, "step": 660 }, { "epoch": 1.2338858195211786, "grad_norm": 0.7778434196490822, "learning_rate": 7.3374708224315725e-06, "loss": 0.2242, "step": 670 }, { "epoch": 1.2523020257826887, "grad_norm": 0.7114711250180328, "learning_rate": 7.2422225511023555e-06, "loss": 0.2261, "step": 680 }, { "epoch": 1.270718232044199, "grad_norm": 0.7059769736094257, "learning_rate": 7.145944619633176e-06, "loss": 0.2218, "step": 690 }, { "epoch": 1.289134438305709, "grad_norm": 0.6846868707521909, "learning_rate": 7.048681240204641e-06, "loss": 0.2305, "step": 700 }, { "epoch": 1.3075506445672191, "grad_norm": 0.636969343049607, "learning_rate": 6.950477077528927e-06, "loss": 0.2242, "step": 710 }, { "epoch": 1.3259668508287292, "grad_norm": 0.633812292072214, "learning_rate": 6.851377228339106e-06, "loss": 0.2302, "step": 720 }, { "epoch": 1.3443830570902393, "grad_norm": 0.6630293802277623, "learning_rate": 6.751427200680109e-06, "loss": 0.2295, "step": 730 }, { "epoch": 1.3627992633517496, "grad_norm": 0.6315090136897826, "learning_rate": 6.650672893010769e-06, "loss": 0.2221, "step": 740 }, { "epoch": 1.3812154696132597, "grad_norm": 0.7126623318254198, "learning_rate": 6.549160573126623e-06, "loss": 0.2293, "step": 750 }, { "epoch": 1.3996316758747698, "grad_norm": 0.6653513173488009, "learning_rate": 6.4469368569130786e-06, "loss": 0.2282, "step": 760 }, { "epoch": 1.4180478821362799, "grad_norm": 0.6343921369098496, "learning_rate": 6.344048686938745e-06, "loss": 0.2273, "step": 770 }, { "epoch": 1.43646408839779, "grad_norm": 0.6210167648750854, "learning_rate": 6.2405433108987456e-06, "loss": 0.2338, "step": 780 }, { "epoch": 1.4548802946593002, "grad_norm": 0.6784962640818855, "learning_rate": 6.136468259917917e-06, "loss": 0.2289, "step": 790 }, { "epoch": 1.4732965009208103, "grad_norm": 0.578659502472278, "learning_rate": 6.031871326723837e-06, "loss": 0.2363, "step": 800 }, { "epoch": 1.4917127071823204, "grad_norm": 0.7040713369273666, "learning_rate": 5.92680054369974e-06, "loss": 0.2233, "step": 810 }, { "epoch": 1.5101289134438307, "grad_norm": 0.7100429365248154, "learning_rate": 5.821304160827371e-06, "loss": 0.2275, "step": 820 }, { "epoch": 1.5285451197053406, "grad_norm": 0.5875774880678092, "learning_rate": 5.71543062352991e-06, "loss": 0.225, "step": 830 }, { "epoch": 1.5469613259668509, "grad_norm": 0.5895563949605374, "learning_rate": 5.609228550425154e-06, "loss": 0.2347, "step": 840 }, { "epoch": 1.565377532228361, "grad_norm": 0.6923142403050372, "learning_rate": 5.50274671099917e-06, "loss": 0.2212, "step": 850 }, { "epoch": 1.583793738489871, "grad_norm": 0.6402043458583734, "learning_rate": 5.3960340032106515e-06, "loss": 0.221, "step": 860 }, { "epoch": 1.6022099447513813, "grad_norm": 0.6583270982368973, "learning_rate": 5.28913943103629e-06, "loss": 0.2263, "step": 870 }, { "epoch": 1.6206261510128912, "grad_norm": 0.6545061377359965, "learning_rate": 5.182112081967467e-06, "loss": 0.2221, "step": 880 }, { "epoch": 1.6390423572744015, "grad_norm": 0.7057814110328757, "learning_rate": 5.075001104468576e-06, "loss": 0.2303, "step": 890 }, { "epoch": 1.6574585635359116, "grad_norm": 0.6738126191275001, "learning_rate": 4.967855685407368e-06, "loss": 0.2235, "step": 900 }, { "epoch": 1.6758747697974217, "grad_norm": 0.6229975201509437, "learning_rate": 4.860725027467641e-06, "loss": 0.2276, "step": 910 }, { "epoch": 1.694290976058932, "grad_norm": 0.6485619857748801, "learning_rate": 4.7536583265546775e-06, "loss": 0.2294, "step": 920 }, { "epoch": 1.7127071823204418, "grad_norm": 0.6952097611858238, "learning_rate": 4.646704749203794e-06, "loss": 0.22, "step": 930 }, { "epoch": 1.7311233885819521, "grad_norm": 0.6993773129475868, "learning_rate": 4.539913410002378e-06, "loss": 0.2253, "step": 940 }, { "epoch": 1.7495395948434622, "grad_norm": 0.7179073268017699, "learning_rate": 4.433333349035773e-06, "loss": 0.2262, "step": 950 }, { "epoch": 1.7679558011049723, "grad_norm": 0.6202631830964883, "learning_rate": 4.327013509367386e-06, "loss": 0.2212, "step": 960 }, { "epoch": 1.7863720073664826, "grad_norm": 0.6608879257543027, "learning_rate": 4.221002714563347e-06, "loss": 0.2205, "step": 970 }, { "epoch": 1.8047882136279927, "grad_norm": 0.59224317673099, "learning_rate": 4.115349646272029e-06, "loss": 0.2205, "step": 980 }, { "epoch": 1.8232044198895028, "grad_norm": 0.5878582444161091, "learning_rate": 4.010102821868762e-06, "loss": 0.2229, "step": 990 }, { "epoch": 1.8416206261510129, "grad_norm": 0.6461746282587334, "learning_rate": 3.90531057217597e-06, "loss": 0.2232, "step": 1000 }, { "epoch": 1.860036832412523, "grad_norm": 0.6205082651968326, "learning_rate": 3.8010210192689688e-06, "loss": 0.2277, "step": 1010 }, { "epoch": 1.8784530386740332, "grad_norm": 0.7181661264692984, "learning_rate": 3.6972820543776404e-06, "loss": 0.2261, "step": 1020 }, { "epoch": 1.8968692449355433, "grad_norm": 0.6649391919825204, "learning_rate": 3.5941413158941086e-06, "loss": 0.2259, "step": 1030 }, { "epoch": 1.9152854511970534, "grad_norm": 0.6044692798894635, "learning_rate": 3.4916461674965074e-06, "loss": 0.222, "step": 1040 }, { "epoch": 1.9337016574585635, "grad_norm": 0.615450434298784, "learning_rate": 3.389843676398925e-06, "loss": 0.2254, "step": 1050 }, { "epoch": 1.9521178637200736, "grad_norm": 0.6691163050124782, "learning_rate": 3.2887805917374736e-06, "loss": 0.2189, "step": 1060 }, { "epoch": 1.9705340699815839, "grad_norm": 0.6022414370557024, "learning_rate": 3.1885033231024253e-06, "loss": 0.2227, "step": 1070 }, { "epoch": 1.988950276243094, "grad_norm": 0.6942520950174368, "learning_rate": 3.089057919226277e-06, "loss": 0.2254, "step": 1080 }, { "epoch": 2.007366482504604, "grad_norm": 0.5986817795885931, "learning_rate": 2.9904900468375298e-06, "loss": 0.2091, "step": 1090 }, { "epoch": 2.0257826887661143, "grad_norm": 0.5655950429797426, "learning_rate": 2.892844969689876e-06, "loss": 0.184, "step": 1100 }, { "epoch": 2.044198895027624, "grad_norm": 0.6492187365821179, "learning_rate": 2.7961675277764498e-06, "loss": 0.1815, "step": 1110 }, { "epoch": 2.0626151012891345, "grad_norm": 0.5494541259912535, "learning_rate": 2.7005021167386804e-06, "loss": 0.1782, "step": 1120 }, { "epoch": 2.0810313075506444, "grad_norm": 0.6018040555497874, "learning_rate": 2.605892667479173e-06, "loss": 0.1806, "step": 1130 }, { "epoch": 2.0994475138121547, "grad_norm": 0.609141353634607, "learning_rate": 2.5123826259880324e-06, "loss": 0.1827, "step": 1140 }, { "epoch": 2.117863720073665, "grad_norm": 0.6522701323065019, "learning_rate": 2.420014933391849e-06, "loss": 0.1834, "step": 1150 }, { "epoch": 2.136279926335175, "grad_norm": 0.6209038557527796, "learning_rate": 2.3288320062345276e-06, "loss": 0.1832, "step": 1160 }, { "epoch": 2.154696132596685, "grad_norm": 0.671734526610132, "learning_rate": 2.238875716999019e-06, "loss": 0.181, "step": 1170 }, { "epoch": 2.1731123388581954, "grad_norm": 0.6211775861295186, "learning_rate": 2.1501873748788804e-06, "loss": 0.1859, "step": 1180 }, { "epoch": 2.1915285451197053, "grad_norm": 0.6161892750894669, "learning_rate": 2.0628077068085173e-06, "loss": 0.1833, "step": 1190 }, { "epoch": 2.2099447513812156, "grad_norm": 0.6171060565609299, "learning_rate": 1.976776838760801e-06, "loss": 0.1823, "step": 1200 }, { "epoch": 2.2283609576427255, "grad_norm": 0.6627457414621876, "learning_rate": 1.8921342773206553e-06, "loss": 0.1802, "step": 1210 }, { "epoch": 2.2467771639042358, "grad_norm": 0.5766657805895978, "learning_rate": 1.8089188915430794e-06, "loss": 0.1801, "step": 1220 }, { "epoch": 2.265193370165746, "grad_norm": 0.6528068390986206, "learning_rate": 1.7271688951039312e-06, "loss": 0.1817, "step": 1230 }, { "epoch": 2.283609576427256, "grad_norm": 0.6108145113882844, "learning_rate": 1.6469218287516664e-06, "loss": 0.1824, "step": 1240 }, { "epoch": 2.3020257826887662, "grad_norm": 0.6223360388944672, "learning_rate": 1.568214543068103e-06, "loss": 0.1817, "step": 1250 }, { "epoch": 2.320441988950276, "grad_norm": 0.5730293037393086, "learning_rate": 1.4910831815461125e-06, "loss": 0.1783, "step": 1260 }, { "epoch": 2.3388581952117864, "grad_norm": 0.6804328273268474, "learning_rate": 1.4155631639920208e-06, "loss": 0.1793, "step": 1270 }, { "epoch": 2.3572744014732967, "grad_norm": 0.5936583186703833, "learning_rate": 1.3416891702603357e-06, "loss": 0.1772, "step": 1280 }, { "epoch": 2.3756906077348066, "grad_norm": 0.6646979614560136, "learning_rate": 1.2694951243282682e-06, "loss": 0.1805, "step": 1290 }, { "epoch": 2.394106813996317, "grad_norm": 0.5867258571902689, "learning_rate": 1.1990141787173648e-06, "loss": 0.1793, "step": 1300 }, { "epoch": 2.4125230202578267, "grad_norm": 0.6356400483556068, "learning_rate": 1.1302786992694049e-06, "loss": 0.1793, "step": 1310 }, { "epoch": 2.430939226519337, "grad_norm": 0.6071714234787918, "learning_rate": 1.0633202502835494e-06, "loss": 0.1831, "step": 1320 }, { "epoch": 2.4493554327808473, "grad_norm": 0.6031318900784148, "learning_rate": 9.981695800215701e-07, "loss": 0.1795, "step": 1330 }, { "epoch": 2.467771639042357, "grad_norm": 0.6195067631818659, "learning_rate": 9.348566065878218e-07, "loss": 0.1807, "step": 1340 }, { "epoch": 2.4861878453038675, "grad_norm": 0.640824325612606, "learning_rate": 8.734104041904129e-07, "loss": 0.1835, "step": 1350 }, { "epoch": 2.5046040515653774, "grad_norm": 0.5801881807297086, "learning_rate": 8.138591897899345e-07, "loss": 0.1792, "step": 1360 }, { "epoch": 2.5230202578268877, "grad_norm": 0.6192805732625175, "learning_rate": 7.56230310141835e-07, "loss": 0.179, "step": 1370 }, { "epoch": 2.541436464088398, "grad_norm": 0.6016304803765787, "learning_rate": 7.005502292383898e-07, "loss": 0.1799, "step": 1380 }, { "epoch": 2.559852670349908, "grad_norm": 0.6115554388128484, "learning_rate": 6.46844516156081e-07, "loss": 0.1758, "step": 1390 }, { "epoch": 2.578268876611418, "grad_norm": 0.5927866192371272, "learning_rate": 5.951378333139118e-07, "loss": 0.1796, "step": 1400 }, { "epoch": 2.596685082872928, "grad_norm": 0.6811377078596266, "learning_rate": 5.454539251480739e-07, "loss": 0.1808, "step": 1410 }, { "epoch": 2.6151012891344383, "grad_norm": 0.6077558733405245, "learning_rate": 4.978156072081669e-07, "loss": 0.1784, "step": 1420 }, { "epoch": 2.6335174953959486, "grad_norm": 0.5643531059218968, "learning_rate": 4.522447556799875e-07, "loss": 0.1795, "step": 1430 }, { "epoch": 2.6519337016574585, "grad_norm": 0.5619239404266387, "learning_rate": 4.0876229733966655e-07, "loss": 0.1759, "step": 1440 }, { "epoch": 2.6703499079189688, "grad_norm": 0.6328209657239623, "learning_rate": 3.6738819994379946e-07, "loss": 0.1808, "step": 1450 }, { "epoch": 2.6887661141804786, "grad_norm": 0.6086714396313364, "learning_rate": 3.281414630599811e-07, "loss": 0.1772, "step": 1460 }, { "epoch": 2.707182320441989, "grad_norm": 0.6266202928122402, "learning_rate": 2.9104010934192795e-07, "loss": 0.1798, "step": 1470 }, { "epoch": 2.7255985267034992, "grad_norm": 0.61076576372029, "learning_rate": 2.561011762532212e-07, "loss": 0.1749, "step": 1480 }, { "epoch": 2.744014732965009, "grad_norm": 0.6615569148493705, "learning_rate": 2.2334070824347243e-07, "loss": 0.1805, "step": 1490 }, { "epoch": 2.7624309392265194, "grad_norm": 0.6278844879654595, "learning_rate": 1.9277374938047989e-07, "loss": 0.1808, "step": 1500 }, { "epoch": 2.7808471454880292, "grad_norm": 0.57717673931454, "learning_rate": 1.644143364417794e-07, "loss": 0.1775, "step": 1510 }, { "epoch": 2.7992633517495396, "grad_norm": 0.5988782652066212, "learning_rate": 1.3827549246876627e-07, "loss": 0.1841, "step": 1520 }, { "epoch": 2.81767955801105, "grad_norm": 0.6220611373904551, "learning_rate": 1.1436922078632395e-07, "loss": 0.1773, "step": 1530 }, { "epoch": 2.8360957642725597, "grad_norm": 0.6549658009335396, "learning_rate": 9.270649949073229e-08, "loss": 0.1796, "step": 1540 }, { "epoch": 2.85451197053407, "grad_norm": 0.5897064017884592, "learning_rate": 7.329727640837059e-08, "loss": 0.1781, "step": 1550 }, { "epoch": 2.87292817679558, "grad_norm": 0.6006548076063878, "learning_rate": 5.615046452753403e-08, "loss": 0.1789, "step": 1560 }, { "epoch": 2.89134438305709, "grad_norm": 0.5940577802255843, "learning_rate": 4.127393790546719e-08, "loss": 0.1801, "step": 1570 }, { "epoch": 2.9097605893186005, "grad_norm": 0.6226132514184528, "learning_rate": 2.8674528052484162e-08, "loss": 0.1774, "step": 1580 }, { "epoch": 2.9281767955801103, "grad_norm": 0.6274949409362496, "learning_rate": 1.8358020794843056e-08, "loss": 0.1827, "step": 1590 }, { "epoch": 2.9465930018416207, "grad_norm": 0.6038116585141188, "learning_rate": 1.0329153617812948e-08, "loss": 0.1781, "step": 1600 }, { "epoch": 2.9650092081031305, "grad_norm": 0.5802100431657281, "learning_rate": 4.5916134901552445e-09, "loss": 0.1825, "step": 1610 }, { "epoch": 2.983425414364641, "grad_norm": 0.6005447860892711, "learning_rate": 1.148035171014139e-09, "loss": 0.1839, "step": 1620 }, { "epoch": 3.0, "step": 1629, "total_flos": 2800162131935232.0, "train_loss": 0.22954999712656582, "train_runtime": 49788.0682, "train_samples_per_second": 2.094, "train_steps_per_second": 0.033 } ], "logging_steps": 10, "max_steps": 1629, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2800162131935232.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }