{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.4, "eval_steps": 500, "global_step": 2000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "grad_norm": 0.5184409618377686, "learning_rate": 3.6e-06, "loss": 1.0866, "step": 10 }, { "grad_norm": 0.13902701437473297, "learning_rate": 7.6e-06, "loss": 1.0761, "step": 20 }, { "grad_norm": 0.0735388919711113, "learning_rate": 1.16e-05, "loss": 1.0654, "step": 30 }, { "grad_norm": 0.0823279619216919, "learning_rate": 1.56e-05, "loss": 1.0577, "step": 40 }, { "grad_norm": 0.0818614587187767, "learning_rate": 1.9600000000000002e-05, "loss": 1.0471, "step": 50 }, { "grad_norm": 0.07863084971904755, "learning_rate": 2.36e-05, "loss": 1.0344, "step": 60 }, { "grad_norm": 0.08459152281284332, "learning_rate": 2.7600000000000003e-05, "loss": 1.0274, "step": 70 }, { "grad_norm": 0.09263663738965988, "learning_rate": 3.16e-05, "loss": 1.0254, "step": 80 }, { "grad_norm": 0.11630931496620178, "learning_rate": 3.56e-05, "loss": 1.0188, "step": 90 }, { "grad_norm": 0.2673693001270294, "learning_rate": 3.960000000000001e-05, "loss": 0.9947, "step": 100 }, { "grad_norm": 0.8061618208885193, "learning_rate": 4.36e-05, "loss": 0.933, "step": 110 }, { "grad_norm": 0.6079975962638855, "learning_rate": 4.76e-05, "loss": 0.843, "step": 120 }, { "grad_norm": 0.5954048037528992, "learning_rate": 5.16e-05, "loss": 0.7637, "step": 130 }, { "grad_norm": 0.856939971446991, "learning_rate": 5.560000000000001e-05, "loss": 0.6892, "step": 140 }, { "grad_norm": 0.987891435623169, "learning_rate": 5.96e-05, "loss": 0.624, "step": 150 }, { "grad_norm": 0.7554893493652344, "learning_rate": 6.36e-05, "loss": 0.5657, "step": 160 }, { "grad_norm": 0.9280672669410706, "learning_rate": 6.76e-05, "loss": 0.5029, "step": 170 }, { "grad_norm": 1.0552935600280762, "learning_rate": 7.16e-05, "loss": 0.4461, "step": 180 }, { "grad_norm": 0.9269078373908997, "learning_rate": 7.560000000000001e-05, "loss": 0.3999, "step": 190 }, { "grad_norm": 0.7659823894500732, "learning_rate": 7.960000000000001e-05, "loss": 0.3686, "step": 200 }, { "grad_norm": 0.8024225831031799, "learning_rate": 8.36e-05, "loss": 0.3347, "step": 210 }, { "grad_norm": 0.9907840490341187, "learning_rate": 8.76e-05, "loss": 0.3038, "step": 220 }, { "grad_norm": 1.0437959432601929, "learning_rate": 9.16e-05, "loss": 0.2782, "step": 230 }, { "grad_norm": 0.7399183511734009, "learning_rate": 9.56e-05, "loss": 0.2469, "step": 240 }, { "grad_norm": 1.2706823348999023, "learning_rate": 9.960000000000001e-05, "loss": 0.2156, "step": 250 }, { "grad_norm": 1.0207575559616089, "learning_rate": 9.999911419878559e-05, "loss": 0.2059, "step": 260 }, { "grad_norm": 0.8645009398460388, "learning_rate": 9.999605221019081e-05, "loss": 0.1927, "step": 270 }, { "grad_norm": 1.3531432151794434, "learning_rate": 9.999080323230761e-05, "loss": 0.1758, "step": 280 }, { "grad_norm": 0.8455585837364197, "learning_rate": 9.998336749474329e-05, "loss": 0.1719, "step": 290 }, { "grad_norm": 1.9373207092285156, "learning_rate": 9.997374532276107e-05, "loss": 0.1467, "step": 300 }, { "grad_norm": 0.7995489239692688, "learning_rate": 9.996193713726596e-05, "loss": 0.1327, "step": 310 }, { "grad_norm": 0.8662717938423157, "learning_rate": 9.994794345478624e-05, "loss": 0.1254, "step": 320 }, { "grad_norm": 0.9200862646102905, "learning_rate": 9.99317648874509e-05, "loss": 0.1229, "step": 330 }, { "grad_norm": 0.8963688611984253, "learning_rate": 9.991340214296292e-05, "loss": 0.1151, "step": 340 }, { "grad_norm": 1.0278245210647583, "learning_rate": 9.989285602456819e-05, "loss": 0.1043, "step": 350 }, { "grad_norm": 0.8198051452636719, "learning_rate": 9.98701274310205e-05, "loss": 0.092, "step": 360 }, { "grad_norm": 0.9469745755195618, "learning_rate": 9.984521735654218e-05, "loss": 0.0756, "step": 370 }, { "grad_norm": 0.8919957280158997, "learning_rate": 9.981812689078057e-05, "loss": 0.0762, "step": 380 }, { "grad_norm": 0.8211736083030701, "learning_rate": 9.978885721876041e-05, "loss": 0.0691, "step": 390 }, { "grad_norm": 0.8344348073005676, "learning_rate": 9.975740962083198e-05, "loss": 0.0675, "step": 400 }, { "grad_norm": 0.8462538719177246, "learning_rate": 9.972378547261504e-05, "loss": 0.0643, "step": 410 }, { "grad_norm": 0.7656953930854797, "learning_rate": 9.968798624493885e-05, "loss": 0.0612, "step": 420 }, { "grad_norm": 0.7218711972236633, "learning_rate": 9.965001350377753e-05, "loss": 0.0644, "step": 430 }, { "grad_norm": 1.0755572319030762, "learning_rate": 9.960986891018183e-05, "loss": 0.0616, "step": 440 }, { "grad_norm": 0.7578702569007874, "learning_rate": 9.95675542202063e-05, "loss": 0.0564, "step": 450 }, { "grad_norm": 0.7378067374229431, "learning_rate": 9.952307128483256e-05, "loss": 0.0524, "step": 460 }, { "grad_norm": 0.7673323750495911, "learning_rate": 9.947642204988835e-05, "loss": 0.0534, "step": 470 }, { "grad_norm": 0.8811172842979431, "learning_rate": 9.942760855596226e-05, "loss": 0.056, "step": 480 }, { "grad_norm": 0.717313826084137, "learning_rate": 9.937663293831471e-05, "loss": 0.0563, "step": 490 }, { "grad_norm": 0.8490850925445557, "learning_rate": 9.932349742678433e-05, "loss": 0.0586, "step": 500 }, { "grad_norm": 0.7520468235015869, "learning_rate": 9.926820434569051e-05, "loss": 0.0532, "step": 510 }, { "grad_norm": 0.7413885593414307, "learning_rate": 9.921075611373179e-05, "loss": 0.056, "step": 520 }, { "grad_norm": 0.7074100375175476, "learning_rate": 9.915115524387988e-05, "loss": 0.0517, "step": 530 }, { "grad_norm": 0.7114731073379517, "learning_rate": 9.908940434326997e-05, "loss": 0.048, "step": 540 }, { "grad_norm": 0.8586488962173462, "learning_rate": 9.902550611308645e-05, "loss": 0.0509, "step": 550 }, { "grad_norm": 0.7280430197715759, "learning_rate": 9.895946334844494e-05, "loss": 0.0485, "step": 560 }, { "grad_norm": 0.7949817180633545, "learning_rate": 9.889127893826989e-05, "loss": 0.05, "step": 570 }, { "grad_norm": 0.6108531355857849, "learning_rate": 9.882095586516831e-05, "loss": 0.0499, "step": 580 }, { "grad_norm": 0.6878312826156616, "learning_rate": 9.874849720529921e-05, "loss": 0.0501, "step": 590 }, { "grad_norm": 0.7149339914321899, "learning_rate": 9.867390612823914e-05, "loss": 0.0505, "step": 600 }, { "grad_norm": 0.8408756256103516, "learning_rate": 9.859718589684344e-05, "loss": 0.0448, "step": 610 }, { "grad_norm": 0.8319779634475708, "learning_rate": 9.851833986710353e-05, "loss": 0.0477, "step": 620 }, { "grad_norm": 0.7106873393058777, "learning_rate": 9.843737148800023e-05, "loss": 0.0458, "step": 630 }, { "grad_norm": 0.7426978349685669, "learning_rate": 9.835428430135271e-05, "loss": 0.0494, "step": 640 }, { "grad_norm": 0.8612025380134583, "learning_rate": 9.82690819416637e-05, "loss": 0.0469, "step": 650 }, { "grad_norm": 0.7378700971603394, "learning_rate": 9.818176813596041e-05, "loss": 0.0483, "step": 660 }, { "grad_norm": 0.6754602789878845, "learning_rate": 9.809234670363159e-05, "loss": 0.0417, "step": 670 }, { "grad_norm": 0.6960591673851013, "learning_rate": 9.800082155626034e-05, "loss": 0.0483, "step": 680 }, { "grad_norm": 0.7274064421653748, "learning_rate": 9.790719669745312e-05, "loss": 0.042, "step": 690 }, { "grad_norm": 0.6894716620445251, "learning_rate": 9.781147622266455e-05, "loss": 0.0389, "step": 700 }, { "grad_norm": 0.776299238204956, "learning_rate": 9.771366431901831e-05, "loss": 0.0415, "step": 710 }, { "grad_norm": 0.6639792919158936, "learning_rate": 9.761376526512394e-05, "loss": 0.0463, "step": 720 }, { "grad_norm": 0.6171531677246094, "learning_rate": 9.751178343088963e-05, "loss": 0.0413, "step": 730 }, { "grad_norm": 0.715189516544342, "learning_rate": 9.740772327733123e-05, "loss": 0.0393, "step": 740 }, { "grad_norm": 0.6551257371902466, "learning_rate": 9.730158935637697e-05, "loss": 0.0417, "step": 750 }, { "grad_norm": 0.6576980948448181, "learning_rate": 9.719338631066834e-05, "loss": 0.0423, "step": 760 }, { "grad_norm": 0.6665154099464417, "learning_rate": 9.708311887335713e-05, "loss": 0.0414, "step": 770 }, { "grad_norm": 0.753499448299408, "learning_rate": 9.697079186789823e-05, "loss": 0.0412, "step": 780 }, { "grad_norm": 0.6803208589553833, "learning_rate": 9.685641020783876e-05, "loss": 0.042, "step": 790 }, { "grad_norm": 0.6924328804016113, "learning_rate": 9.67399788966031e-05, "loss": 0.0415, "step": 800 }, { "grad_norm": 0.7184621691703796, "learning_rate": 9.662150302727395e-05, "loss": 0.0377, "step": 810 }, { "grad_norm": 0.6149929761886597, "learning_rate": 9.650098778236968e-05, "loss": 0.0405, "step": 820 }, { "grad_norm": 0.6302839517593384, "learning_rate": 9.637843843361749e-05, "loss": 0.0371, "step": 830 }, { "grad_norm": 0.5969130396842957, "learning_rate": 9.62538603417229e-05, "loss": 0.0404, "step": 840 }, { "grad_norm": 0.56196528673172, "learning_rate": 9.612725895613526e-05, "loss": 0.04, "step": 850 }, { "grad_norm": 0.7191771864891052, "learning_rate": 9.599863981480926e-05, "loss": 0.039, "step": 860 }, { "grad_norm": 0.5886449813842773, "learning_rate": 9.586800854396283e-05, "loss": 0.0369, "step": 870 }, { "grad_norm": 0.5173853039741516, "learning_rate": 9.573537085783095e-05, "loss": 0.0358, "step": 880 }, { "grad_norm": 0.5850680470466614, "learning_rate": 9.560073255841571e-05, "loss": 0.0427, "step": 890 }, { "grad_norm": 0.758872926235199, "learning_rate": 9.546409953523247e-05, "loss": 0.04, "step": 900 }, { "grad_norm": 0.631585955619812, "learning_rate": 9.532547776505229e-05, "loss": 0.0385, "step": 910 }, { "grad_norm": 0.42788979411125183, "learning_rate": 9.518487331164048e-05, "loss": 0.0359, "step": 920 }, { "grad_norm": 0.8113523125648499, "learning_rate": 9.504229232549134e-05, "loss": 0.036, "step": 930 }, { "grad_norm": 0.5256606340408325, "learning_rate": 9.489774104355909e-05, "loss": 0.0369, "step": 940 }, { "grad_norm": 0.680260956287384, "learning_rate": 9.475122578898507e-05, "loss": 0.0356, "step": 950 }, { "grad_norm": 0.6082437038421631, "learning_rate": 9.460275297082119e-05, "loss": 0.035, "step": 960 }, { "grad_norm": 0.6395261883735657, "learning_rate": 9.445232908374948e-05, "loss": 0.0398, "step": 970 }, { "grad_norm": 0.5504397749900818, "learning_rate": 9.429996070779808e-05, "loss": 0.0347, "step": 980 }, { "grad_norm": 0.6658892035484314, "learning_rate": 9.414565450805333e-05, "loss": 0.038, "step": 990 }, { "grad_norm": 0.6586282253265381, "learning_rate": 9.398941723436831e-05, "loss": 0.0348, "step": 1000 }, { "grad_norm": 0.563089907169342, "learning_rate": 9.383125572106752e-05, "loss": 0.0365, "step": 1010 }, { "grad_norm": 0.6453968286514282, "learning_rate": 9.367117688664791e-05, "loss": 0.0352, "step": 1020 }, { "grad_norm": 0.7110365629196167, "learning_rate": 9.35091877334763e-05, "loss": 0.0373, "step": 1030 }, { "grad_norm": 0.8032221794128418, "learning_rate": 9.334529534748297e-05, "loss": 0.0373, "step": 1040 }, { "grad_norm": 0.5992754697799683, "learning_rate": 9.317950689785188e-05, "loss": 0.0372, "step": 1050 }, { "grad_norm": 0.5257222652435303, "learning_rate": 9.301182963670688e-05, "loss": 0.0349, "step": 1060 }, { "grad_norm": 0.5869795083999634, "learning_rate": 9.284227089879456e-05, "loss": 0.0379, "step": 1070 }, { "grad_norm": 0.6808982491493225, "learning_rate": 9.26708381011634e-05, "loss": 0.0321, "step": 1080 }, { "grad_norm": 0.8073207139968872, "learning_rate": 9.249753874283937e-05, "loss": 0.0363, "step": 1090 }, { "grad_norm": 0.6191753149032593, "learning_rate": 9.232238040449779e-05, "loss": 0.0338, "step": 1100 }, { "grad_norm": 0.6077655553817749, "learning_rate": 9.214537074813181e-05, "loss": 0.034, "step": 1110 }, { "grad_norm": 0.5800149440765381, "learning_rate": 9.196651751671724e-05, "loss": 0.0342, "step": 1120 }, { "grad_norm": 0.5005490183830261, "learning_rate": 9.178582853387384e-05, "loss": 0.0396, "step": 1130 }, { "grad_norm": 0.5275071859359741, "learning_rate": 9.160331170352304e-05, "loss": 0.0364, "step": 1140 }, { "grad_norm": 0.5516709685325623, "learning_rate": 9.141897500954229e-05, "loss": 0.0326, "step": 1150 }, { "grad_norm": 0.6132859587669373, "learning_rate": 9.123282651541576e-05, "loss": 0.0345, "step": 1160 }, { "grad_norm": 0.6618711948394775, "learning_rate": 9.104487436388161e-05, "loss": 0.0364, "step": 1170 }, { "grad_norm": 0.48607364296913147, "learning_rate": 9.085512677657582e-05, "loss": 0.0347, "step": 1180 }, { "grad_norm": 0.6589488983154297, "learning_rate": 9.066359205367258e-05, "loss": 0.0343, "step": 1190 }, { "grad_norm": 0.5488317012786865, "learning_rate": 9.047027857352112e-05, "loss": 0.0344, "step": 1200 }, { "grad_norm": 0.4591907262802124, "learning_rate": 9.027519479227935e-05, "loss": 0.0314, "step": 1210 }, { "grad_norm": 0.6019390821456909, "learning_rate": 9.007834924354383e-05, "loss": 0.0338, "step": 1220 }, { "grad_norm": 0.5816202759742737, "learning_rate": 8.987975053797655e-05, "loss": 0.0351, "step": 1230 }, { "grad_norm": 0.5801373720169067, "learning_rate": 8.967940736292825e-05, "loss": 0.0345, "step": 1240 }, { "grad_norm": 0.6135644316673279, "learning_rate": 8.947732848205846e-05, "loss": 0.0341, "step": 1250 }, { "grad_norm": 0.5733956098556519, "learning_rate": 8.927352273495204e-05, "loss": 0.0347, "step": 1260 }, { "grad_norm": 0.5448894500732422, "learning_rate": 8.906799903673265e-05, "loss": 0.0314, "step": 1270 }, { "grad_norm": 0.474997878074646, "learning_rate": 8.88607663776726e-05, "loss": 0.0337, "step": 1280 }, { "grad_norm": 0.5049208998680115, "learning_rate": 8.865183382279978e-05, "loss": 0.0334, "step": 1290 }, { "grad_norm": 0.6067247986793518, "learning_rate": 8.844121051150096e-05, "loss": 0.0325, "step": 1300 }, { "grad_norm": 0.5232999920845032, "learning_rate": 8.822890565712211e-05, "loss": 0.0318, "step": 1310 }, { "grad_norm": 0.5644145607948303, "learning_rate": 8.801492854656536e-05, "loss": 0.0333, "step": 1320 }, { "grad_norm": 0.5336611866950989, "learning_rate": 8.779928853988268e-05, "loss": 0.0322, "step": 1330 }, { "grad_norm": 0.6173704862594604, "learning_rate": 8.758199506986655e-05, "loss": 0.0318, "step": 1340 }, { "grad_norm": 0.5638954639434814, "learning_rate": 8.73630576416373e-05, "loss": 0.0327, "step": 1350 }, { "grad_norm": 0.5919230580329895, "learning_rate": 8.714248583222726e-05, "loss": 0.0337, "step": 1360 }, { "grad_norm": 0.4524107575416565, "learning_rate": 8.692028929016196e-05, "loss": 0.0329, "step": 1370 }, { "grad_norm": 0.5694531798362732, "learning_rate": 8.669647773503797e-05, "loss": 0.0326, "step": 1380 }, { "grad_norm": 0.5108136534690857, "learning_rate": 8.647106095709773e-05, "loss": 0.031, "step": 1390 }, { "grad_norm": 0.6212553977966309, "learning_rate": 8.624404881680139e-05, "loss": 0.0324, "step": 1400 }, { "grad_norm": 0.4589623808860779, "learning_rate": 8.601545124439535e-05, "loss": 0.0281, "step": 1410 }, { "grad_norm": 0.4673466682434082, "learning_rate": 8.5785278239478e-05, "loss": 0.0328, "step": 1420 }, { "grad_norm": 0.6234567165374756, "learning_rate": 8.555353987056224e-05, "loss": 0.0301, "step": 1430 }, { "grad_norm": 0.7200560569763184, "learning_rate": 8.532024627463505e-05, "loss": 0.0302, "step": 1440 }, { "grad_norm": 0.6383592486381531, "learning_rate": 8.508540765671407e-05, "loss": 0.0291, "step": 1450 }, { "grad_norm": 0.43400031328201294, "learning_rate": 8.484903428940121e-05, "loss": 0.0294, "step": 1460 }, { "grad_norm": 0.4296083152294159, "learning_rate": 8.461113651243334e-05, "loss": 0.0329, "step": 1470 }, { "grad_norm": 0.5975046157836914, "learning_rate": 8.437172473222987e-05, "loss": 0.0344, "step": 1480 }, { "grad_norm": 0.5191563367843628, "learning_rate": 8.413080942143767e-05, "loss": 0.0294, "step": 1490 }, { "grad_norm": 0.4823097884654999, "learning_rate": 8.388840111847288e-05, "loss": 0.0304, "step": 1500 }, { "grad_norm": 0.6224477887153625, "learning_rate": 8.364451042705998e-05, "loss": 0.0324, "step": 1510 }, { "grad_norm": 0.520642876625061, "learning_rate": 8.33991480157679e-05, "loss": 0.0282, "step": 1520 }, { "grad_norm": 0.5692772269248962, "learning_rate": 8.315232461754338e-05, "loss": 0.0296, "step": 1530 }, { "grad_norm": 0.41501984000205994, "learning_rate": 8.290405102924144e-05, "loss": 0.031, "step": 1540 }, { "grad_norm": 0.5292897820472717, "learning_rate": 8.265433811115316e-05, "loss": 0.0312, "step": 1550 }, { "grad_norm": 0.4766462445259094, "learning_rate": 8.240319678653049e-05, "loss": 0.0314, "step": 1560 }, { "grad_norm": 0.625045895576477, "learning_rate": 8.215063804110857e-05, "loss": 0.0327, "step": 1570 }, { "grad_norm": 0.4520655572414398, "learning_rate": 8.189667292262512e-05, "loss": 0.0297, "step": 1580 }, { "grad_norm": 0.4334840178489685, "learning_rate": 8.164131254033716e-05, "loss": 0.0302, "step": 1590 }, { "grad_norm": 0.5579179525375366, "learning_rate": 8.138456806453503e-05, "loss": 0.0304, "step": 1600 }, { "grad_norm": 0.5523282289505005, "learning_rate": 8.112645072605386e-05, "loss": 0.0291, "step": 1610 }, { "grad_norm": 0.47829750180244446, "learning_rate": 8.086697181578222e-05, "loss": 0.0332, "step": 1620 }, { "grad_norm": 0.5201600790023804, "learning_rate": 8.060614268416823e-05, "loss": 0.0305, "step": 1630 }, { "grad_norm": 0.5666531920433044, "learning_rate": 8.034397474072309e-05, "loss": 0.0288, "step": 1640 }, { "grad_norm": 0.5095893144607544, "learning_rate": 8.008047945352193e-05, "loss": 0.03, "step": 1650 }, { "grad_norm": 0.4521785080432892, "learning_rate": 7.981566834870225e-05, "loss": 0.0315, "step": 1660 }, { "grad_norm": 0.4474700093269348, "learning_rate": 7.954955300995961e-05, "loss": 0.0289, "step": 1670 }, { "grad_norm": 0.5986289978027344, "learning_rate": 7.928214507804104e-05, "loss": 0.0293, "step": 1680 }, { "grad_norm": 0.5038372278213501, "learning_rate": 7.901345625023576e-05, "loss": 0.028, "step": 1690 }, { "grad_norm": 0.4206600785255432, "learning_rate": 7.874349827986354e-05, "loss": 0.0305, "step": 1700 }, { "grad_norm": 0.4808930456638336, "learning_rate": 7.847228297576053e-05, "loss": 0.0311, "step": 1710 }, { "grad_norm": 0.4412468671798706, "learning_rate": 7.819982220176276e-05, "loss": 0.0284, "step": 1720 }, { "grad_norm": 0.47601237893104553, "learning_rate": 7.792612787618714e-05, "loss": 0.0276, "step": 1730 }, { "grad_norm": 0.475136399269104, "learning_rate": 7.765121197131009e-05, "loss": 0.0312, "step": 1740 }, { "grad_norm": 0.4572272300720215, "learning_rate": 7.737508651284391e-05, "loss": 0.0274, "step": 1750 }, { "grad_norm": 0.4734186828136444, "learning_rate": 7.709776357941069e-05, "loss": 0.0301, "step": 1760 }, { "grad_norm": 0.5564247965812683, "learning_rate": 7.681925530201392e-05, "loss": 0.0289, "step": 1770 }, { "grad_norm": 0.513393759727478, "learning_rate": 7.65395738635079e-05, "loss": 0.0292, "step": 1780 }, { "grad_norm": 0.4534357488155365, "learning_rate": 7.62587314980648e-05, "loss": 0.0296, "step": 1790 }, { "grad_norm": 0.4454066753387451, "learning_rate": 7.597674049063947e-05, "loss": 0.0274, "step": 1800 }, { "grad_norm": 0.46227017045021057, "learning_rate": 7.569361317643211e-05, "loss": 0.0273, "step": 1810 }, { "grad_norm": 0.38878199458122253, "learning_rate": 7.540936194034865e-05, "loss": 0.0287, "step": 1820 }, { "grad_norm": 0.5070212483406067, "learning_rate": 7.512399921645901e-05, "loss": 0.0261, "step": 1830 }, { "grad_norm": 0.41168951988220215, "learning_rate": 7.483753748745317e-05, "loss": 0.0265, "step": 1840 }, { "grad_norm": 0.43798619508743286, "learning_rate": 7.454998928409516e-05, "loss": 0.028, "step": 1850 }, { "grad_norm": 0.5544193387031555, "learning_rate": 7.426136718467493e-05, "loss": 0.0264, "step": 1860 }, { "grad_norm": 0.6372482180595398, "learning_rate": 7.397168381445812e-05, "loss": 0.0295, "step": 1870 }, { "grad_norm": 0.41932424902915955, "learning_rate": 7.368095184513377e-05, "loss": 0.0298, "step": 1880 }, { "grad_norm": 0.4176802933216095, "learning_rate": 7.338918399426005e-05, "loss": 0.0259, "step": 1890 }, { "grad_norm": 0.47394803166389465, "learning_rate": 7.309639302470801e-05, "loss": 0.0272, "step": 1900 }, { "grad_norm": 0.5122660398483276, "learning_rate": 7.280259174410312e-05, "loss": 0.031, "step": 1910 }, { "grad_norm": 0.5513008236885071, "learning_rate": 7.250779300426517e-05, "loss": 0.0261, "step": 1920 }, { "grad_norm": 0.39191436767578125, "learning_rate": 7.22120097006461e-05, "loss": 0.0287, "step": 1930 }, { "grad_norm": 0.48576945066452026, "learning_rate": 7.191525477176577e-05, "loss": 0.0262, "step": 1940 }, { "grad_norm": 0.47672203183174133, "learning_rate": 7.161754119864616e-05, "loss": 0.0308, "step": 1950 }, { "grad_norm": 0.5489500164985657, "learning_rate": 7.131888200424339e-05, "loss": 0.0304, "step": 1960 }, { "grad_norm": 0.5475249290466309, "learning_rate": 7.101929025287816e-05, "loss": 0.0284, "step": 1970 }, { "grad_norm": 0.3720989227294922, "learning_rate": 7.071877904966423e-05, "loss": 0.0268, "step": 1980 }, { "grad_norm": 0.45520398020744324, "learning_rate": 7.04173615399351e-05, "loss": 0.0245, "step": 1990 }, { "grad_norm": 0.4413450360298157, "learning_rate": 7.011505090866913e-05, "loss": 0.0267, "step": 2000 } ], "logging_steps": 10, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 9223372036854775807, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 0.0, "train_batch_size": 96, "trial_name": null, "trial_params": null }