| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.999853250501394, | |
| "eval_steps": 500, | |
| "global_step": 30663, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.009783299907058651, | |
| "grad_norm": 33.780609130859375, | |
| "learning_rate": 1.6302575806977503e-06, | |
| "loss": 6.574, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.019566599814117302, | |
| "grad_norm": 33.439022064208984, | |
| "learning_rate": 3.2605151613955006e-06, | |
| "loss": 6.1653, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.02934989972117595, | |
| "grad_norm": 27.25751304626465, | |
| "learning_rate": 4.890772742093251e-06, | |
| "loss": 5.5515, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.039133199628234604, | |
| "grad_norm": 38.36979675292969, | |
| "learning_rate": 6.521030322791001e-06, | |
| "loss": 5.0531, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.048916499535293256, | |
| "grad_norm": 29.350488662719727, | |
| "learning_rate": 8.15128790348875e-06, | |
| "loss": 4.9225, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.048916499535293256, | |
| "eval_runtime": 181.5812, | |
| "eval_samples_per_second": 112.578, | |
| "eval_steps_per_second": 14.076, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.0586997994423519, | |
| "grad_norm": 33.02122497558594, | |
| "learning_rate": 9.781545484186502e-06, | |
| "loss": 4.8186, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.06848309934941056, | |
| "grad_norm": 42.41593933105469, | |
| "learning_rate": 1.1411803064884251e-05, | |
| "loss": 4.5769, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.07826639925646921, | |
| "grad_norm": 40.29044723510742, | |
| "learning_rate": 1.3042060645582003e-05, | |
| "loss": 4.3963, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.08804969916352785, | |
| "grad_norm": 38.0811653137207, | |
| "learning_rate": 1.4672318226279752e-05, | |
| "loss": 4.3393, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.09783299907058651, | |
| "grad_norm": 36.08370590209961, | |
| "learning_rate": 1.63025758069775e-05, | |
| "loss": 4.2421, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.09783299907058651, | |
| "eval_runtime": 181.886, | |
| "eval_samples_per_second": 112.389, | |
| "eval_steps_per_second": 14.053, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.10761629897764516, | |
| "grad_norm": 37.253684997558594, | |
| "learning_rate": 1.7932833387675256e-05, | |
| "loss": 4.1156, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.1173995988847038, | |
| "grad_norm": 33.003475189208984, | |
| "learning_rate": 1.9563090968373004e-05, | |
| "loss": 4.0112, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.12718289879176245, | |
| "grad_norm": 30.727867126464844, | |
| "learning_rate": 2.1193348549070755e-05, | |
| "loss": 3.9969, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.13696619869882112, | |
| "grad_norm": 37.471092224121094, | |
| "learning_rate": 2.2823606129768503e-05, | |
| "loss": 3.874, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.14674949860587977, | |
| "grad_norm": 42.32167434692383, | |
| "learning_rate": 2.4453863710466254e-05, | |
| "loss": 3.8518, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.14674949860587977, | |
| "eval_runtime": 181.9332, | |
| "eval_samples_per_second": 112.36, | |
| "eval_steps_per_second": 14.049, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.15653279851293841, | |
| "grad_norm": 38.00124740600586, | |
| "learning_rate": 2.6084121291164005e-05, | |
| "loss": 3.918, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.16631609841999706, | |
| "grad_norm": 44.637386322021484, | |
| "learning_rate": 2.7714378871861756e-05, | |
| "loss": 3.9134, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.1760993983270557, | |
| "grad_norm": 49.578609466552734, | |
| "learning_rate": 2.9344636452559504e-05, | |
| "loss": 3.7507, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.18588269823411438, | |
| "grad_norm": 36.65715789794922, | |
| "learning_rate": 3.0974894033257255e-05, | |
| "loss": 3.7551, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.19566599814117303, | |
| "grad_norm": 36.873443603515625, | |
| "learning_rate": 3.2605151613955e-05, | |
| "loss": 3.6951, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.19566599814117303, | |
| "eval_runtime": 181.8273, | |
| "eval_samples_per_second": 112.425, | |
| "eval_steps_per_second": 14.057, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.20544929804823167, | |
| "grad_norm": 33.025413513183594, | |
| "learning_rate": 3.423540919465276e-05, | |
| "loss": 3.6603, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.21523259795529032, | |
| "grad_norm": 30.105051040649414, | |
| "learning_rate": 3.586566677535051e-05, | |
| "loss": 3.525, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.22501589786234896, | |
| "grad_norm": 34.5129280090332, | |
| "learning_rate": 3.749592435604825e-05, | |
| "loss": 3.6454, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.2347991977694076, | |
| "grad_norm": 33.16934585571289, | |
| "learning_rate": 3.912618193674601e-05, | |
| "loss": 3.6356, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.24458249767646628, | |
| "grad_norm": 33.5789794921875, | |
| "learning_rate": 4.0756439517443756e-05, | |
| "loss": 3.5605, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.24458249767646628, | |
| "eval_runtime": 181.7254, | |
| "eval_samples_per_second": 112.488, | |
| "eval_steps_per_second": 14.065, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.2543657975835249, | |
| "grad_norm": 34.30876159667969, | |
| "learning_rate": 4.238669709814151e-05, | |
| "loss": 3.5447, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.2641490974905836, | |
| "grad_norm": 29.907989501953125, | |
| "learning_rate": 4.401695467883926e-05, | |
| "loss": 3.5116, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.27393239739764225, | |
| "grad_norm": 34.08231735229492, | |
| "learning_rate": 4.5647212259537006e-05, | |
| "loss": 3.4941, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.28371569730470086, | |
| "grad_norm": 25.034149169921875, | |
| "learning_rate": 4.727746984023476e-05, | |
| "loss": 3.4863, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.29349899721175954, | |
| "grad_norm": 32.21685028076172, | |
| "learning_rate": 4.890772742093251e-05, | |
| "loss": 3.5096, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.29349899721175954, | |
| "eval_runtime": 181.6612, | |
| "eval_samples_per_second": 112.528, | |
| "eval_steps_per_second": 14.07, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.30328229711881816, | |
| "grad_norm": 24.290380477905273, | |
| "learning_rate": 4.9940208725902305e-05, | |
| "loss": 3.3867, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.31306559702587683, | |
| "grad_norm": 22.924575805664062, | |
| "learning_rate": 4.975902304681838e-05, | |
| "loss": 3.398, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.3228488969329355, | |
| "grad_norm": 19.540430068969727, | |
| "learning_rate": 4.957783736773446e-05, | |
| "loss": 3.3727, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.3326321968399941, | |
| "grad_norm": 22.529376983642578, | |
| "learning_rate": 4.939665168865053e-05, | |
| "loss": 3.3364, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.3424154967470528, | |
| "grad_norm": 20.821264266967773, | |
| "learning_rate": 4.921546600956661e-05, | |
| "loss": 3.3126, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.3424154967470528, | |
| "eval_runtime": 181.7582, | |
| "eval_samples_per_second": 112.468, | |
| "eval_steps_per_second": 14.063, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.3521987966541114, | |
| "grad_norm": 24.346153259277344, | |
| "learning_rate": 4.903428033048268e-05, | |
| "loss": 3.2678, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.3619820965611701, | |
| "grad_norm": 19.89035415649414, | |
| "learning_rate": 4.8853094651398754e-05, | |
| "loss": 3.3233, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.37176539646822876, | |
| "grad_norm": 17.938880920410156, | |
| "learning_rate": 4.8671908972314825e-05, | |
| "loss": 3.2822, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.3815486963752874, | |
| "grad_norm": 16.92071533203125, | |
| "learning_rate": 4.84907232932309e-05, | |
| "loss": 3.2254, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.39133199628234605, | |
| "grad_norm": 18.241249084472656, | |
| "learning_rate": 4.830953761414698e-05, | |
| "loss": 3.2116, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.39133199628234605, | |
| "eval_runtime": 182.8906, | |
| "eval_samples_per_second": 111.772, | |
| "eval_steps_per_second": 13.976, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.40111529618940467, | |
| "grad_norm": 17.56020736694336, | |
| "learning_rate": 4.812835193506305e-05, | |
| "loss": 3.2232, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.41089859609646334, | |
| "grad_norm": 17.81117057800293, | |
| "learning_rate": 4.794716625597913e-05, | |
| "loss": 3.1936, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.420681896003522, | |
| "grad_norm": 19.89581871032715, | |
| "learning_rate": 4.77659805768952e-05, | |
| "loss": 3.1443, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.43046519591058063, | |
| "grad_norm": 22.968582153320312, | |
| "learning_rate": 4.758479489781128e-05, | |
| "loss": 3.2084, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.4402484958176393, | |
| "grad_norm": 17.119598388671875, | |
| "learning_rate": 4.740360921872735e-05, | |
| "loss": 3.1263, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.4402484958176393, | |
| "eval_runtime": 182.3246, | |
| "eval_samples_per_second": 112.119, | |
| "eval_steps_per_second": 14.019, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.4500317957246979, | |
| "grad_norm": 19.294527053833008, | |
| "learning_rate": 4.722242353964343e-05, | |
| "loss": 3.1327, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.4598150956317566, | |
| "grad_norm": 16.941057205200195, | |
| "learning_rate": 4.704123786055951e-05, | |
| "loss": 3.0944, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.4695983955388152, | |
| "grad_norm": 22.43411636352539, | |
| "learning_rate": 4.686005218147558e-05, | |
| "loss": 3.1093, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.4793816954458739, | |
| "grad_norm": 19.64097023010254, | |
| "learning_rate": 4.667886650239166e-05, | |
| "loss": 3.0597, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.48916499535293256, | |
| "grad_norm": 19.343788146972656, | |
| "learning_rate": 4.649768082330773e-05, | |
| "loss": 3.1659, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.48916499535293256, | |
| "eval_runtime": 181.8771, | |
| "eval_samples_per_second": 112.395, | |
| "eval_steps_per_second": 14.053, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.4989482952599912, | |
| "grad_norm": 19.657760620117188, | |
| "learning_rate": 4.63164951442238e-05, | |
| "loss": 3.0506, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.5087315951670498, | |
| "grad_norm": 16.2425537109375, | |
| "learning_rate": 4.613530946513987e-05, | |
| "loss": 3.0524, | |
| "step": 5200 | |
| }, | |
| { | |
| "epoch": 0.5185148950741085, | |
| "grad_norm": 19.64779281616211, | |
| "learning_rate": 4.595412378605595e-05, | |
| "loss": 2.9995, | |
| "step": 5300 | |
| }, | |
| { | |
| "epoch": 0.5282981949811671, | |
| "grad_norm": 17.29520606994629, | |
| "learning_rate": 4.577293810697203e-05, | |
| "loss": 3.0932, | |
| "step": 5400 | |
| }, | |
| { | |
| "epoch": 0.5380814948882258, | |
| "grad_norm": 17.694602966308594, | |
| "learning_rate": 4.55917524278881e-05, | |
| "loss": 3.0309, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.5380814948882258, | |
| "eval_runtime": 181.7231, | |
| "eval_samples_per_second": 112.49, | |
| "eval_steps_per_second": 14.065, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.5478647947952845, | |
| "grad_norm": 21.030174255371094, | |
| "learning_rate": 4.541056674880418e-05, | |
| "loss": 3.0313, | |
| "step": 5600 | |
| }, | |
| { | |
| "epoch": 0.5576480947023431, | |
| "grad_norm": 12.339129447937012, | |
| "learning_rate": 4.522938106972025e-05, | |
| "loss": 3.047, | |
| "step": 5700 | |
| }, | |
| { | |
| "epoch": 0.5674313946094017, | |
| "grad_norm": 16.496389389038086, | |
| "learning_rate": 4.504819539063633e-05, | |
| "loss": 2.9961, | |
| "step": 5800 | |
| }, | |
| { | |
| "epoch": 0.5772146945164603, | |
| "grad_norm": 15.456297874450684, | |
| "learning_rate": 4.48670097115524e-05, | |
| "loss": 2.9821, | |
| "step": 5900 | |
| }, | |
| { | |
| "epoch": 0.5869979944235191, | |
| "grad_norm": 17.8603572845459, | |
| "learning_rate": 4.468582403246848e-05, | |
| "loss": 2.9294, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.5869979944235191, | |
| "eval_runtime": 181.8258, | |
| "eval_samples_per_second": 112.426, | |
| "eval_steps_per_second": 14.057, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.5967812943305777, | |
| "grad_norm": 18.85349464416504, | |
| "learning_rate": 4.450463835338455e-05, | |
| "loss": 2.9929, | |
| "step": 6100 | |
| }, | |
| { | |
| "epoch": 0.6065645942376363, | |
| "grad_norm": 22.971813201904297, | |
| "learning_rate": 4.432345267430063e-05, | |
| "loss": 2.9684, | |
| "step": 6200 | |
| }, | |
| { | |
| "epoch": 0.616347894144695, | |
| "grad_norm": 15.877230644226074, | |
| "learning_rate": 4.4142266995216706e-05, | |
| "loss": 2.9399, | |
| "step": 6300 | |
| }, | |
| { | |
| "epoch": 0.6261311940517537, | |
| "grad_norm": 19.847482681274414, | |
| "learning_rate": 4.396108131613278e-05, | |
| "loss": 2.88, | |
| "step": 6400 | |
| }, | |
| { | |
| "epoch": 0.6359144939588123, | |
| "grad_norm": 15.004170417785645, | |
| "learning_rate": 4.377989563704885e-05, | |
| "loss": 2.9719, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.6359144939588123, | |
| "eval_runtime": 182.6045, | |
| "eval_samples_per_second": 111.947, | |
| "eval_steps_per_second": 13.997, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.645697793865871, | |
| "grad_norm": 19.473665237426758, | |
| "learning_rate": 4.359870995796492e-05, | |
| "loss": 2.9246, | |
| "step": 6600 | |
| }, | |
| { | |
| "epoch": 0.6554810937729296, | |
| "grad_norm": 18.071683883666992, | |
| "learning_rate": 4.3417524278881e-05, | |
| "loss": 2.9031, | |
| "step": 6700 | |
| }, | |
| { | |
| "epoch": 0.6652643936799882, | |
| "grad_norm": 17.544504165649414, | |
| "learning_rate": 4.323633859979707e-05, | |
| "loss": 2.8313, | |
| "step": 6800 | |
| }, | |
| { | |
| "epoch": 0.6750476935870469, | |
| "grad_norm": 18.936140060424805, | |
| "learning_rate": 4.305515292071315e-05, | |
| "loss": 2.8536, | |
| "step": 6900 | |
| }, | |
| { | |
| "epoch": 0.6848309934941056, | |
| "grad_norm": 14.77696418762207, | |
| "learning_rate": 4.2873967241629226e-05, | |
| "loss": 2.9104, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.6848309934941056, | |
| "eval_runtime": 181.938, | |
| "eval_samples_per_second": 112.357, | |
| "eval_steps_per_second": 14.049, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.6946142934011642, | |
| "grad_norm": 14.303226470947266, | |
| "learning_rate": 4.26927815625453e-05, | |
| "loss": 2.8386, | |
| "step": 7100 | |
| }, | |
| { | |
| "epoch": 0.7043975933082228, | |
| "grad_norm": 17.11782455444336, | |
| "learning_rate": 4.2511595883461376e-05, | |
| "loss": 2.9013, | |
| "step": 7200 | |
| }, | |
| { | |
| "epoch": 0.7141808932152816, | |
| "grad_norm": 18.661100387573242, | |
| "learning_rate": 4.233041020437745e-05, | |
| "loss": 2.9428, | |
| "step": 7300 | |
| }, | |
| { | |
| "epoch": 0.7239641931223402, | |
| "grad_norm": 15.535719871520996, | |
| "learning_rate": 4.2149224525293525e-05, | |
| "loss": 2.8582, | |
| "step": 7400 | |
| }, | |
| { | |
| "epoch": 0.7337474930293988, | |
| "grad_norm": 15.3306303024292, | |
| "learning_rate": 4.19680388462096e-05, | |
| "loss": 2.8896, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.7337474930293988, | |
| "eval_runtime": 181.8938, | |
| "eval_samples_per_second": 112.384, | |
| "eval_steps_per_second": 14.052, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.7435307929364575, | |
| "grad_norm": 16.730344772338867, | |
| "learning_rate": 4.1786853167125675e-05, | |
| "loss": 2.9097, | |
| "step": 7600 | |
| }, | |
| { | |
| "epoch": 0.7533140928435161, | |
| "grad_norm": 18.755483627319336, | |
| "learning_rate": 4.1605667488041746e-05, | |
| "loss": 2.8815, | |
| "step": 7700 | |
| }, | |
| { | |
| "epoch": 0.7630973927505748, | |
| "grad_norm": 18.737581253051758, | |
| "learning_rate": 4.1424481808957824e-05, | |
| "loss": 2.9202, | |
| "step": 7800 | |
| }, | |
| { | |
| "epoch": 0.7728806926576334, | |
| "grad_norm": 14.711681365966797, | |
| "learning_rate": 4.1243296129873896e-05, | |
| "loss": 2.806, | |
| "step": 7900 | |
| }, | |
| { | |
| "epoch": 0.7826639925646921, | |
| "grad_norm": 17.5069580078125, | |
| "learning_rate": 4.106211045078997e-05, | |
| "loss": 2.8576, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.7826639925646921, | |
| "eval_runtime": 181.9442, | |
| "eval_samples_per_second": 112.353, | |
| "eval_steps_per_second": 14.048, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.7924472924717507, | |
| "grad_norm": 17.678852081298828, | |
| "learning_rate": 4.0880924771706046e-05, | |
| "loss": 2.8035, | |
| "step": 8100 | |
| }, | |
| { | |
| "epoch": 0.8022305923788093, | |
| "grad_norm": 17.644638061523438, | |
| "learning_rate": 4.069973909262212e-05, | |
| "loss": 2.7958, | |
| "step": 8200 | |
| }, | |
| { | |
| "epoch": 0.8120138922858681, | |
| "grad_norm": 18.377134323120117, | |
| "learning_rate": 4.0518553413538195e-05, | |
| "loss": 2.8055, | |
| "step": 8300 | |
| }, | |
| { | |
| "epoch": 0.8217971921929267, | |
| "grad_norm": 18.026033401489258, | |
| "learning_rate": 4.0337367734454273e-05, | |
| "loss": 2.7334, | |
| "step": 8400 | |
| }, | |
| { | |
| "epoch": 0.8315804920999853, | |
| "grad_norm": 14.77315616607666, | |
| "learning_rate": 4.0156182055370345e-05, | |
| "loss": 2.8082, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.8315804920999853, | |
| "eval_runtime": 182.4176, | |
| "eval_samples_per_second": 112.062, | |
| "eval_steps_per_second": 14.012, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.841363792007044, | |
| "grad_norm": 13.729479789733887, | |
| "learning_rate": 3.997499637628642e-05, | |
| "loss": 2.7939, | |
| "step": 8600 | |
| }, | |
| { | |
| "epoch": 0.8511470919141026, | |
| "grad_norm": 16.34333610534668, | |
| "learning_rate": 3.9793810697202494e-05, | |
| "loss": 2.8517, | |
| "step": 8700 | |
| }, | |
| { | |
| "epoch": 0.8609303918211613, | |
| "grad_norm": 22.484411239624023, | |
| "learning_rate": 3.961262501811857e-05, | |
| "loss": 2.776, | |
| "step": 8800 | |
| }, | |
| { | |
| "epoch": 0.8707136917282199, | |
| "grad_norm": 15.922870635986328, | |
| "learning_rate": 3.9431439339034644e-05, | |
| "loss": 2.7909, | |
| "step": 8900 | |
| }, | |
| { | |
| "epoch": 0.8804969916352786, | |
| "grad_norm": 15.06955623626709, | |
| "learning_rate": 3.925025365995072e-05, | |
| "loss": 2.8416, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.8804969916352786, | |
| "eval_runtime": 181.9314, | |
| "eval_samples_per_second": 112.361, | |
| "eval_steps_per_second": 14.049, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 0.8902802915423372, | |
| "grad_norm": 16.060428619384766, | |
| "learning_rate": 3.9069067980866794e-05, | |
| "loss": 2.7803, | |
| "step": 9100 | |
| }, | |
| { | |
| "epoch": 0.9000635914493959, | |
| "grad_norm": 16.80124855041504, | |
| "learning_rate": 3.888788230178287e-05, | |
| "loss": 2.7548, | |
| "step": 9200 | |
| }, | |
| { | |
| "epoch": 0.9098468913564546, | |
| "grad_norm": 16.608434677124023, | |
| "learning_rate": 3.870669662269894e-05, | |
| "loss": 2.8606, | |
| "step": 9300 | |
| }, | |
| { | |
| "epoch": 0.9196301912635132, | |
| "grad_norm": 14.83870792388916, | |
| "learning_rate": 3.8525510943615015e-05, | |
| "loss": 2.7833, | |
| "step": 9400 | |
| }, | |
| { | |
| "epoch": 0.9294134911705718, | |
| "grad_norm": 25.778181076049805, | |
| "learning_rate": 3.834432526453109e-05, | |
| "loss": 2.7434, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.9294134911705718, | |
| "eval_runtime": 181.99, | |
| "eval_samples_per_second": 112.325, | |
| "eval_steps_per_second": 14.045, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 0.9391967910776304, | |
| "grad_norm": 17.374011993408203, | |
| "learning_rate": 3.8163139585447164e-05, | |
| "loss": 2.7258, | |
| "step": 9600 | |
| }, | |
| { | |
| "epoch": 0.9489800909846892, | |
| "grad_norm": 17.551128387451172, | |
| "learning_rate": 3.798195390636324e-05, | |
| "loss": 2.824, | |
| "step": 9700 | |
| }, | |
| { | |
| "epoch": 0.9587633908917478, | |
| "grad_norm": 14.35797119140625, | |
| "learning_rate": 3.7800768227279314e-05, | |
| "loss": 2.745, | |
| "step": 9800 | |
| }, | |
| { | |
| "epoch": 0.9685466907988064, | |
| "grad_norm": 20.098552703857422, | |
| "learning_rate": 3.761958254819539e-05, | |
| "loss": 2.7025, | |
| "step": 9900 | |
| }, | |
| { | |
| "epoch": 0.9783299907058651, | |
| "grad_norm": 16.218109130859375, | |
| "learning_rate": 3.743839686911147e-05, | |
| "loss": 2.8093, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.9783299907058651, | |
| "eval_runtime": 181.8987, | |
| "eval_samples_per_second": 112.381, | |
| "eval_steps_per_second": 14.052, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 0.9881132906129237, | |
| "grad_norm": 17.198423385620117, | |
| "learning_rate": 3.725721119002754e-05, | |
| "loss": 2.7124, | |
| "step": 10100 | |
| }, | |
| { | |
| "epoch": 0.9978965905199824, | |
| "grad_norm": 18.021198272705078, | |
| "learning_rate": 3.707602551094362e-05, | |
| "loss": 2.6922, | |
| "step": 10200 | |
| }, | |
| { | |
| "epoch": 1.007679890427041, | |
| "grad_norm": 15.27678108215332, | |
| "learning_rate": 3.689483983185969e-05, | |
| "loss": 2.6743, | |
| "step": 10300 | |
| }, | |
| { | |
| "epoch": 1.0174631903340996, | |
| "grad_norm": 16.770511627197266, | |
| "learning_rate": 3.671365415277577e-05, | |
| "loss": 2.857, | |
| "step": 10400 | |
| }, | |
| { | |
| "epoch": 1.0272464902411584, | |
| "grad_norm": 18.810932159423828, | |
| "learning_rate": 3.653246847369184e-05, | |
| "loss": 2.7269, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.0272464902411584, | |
| "eval_runtime": 181.8537, | |
| "eval_samples_per_second": 112.409, | |
| "eval_steps_per_second": 14.055, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.037029790148217, | |
| "grad_norm": 18.56201171875, | |
| "learning_rate": 3.635128279460791e-05, | |
| "loss": 2.7325, | |
| "step": 10600 | |
| }, | |
| { | |
| "epoch": 1.0468130900552757, | |
| "grad_norm": 15.063011169433594, | |
| "learning_rate": 3.617009711552399e-05, | |
| "loss": 2.7827, | |
| "step": 10700 | |
| }, | |
| { | |
| "epoch": 1.0565963899623343, | |
| "grad_norm": 15.339439392089844, | |
| "learning_rate": 3.598891143644006e-05, | |
| "loss": 2.7472, | |
| "step": 10800 | |
| }, | |
| { | |
| "epoch": 1.066379689869393, | |
| "grad_norm": 17.466033935546875, | |
| "learning_rate": 3.580772575735614e-05, | |
| "loss": 2.7859, | |
| "step": 10900 | |
| }, | |
| { | |
| "epoch": 1.0761629897764515, | |
| "grad_norm": 20.727872848510742, | |
| "learning_rate": 3.562654007827221e-05, | |
| "loss": 2.7278, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.0761629897764515, | |
| "eval_runtime": 181.8566, | |
| "eval_samples_per_second": 112.407, | |
| "eval_steps_per_second": 14.055, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.0859462896835101, | |
| "grad_norm": 16.02055549621582, | |
| "learning_rate": 3.544535439918829e-05, | |
| "loss": 2.6307, | |
| "step": 11100 | |
| }, | |
| { | |
| "epoch": 1.095729589590569, | |
| "grad_norm": 20.069686889648438, | |
| "learning_rate": 3.526416872010436e-05, | |
| "loss": 2.711, | |
| "step": 11200 | |
| }, | |
| { | |
| "epoch": 1.1055128894976276, | |
| "grad_norm": 14.833261489868164, | |
| "learning_rate": 3.508298304102044e-05, | |
| "loss": 2.6141, | |
| "step": 11300 | |
| }, | |
| { | |
| "epoch": 1.1152961894046862, | |
| "grad_norm": 14.86436653137207, | |
| "learning_rate": 3.490179736193652e-05, | |
| "loss": 2.6816, | |
| "step": 11400 | |
| }, | |
| { | |
| "epoch": 1.1250794893117448, | |
| "grad_norm": 17.955862045288086, | |
| "learning_rate": 3.472061168285259e-05, | |
| "loss": 2.6924, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.1250794893117448, | |
| "eval_runtime": 181.8085, | |
| "eval_samples_per_second": 112.437, | |
| "eval_steps_per_second": 14.059, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.1348627892188035, | |
| "grad_norm": 18.360109329223633, | |
| "learning_rate": 3.453942600376867e-05, | |
| "loss": 2.6181, | |
| "step": 11600 | |
| }, | |
| { | |
| "epoch": 1.144646089125862, | |
| "grad_norm": 17.547542572021484, | |
| "learning_rate": 3.435824032468474e-05, | |
| "loss": 2.6394, | |
| "step": 11700 | |
| }, | |
| { | |
| "epoch": 1.154429389032921, | |
| "grad_norm": 12.194833755493164, | |
| "learning_rate": 3.417705464560082e-05, | |
| "loss": 2.6684, | |
| "step": 11800 | |
| }, | |
| { | |
| "epoch": 1.1642126889399795, | |
| "grad_norm": 17.095104217529297, | |
| "learning_rate": 3.399586896651689e-05, | |
| "loss": 2.6129, | |
| "step": 11900 | |
| }, | |
| { | |
| "epoch": 1.1739959888470382, | |
| "grad_norm": 20.788406372070312, | |
| "learning_rate": 3.381468328743296e-05, | |
| "loss": 2.5663, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.1739959888470382, | |
| "eval_runtime": 181.8035, | |
| "eval_samples_per_second": 112.44, | |
| "eval_steps_per_second": 14.059, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.1837792887540968, | |
| "grad_norm": 14.261167526245117, | |
| "learning_rate": 3.363349760834904e-05, | |
| "loss": 2.6544, | |
| "step": 12100 | |
| }, | |
| { | |
| "epoch": 1.1935625886611554, | |
| "grad_norm": 24.68012046813965, | |
| "learning_rate": 3.345231192926511e-05, | |
| "loss": 2.6632, | |
| "step": 12200 | |
| }, | |
| { | |
| "epoch": 1.203345888568214, | |
| "grad_norm": 16.10886573791504, | |
| "learning_rate": 3.327112625018119e-05, | |
| "loss": 2.6366, | |
| "step": 12300 | |
| }, | |
| { | |
| "epoch": 1.2131291884752726, | |
| "grad_norm": 18.038848876953125, | |
| "learning_rate": 3.308994057109726e-05, | |
| "loss": 2.6563, | |
| "step": 12400 | |
| }, | |
| { | |
| "epoch": 1.2229124883823315, | |
| "grad_norm": 17.40920639038086, | |
| "learning_rate": 3.290875489201334e-05, | |
| "loss": 2.718, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.2229124883823315, | |
| "eval_runtime": 181.9491, | |
| "eval_samples_per_second": 112.35, | |
| "eval_steps_per_second": 14.048, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.23269578828939, | |
| "grad_norm": 15.097307205200195, | |
| "learning_rate": 3.272756921292941e-05, | |
| "loss": 2.7282, | |
| "step": 12600 | |
| }, | |
| { | |
| "epoch": 1.2424790881964487, | |
| "grad_norm": 17.63008689880371, | |
| "learning_rate": 3.254638353384549e-05, | |
| "loss": 2.7104, | |
| "step": 12700 | |
| }, | |
| { | |
| "epoch": 1.2522623881035073, | |
| "grad_norm": 16.161130905151367, | |
| "learning_rate": 3.236519785476156e-05, | |
| "loss": 2.6427, | |
| "step": 12800 | |
| }, | |
| { | |
| "epoch": 1.262045688010566, | |
| "grad_norm": 18.786882400512695, | |
| "learning_rate": 3.218401217567764e-05, | |
| "loss": 2.6105, | |
| "step": 12900 | |
| }, | |
| { | |
| "epoch": 1.2718289879176246, | |
| "grad_norm": 24.145421981811523, | |
| "learning_rate": 3.2002826496593715e-05, | |
| "loss": 2.6322, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.2718289879176246, | |
| "eval_runtime": 182.5613, | |
| "eval_samples_per_second": 111.973, | |
| "eval_steps_per_second": 14.001, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.2816122878246832, | |
| "grad_norm": 15.286133766174316, | |
| "learning_rate": 3.1821640817509786e-05, | |
| "loss": 2.6465, | |
| "step": 13100 | |
| }, | |
| { | |
| "epoch": 1.291395587731742, | |
| "grad_norm": 21.22935676574707, | |
| "learning_rate": 3.1640455138425865e-05, | |
| "loss": 2.6691, | |
| "step": 13200 | |
| }, | |
| { | |
| "epoch": 1.3011788876388006, | |
| "grad_norm": 18.064428329467773, | |
| "learning_rate": 3.1459269459341936e-05, | |
| "loss": 2.5904, | |
| "step": 13300 | |
| }, | |
| { | |
| "epoch": 1.3109621875458592, | |
| "grad_norm": 14.45976448059082, | |
| "learning_rate": 3.127808378025801e-05, | |
| "loss": 2.6602, | |
| "step": 13400 | |
| }, | |
| { | |
| "epoch": 1.3207454874529179, | |
| "grad_norm": 19.72386360168457, | |
| "learning_rate": 3.109689810117408e-05, | |
| "loss": 2.6337, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.3207454874529179, | |
| "eval_runtime": 182.4053, | |
| "eval_samples_per_second": 112.069, | |
| "eval_steps_per_second": 14.013, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.3305287873599765, | |
| "grad_norm": 17.639583587646484, | |
| "learning_rate": 3.091571242209016e-05, | |
| "loss": 2.6135, | |
| "step": 13600 | |
| }, | |
| { | |
| "epoch": 1.340312087267035, | |
| "grad_norm": 19.71700096130371, | |
| "learning_rate": 3.0734526743006235e-05, | |
| "loss": 2.6252, | |
| "step": 13700 | |
| }, | |
| { | |
| "epoch": 1.3500953871740937, | |
| "grad_norm": 16.715856552124023, | |
| "learning_rate": 3.055334106392231e-05, | |
| "loss": 2.6475, | |
| "step": 13800 | |
| }, | |
| { | |
| "epoch": 1.3598786870811526, | |
| "grad_norm": 12.645075798034668, | |
| "learning_rate": 3.0372155384838385e-05, | |
| "loss": 2.6199, | |
| "step": 13900 | |
| }, | |
| { | |
| "epoch": 1.3696619869882112, | |
| "grad_norm": 20.150625228881836, | |
| "learning_rate": 3.0190969705754456e-05, | |
| "loss": 2.5567, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.3696619869882112, | |
| "eval_runtime": 181.9086, | |
| "eval_samples_per_second": 112.375, | |
| "eval_steps_per_second": 14.051, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.3794452868952698, | |
| "grad_norm": 19.111286163330078, | |
| "learning_rate": 3.0009784026670535e-05, | |
| "loss": 2.59, | |
| "step": 14100 | |
| }, | |
| { | |
| "epoch": 1.3892285868023284, | |
| "grad_norm": 17.12226104736328, | |
| "learning_rate": 2.9828598347586606e-05, | |
| "loss": 2.5913, | |
| "step": 14200 | |
| }, | |
| { | |
| "epoch": 1.399011886709387, | |
| "grad_norm": 19.741445541381836, | |
| "learning_rate": 2.9647412668502684e-05, | |
| "loss": 2.5617, | |
| "step": 14300 | |
| }, | |
| { | |
| "epoch": 1.4087951866164456, | |
| "grad_norm": 17.605525970458984, | |
| "learning_rate": 2.946622698941876e-05, | |
| "loss": 2.6077, | |
| "step": 14400 | |
| }, | |
| { | |
| "epoch": 1.4185784865235043, | |
| "grad_norm": 17.433218002319336, | |
| "learning_rate": 2.928504131033483e-05, | |
| "loss": 2.5713, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.4185784865235043, | |
| "eval_runtime": 181.9305, | |
| "eval_samples_per_second": 112.362, | |
| "eval_steps_per_second": 14.049, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.428361786430563, | |
| "grad_norm": 15.442538261413574, | |
| "learning_rate": 2.910385563125091e-05, | |
| "loss": 2.6499, | |
| "step": 14600 | |
| }, | |
| { | |
| "epoch": 1.4381450863376217, | |
| "grad_norm": 15.078730583190918, | |
| "learning_rate": 2.892266995216698e-05, | |
| "loss": 2.6517, | |
| "step": 14700 | |
| }, | |
| { | |
| "epoch": 1.4479283862446803, | |
| "grad_norm": 23.07891273498535, | |
| "learning_rate": 2.874148427308306e-05, | |
| "loss": 2.594, | |
| "step": 14800 | |
| }, | |
| { | |
| "epoch": 1.457711686151739, | |
| "grad_norm": 16.707923889160156, | |
| "learning_rate": 2.856029859399913e-05, | |
| "loss": 2.6613, | |
| "step": 14900 | |
| }, | |
| { | |
| "epoch": 1.4674949860587976, | |
| "grad_norm": 16.731164932250977, | |
| "learning_rate": 2.8379112914915208e-05, | |
| "loss": 2.5927, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.4674949860587976, | |
| "eval_runtime": 181.9649, | |
| "eval_samples_per_second": 112.34, | |
| "eval_steps_per_second": 14.047, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.4772782859658564, | |
| "grad_norm": 16.020864486694336, | |
| "learning_rate": 2.819792723583128e-05, | |
| "loss": 2.6464, | |
| "step": 15100 | |
| }, | |
| { | |
| "epoch": 1.4870615858729148, | |
| "grad_norm": 16.674760818481445, | |
| "learning_rate": 2.8016741556747354e-05, | |
| "loss": 2.5853, | |
| "step": 15200 | |
| }, | |
| { | |
| "epoch": 1.4968448857799737, | |
| "grad_norm": 16.890748977661133, | |
| "learning_rate": 2.7835555877663432e-05, | |
| "loss": 2.5748, | |
| "step": 15300 | |
| }, | |
| { | |
| "epoch": 1.5066281856870323, | |
| "grad_norm": 20.217845916748047, | |
| "learning_rate": 2.7654370198579504e-05, | |
| "loss": 2.6204, | |
| "step": 15400 | |
| }, | |
| { | |
| "epoch": 1.516411485594091, | |
| "grad_norm": 20.459087371826172, | |
| "learning_rate": 2.7473184519495582e-05, | |
| "loss": 2.6103, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.516411485594091, | |
| "eval_runtime": 181.9454, | |
| "eval_samples_per_second": 112.352, | |
| "eval_steps_per_second": 14.048, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.5261947855011495, | |
| "grad_norm": 18.207612991333008, | |
| "learning_rate": 2.7291998840411654e-05, | |
| "loss": 2.5786, | |
| "step": 15600 | |
| }, | |
| { | |
| "epoch": 1.5359780854082081, | |
| "grad_norm": 18.084758758544922, | |
| "learning_rate": 2.7110813161327732e-05, | |
| "loss": 2.6535, | |
| "step": 15700 | |
| }, | |
| { | |
| "epoch": 1.545761385315267, | |
| "grad_norm": 15.03881549835205, | |
| "learning_rate": 2.6929627482243803e-05, | |
| "loss": 2.6061, | |
| "step": 15800 | |
| }, | |
| { | |
| "epoch": 1.5555446852223254, | |
| "grad_norm": 16.99995231628418, | |
| "learning_rate": 2.6748441803159878e-05, | |
| "loss": 2.6151, | |
| "step": 15900 | |
| }, | |
| { | |
| "epoch": 1.5653279851293842, | |
| "grad_norm": 15.581089973449707, | |
| "learning_rate": 2.6567256124075956e-05, | |
| "loss": 2.6163, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.5653279851293842, | |
| "eval_runtime": 181.8152, | |
| "eval_samples_per_second": 112.433, | |
| "eval_steps_per_second": 14.058, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.5751112850364428, | |
| "grad_norm": 21.4382266998291, | |
| "learning_rate": 2.6386070444992028e-05, | |
| "loss": 2.5975, | |
| "step": 16100 | |
| }, | |
| { | |
| "epoch": 1.5848945849435014, | |
| "grad_norm": 15.874536514282227, | |
| "learning_rate": 2.6204884765908106e-05, | |
| "loss": 2.5851, | |
| "step": 16200 | |
| }, | |
| { | |
| "epoch": 1.59467788485056, | |
| "grad_norm": 17.902137756347656, | |
| "learning_rate": 2.6023699086824177e-05, | |
| "loss": 2.6027, | |
| "step": 16300 | |
| }, | |
| { | |
| "epoch": 1.6044611847576187, | |
| "grad_norm": 17.04872703552246, | |
| "learning_rate": 2.5842513407740255e-05, | |
| "loss": 2.5854, | |
| "step": 16400 | |
| }, | |
| { | |
| "epoch": 1.6142444846646775, | |
| "grad_norm": 15.406013488769531, | |
| "learning_rate": 2.5661327728656327e-05, | |
| "loss": 2.5158, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.6142444846646775, | |
| "eval_runtime": 181.8647, | |
| "eval_samples_per_second": 112.402, | |
| "eval_steps_per_second": 14.054, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.624027784571736, | |
| "grad_norm": 19.62627601623535, | |
| "learning_rate": 2.5480142049572402e-05, | |
| "loss": 2.5378, | |
| "step": 16600 | |
| }, | |
| { | |
| "epoch": 1.6338110844787948, | |
| "grad_norm": 17.825178146362305, | |
| "learning_rate": 2.529895637048848e-05, | |
| "loss": 2.6162, | |
| "step": 16700 | |
| }, | |
| { | |
| "epoch": 1.6435943843858534, | |
| "grad_norm": 15.442023277282715, | |
| "learning_rate": 2.511777069140455e-05, | |
| "loss": 2.5802, | |
| "step": 16800 | |
| }, | |
| { | |
| "epoch": 1.653377684292912, | |
| "grad_norm": 18.695241928100586, | |
| "learning_rate": 2.4936585012320626e-05, | |
| "loss": 2.585, | |
| "step": 16900 | |
| }, | |
| { | |
| "epoch": 1.6631609841999706, | |
| "grad_norm": 18.992969512939453, | |
| "learning_rate": 2.4755399333236704e-05, | |
| "loss": 2.5448, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.6631609841999706, | |
| "eval_runtime": 181.91, | |
| "eval_samples_per_second": 112.374, | |
| "eval_steps_per_second": 14.051, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.6729442841070292, | |
| "grad_norm": 19.065349578857422, | |
| "learning_rate": 2.457421365415278e-05, | |
| "loss": 2.6565, | |
| "step": 17100 | |
| }, | |
| { | |
| "epoch": 1.682727584014088, | |
| "grad_norm": 20.110734939575195, | |
| "learning_rate": 2.439302797506885e-05, | |
| "loss": 2.5519, | |
| "step": 17200 | |
| }, | |
| { | |
| "epoch": 1.6925108839211465, | |
| "grad_norm": 15.886931419372559, | |
| "learning_rate": 2.4211842295984925e-05, | |
| "loss": 2.5589, | |
| "step": 17300 | |
| }, | |
| { | |
| "epoch": 1.7022941838282053, | |
| "grad_norm": 19.213207244873047, | |
| "learning_rate": 2.4030656616901e-05, | |
| "loss": 2.5714, | |
| "step": 17400 | |
| }, | |
| { | |
| "epoch": 1.712077483735264, | |
| "grad_norm": 17.117481231689453, | |
| "learning_rate": 2.3849470937817075e-05, | |
| "loss": 2.6682, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.712077483735264, | |
| "eval_runtime": 181.766, | |
| "eval_samples_per_second": 112.463, | |
| "eval_steps_per_second": 14.062, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.7218607836423225, | |
| "grad_norm": 17.19162940979004, | |
| "learning_rate": 2.366828525873315e-05, | |
| "loss": 2.5591, | |
| "step": 17600 | |
| }, | |
| { | |
| "epoch": 1.7316440835493812, | |
| "grad_norm": 15.454411506652832, | |
| "learning_rate": 2.3487099579649225e-05, | |
| "loss": 2.469, | |
| "step": 17700 | |
| }, | |
| { | |
| "epoch": 1.7414273834564398, | |
| "grad_norm": 15.227791786193848, | |
| "learning_rate": 2.3305913900565303e-05, | |
| "loss": 2.664, | |
| "step": 17800 | |
| }, | |
| { | |
| "epoch": 1.7512106833634986, | |
| "grad_norm": 18.5739688873291, | |
| "learning_rate": 2.3124728221481374e-05, | |
| "loss": 2.5991, | |
| "step": 17900 | |
| }, | |
| { | |
| "epoch": 1.760993983270557, | |
| "grad_norm": 12.589066505432129, | |
| "learning_rate": 2.294354254239745e-05, | |
| "loss": 2.6593, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.760993983270557, | |
| "eval_runtime": 181.9699, | |
| "eval_samples_per_second": 112.337, | |
| "eval_steps_per_second": 14.046, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 1.7707772831776158, | |
| "grad_norm": 20.695772171020508, | |
| "learning_rate": 2.2762356863313524e-05, | |
| "loss": 2.5555, | |
| "step": 18100 | |
| }, | |
| { | |
| "epoch": 1.7805605830846745, | |
| "grad_norm": 12.731703758239746, | |
| "learning_rate": 2.25811711842296e-05, | |
| "loss": 2.4617, | |
| "step": 18200 | |
| }, | |
| { | |
| "epoch": 1.790343882991733, | |
| "grad_norm": 18.506074905395508, | |
| "learning_rate": 2.2399985505145674e-05, | |
| "loss": 2.6061, | |
| "step": 18300 | |
| }, | |
| { | |
| "epoch": 1.800127182898792, | |
| "grad_norm": 14.8694486618042, | |
| "learning_rate": 2.221879982606175e-05, | |
| "loss": 2.5779, | |
| "step": 18400 | |
| }, | |
| { | |
| "epoch": 1.8099104828058503, | |
| "grad_norm": 22.47985076904297, | |
| "learning_rate": 2.2037614146977827e-05, | |
| "loss": 2.5012, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.8099104828058503, | |
| "eval_runtime": 182.3919, | |
| "eval_samples_per_second": 112.077, | |
| "eval_steps_per_second": 14.014, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 1.8196937827129092, | |
| "grad_norm": 25.74334144592285, | |
| "learning_rate": 2.1856428467893898e-05, | |
| "loss": 2.5265, | |
| "step": 18600 | |
| }, | |
| { | |
| "epoch": 1.8294770826199676, | |
| "grad_norm": 18.477630615234375, | |
| "learning_rate": 2.1675242788809973e-05, | |
| "loss": 2.5555, | |
| "step": 18700 | |
| }, | |
| { | |
| "epoch": 1.8392603825270264, | |
| "grad_norm": 14.832316398620605, | |
| "learning_rate": 2.1494057109726048e-05, | |
| "loss": 2.4609, | |
| "step": 18800 | |
| }, | |
| { | |
| "epoch": 1.849043682434085, | |
| "grad_norm": 17.025096893310547, | |
| "learning_rate": 2.1312871430642123e-05, | |
| "loss": 2.5119, | |
| "step": 18900 | |
| }, | |
| { | |
| "epoch": 1.8588269823411436, | |
| "grad_norm": 16.852436065673828, | |
| "learning_rate": 2.1131685751558197e-05, | |
| "loss": 2.5369, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.8588269823411436, | |
| "eval_runtime": 181.7443, | |
| "eval_samples_per_second": 112.477, | |
| "eval_steps_per_second": 14.064, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 1.8686102822482025, | |
| "grad_norm": 15.160259246826172, | |
| "learning_rate": 2.0950500072474272e-05, | |
| "loss": 2.6297, | |
| "step": 19100 | |
| }, | |
| { | |
| "epoch": 1.8783935821552609, | |
| "grad_norm": 15.909671783447266, | |
| "learning_rate": 2.0769314393390347e-05, | |
| "loss": 2.4696, | |
| "step": 19200 | |
| }, | |
| { | |
| "epoch": 1.8881768820623197, | |
| "grad_norm": 14.201844215393066, | |
| "learning_rate": 2.0588128714306422e-05, | |
| "loss": 2.5653, | |
| "step": 19300 | |
| }, | |
| { | |
| "epoch": 1.8979601819693783, | |
| "grad_norm": 16.351415634155273, | |
| "learning_rate": 2.0406943035222497e-05, | |
| "loss": 2.4962, | |
| "step": 19400 | |
| }, | |
| { | |
| "epoch": 1.907743481876437, | |
| "grad_norm": 16.943771362304688, | |
| "learning_rate": 2.022575735613857e-05, | |
| "loss": 2.5091, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.907743481876437, | |
| "eval_runtime": 181.6486, | |
| "eval_samples_per_second": 112.536, | |
| "eval_steps_per_second": 14.071, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 1.9175267817834956, | |
| "grad_norm": 15.006349563598633, | |
| "learning_rate": 2.0044571677054646e-05, | |
| "loss": 2.5214, | |
| "step": 19600 | |
| }, | |
| { | |
| "epoch": 1.9273100816905542, | |
| "grad_norm": 17.305580139160156, | |
| "learning_rate": 1.986338599797072e-05, | |
| "loss": 2.4989, | |
| "step": 19700 | |
| }, | |
| { | |
| "epoch": 1.937093381597613, | |
| "grad_norm": 17.28044891357422, | |
| "learning_rate": 1.9682200318886796e-05, | |
| "loss": 2.4008, | |
| "step": 19800 | |
| }, | |
| { | |
| "epoch": 1.9468766815046714, | |
| "grad_norm": 18.25079917907715, | |
| "learning_rate": 1.950101463980287e-05, | |
| "loss": 2.6015, | |
| "step": 19900 | |
| }, | |
| { | |
| "epoch": 1.9566599814117303, | |
| "grad_norm": 20.741668701171875, | |
| "learning_rate": 1.9319828960718946e-05, | |
| "loss": 2.4081, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 1.9566599814117303, | |
| "eval_runtime": 181.7745, | |
| "eval_samples_per_second": 112.458, | |
| "eval_steps_per_second": 14.061, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 1.9664432813187889, | |
| "grad_norm": 16.1226863861084, | |
| "learning_rate": 1.913864328163502e-05, | |
| "loss": 2.5418, | |
| "step": 20100 | |
| }, | |
| { | |
| "epoch": 1.9762265812258475, | |
| "grad_norm": 13.914982795715332, | |
| "learning_rate": 1.8957457602551095e-05, | |
| "loss": 2.5248, | |
| "step": 20200 | |
| }, | |
| { | |
| "epoch": 1.986009881132906, | |
| "grad_norm": 15.072690963745117, | |
| "learning_rate": 1.877627192346717e-05, | |
| "loss": 2.5488, | |
| "step": 20300 | |
| }, | |
| { | |
| "epoch": 1.9957931810399647, | |
| "grad_norm": 15.510763168334961, | |
| "learning_rate": 1.8595086244383245e-05, | |
| "loss": 2.4605, | |
| "step": 20400 | |
| }, | |
| { | |
| "epoch": 2.0055764809470236, | |
| "grad_norm": 18.463842391967773, | |
| "learning_rate": 1.841390056529932e-05, | |
| "loss": 2.522, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.0055764809470236, | |
| "eval_runtime": 182.07, | |
| "eval_samples_per_second": 112.276, | |
| "eval_steps_per_second": 14.039, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.015359780854082, | |
| "grad_norm": 16.670269012451172, | |
| "learning_rate": 1.8232714886215394e-05, | |
| "loss": 2.5585, | |
| "step": 20600 | |
| }, | |
| { | |
| "epoch": 2.025143080761141, | |
| "grad_norm": 20.60368537902832, | |
| "learning_rate": 1.805152920713147e-05, | |
| "loss": 2.5381, | |
| "step": 20700 | |
| }, | |
| { | |
| "epoch": 2.034926380668199, | |
| "grad_norm": 15.686981201171875, | |
| "learning_rate": 1.7870343528047544e-05, | |
| "loss": 2.5721, | |
| "step": 20800 | |
| }, | |
| { | |
| "epoch": 2.044709680575258, | |
| "grad_norm": 14.691718101501465, | |
| "learning_rate": 1.768915784896362e-05, | |
| "loss": 2.5187, | |
| "step": 20900 | |
| }, | |
| { | |
| "epoch": 2.054492980482317, | |
| "grad_norm": 16.31734848022461, | |
| "learning_rate": 1.7507972169879694e-05, | |
| "loss": 2.5202, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.054492980482317, | |
| "eval_runtime": 181.9896, | |
| "eval_samples_per_second": 112.325, | |
| "eval_steps_per_second": 14.045, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.0642762803893753, | |
| "grad_norm": 12.698554992675781, | |
| "learning_rate": 1.732678649079577e-05, | |
| "loss": 2.4228, | |
| "step": 21100 | |
| }, | |
| { | |
| "epoch": 2.074059580296434, | |
| "grad_norm": 16.34201431274414, | |
| "learning_rate": 1.7145600811711843e-05, | |
| "loss": 2.3963, | |
| "step": 21200 | |
| }, | |
| { | |
| "epoch": 2.0838428802034925, | |
| "grad_norm": 16.52840232849121, | |
| "learning_rate": 1.6964415132627918e-05, | |
| "loss": 2.4759, | |
| "step": 21300 | |
| }, | |
| { | |
| "epoch": 2.0936261801105513, | |
| "grad_norm": 14.856452941894531, | |
| "learning_rate": 1.6783229453543993e-05, | |
| "loss": 2.4675, | |
| "step": 21400 | |
| }, | |
| { | |
| "epoch": 2.1034094800176097, | |
| "grad_norm": 19.68895721435547, | |
| "learning_rate": 1.6602043774460068e-05, | |
| "loss": 2.5324, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.1034094800176097, | |
| "eval_runtime": 182.1877, | |
| "eval_samples_per_second": 112.203, | |
| "eval_steps_per_second": 14.029, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.1131927799246686, | |
| "grad_norm": 23.248056411743164, | |
| "learning_rate": 1.6420858095376143e-05, | |
| "loss": 2.5231, | |
| "step": 21600 | |
| }, | |
| { | |
| "epoch": 2.1229760798317274, | |
| "grad_norm": 25.471004486083984, | |
| "learning_rate": 1.6239672416292217e-05, | |
| "loss": 2.5871, | |
| "step": 21700 | |
| }, | |
| { | |
| "epoch": 2.132759379738786, | |
| "grad_norm": 17.794851303100586, | |
| "learning_rate": 1.6058486737208292e-05, | |
| "loss": 2.5008, | |
| "step": 21800 | |
| }, | |
| { | |
| "epoch": 2.1425426796458447, | |
| "grad_norm": 15.450346946716309, | |
| "learning_rate": 1.5877301058124367e-05, | |
| "loss": 2.4194, | |
| "step": 21900 | |
| }, | |
| { | |
| "epoch": 2.152325979552903, | |
| "grad_norm": 13.243645668029785, | |
| "learning_rate": 1.5696115379040442e-05, | |
| "loss": 2.5018, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.152325979552903, | |
| "eval_runtime": 181.9841, | |
| "eval_samples_per_second": 112.328, | |
| "eval_steps_per_second": 14.045, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.162109279459962, | |
| "grad_norm": 16.996198654174805, | |
| "learning_rate": 1.5514929699956517e-05, | |
| "loss": 2.4492, | |
| "step": 22100 | |
| }, | |
| { | |
| "epoch": 2.1718925793670203, | |
| "grad_norm": 20.05558967590332, | |
| "learning_rate": 1.5333744020872588e-05, | |
| "loss": 2.489, | |
| "step": 22200 | |
| }, | |
| { | |
| "epoch": 2.181675879274079, | |
| "grad_norm": 15.66326904296875, | |
| "learning_rate": 1.5152558341788666e-05, | |
| "loss": 2.5089, | |
| "step": 22300 | |
| }, | |
| { | |
| "epoch": 2.191459179181138, | |
| "grad_norm": 17.83564567565918, | |
| "learning_rate": 1.4971372662704741e-05, | |
| "loss": 2.4945, | |
| "step": 22400 | |
| }, | |
| { | |
| "epoch": 2.2012424790881964, | |
| "grad_norm": 21.466899871826172, | |
| "learning_rate": 1.4790186983620816e-05, | |
| "loss": 2.5467, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 2.2012424790881964, | |
| "eval_runtime": 182.8328, | |
| "eval_samples_per_second": 111.807, | |
| "eval_steps_per_second": 13.98, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 2.211025778995255, | |
| "grad_norm": 17.91064453125, | |
| "learning_rate": 1.4609001304536891e-05, | |
| "loss": 2.5144, | |
| "step": 22600 | |
| }, | |
| { | |
| "epoch": 2.2208090789023136, | |
| "grad_norm": 17.678396224975586, | |
| "learning_rate": 1.4427815625452964e-05, | |
| "loss": 2.5018, | |
| "step": 22700 | |
| }, | |
| { | |
| "epoch": 2.2305923788093724, | |
| "grad_norm": 17.510461807250977, | |
| "learning_rate": 1.4246629946369039e-05, | |
| "loss": 2.4228, | |
| "step": 22800 | |
| }, | |
| { | |
| "epoch": 2.240375678716431, | |
| "grad_norm": 24.923967361450195, | |
| "learning_rate": 1.4065444267285114e-05, | |
| "loss": 2.5249, | |
| "step": 22900 | |
| }, | |
| { | |
| "epoch": 2.2501589786234897, | |
| "grad_norm": 17.82384490966797, | |
| "learning_rate": 1.388425858820119e-05, | |
| "loss": 2.4282, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 2.2501589786234897, | |
| "eval_runtime": 182.0459, | |
| "eval_samples_per_second": 112.29, | |
| "eval_steps_per_second": 14.04, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 2.2599422785305485, | |
| "grad_norm": 16.13028335571289, | |
| "learning_rate": 1.3703072909117265e-05, | |
| "loss": 2.4472, | |
| "step": 23100 | |
| }, | |
| { | |
| "epoch": 2.269725578437607, | |
| "grad_norm": 15.137242317199707, | |
| "learning_rate": 1.352188723003334e-05, | |
| "loss": 2.5985, | |
| "step": 23200 | |
| }, | |
| { | |
| "epoch": 2.2795088783446658, | |
| "grad_norm": 16.187530517578125, | |
| "learning_rate": 1.3340701550949415e-05, | |
| "loss": 2.4862, | |
| "step": 23300 | |
| }, | |
| { | |
| "epoch": 2.289292178251724, | |
| "grad_norm": 18.84433937072754, | |
| "learning_rate": 1.3159515871865488e-05, | |
| "loss": 2.516, | |
| "step": 23400 | |
| }, | |
| { | |
| "epoch": 2.299075478158783, | |
| "grad_norm": 20.209121704101562, | |
| "learning_rate": 1.2978330192781563e-05, | |
| "loss": 2.5031, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 2.299075478158783, | |
| "eval_runtime": 181.9806, | |
| "eval_samples_per_second": 112.331, | |
| "eval_steps_per_second": 14.045, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 2.308858778065842, | |
| "grad_norm": 67.4502182006836, | |
| "learning_rate": 1.2797144513697637e-05, | |
| "loss": 2.4491, | |
| "step": 23600 | |
| }, | |
| { | |
| "epoch": 2.3186420779729002, | |
| "grad_norm": 14.940401077270508, | |
| "learning_rate": 1.2615958834613712e-05, | |
| "loss": 2.5669, | |
| "step": 23700 | |
| }, | |
| { | |
| "epoch": 2.328425377879959, | |
| "grad_norm": 16.591793060302734, | |
| "learning_rate": 1.2434773155529787e-05, | |
| "loss": 2.4565, | |
| "step": 23800 | |
| }, | |
| { | |
| "epoch": 2.3382086777870175, | |
| "grad_norm": 16.798791885375977, | |
| "learning_rate": 1.2253587476445862e-05, | |
| "loss": 2.4046, | |
| "step": 23900 | |
| }, | |
| { | |
| "epoch": 2.3479919776940763, | |
| "grad_norm": 17.712255477905273, | |
| "learning_rate": 1.2072401797361937e-05, | |
| "loss": 2.4453, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 2.3479919776940763, | |
| "eval_runtime": 182.0401, | |
| "eval_samples_per_second": 112.294, | |
| "eval_steps_per_second": 14.041, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 2.3577752776011347, | |
| "grad_norm": 18.64284324645996, | |
| "learning_rate": 1.1891216118278011e-05, | |
| "loss": 2.3973, | |
| "step": 24100 | |
| }, | |
| { | |
| "epoch": 2.3675585775081935, | |
| "grad_norm": 18.185895919799805, | |
| "learning_rate": 1.1710030439194086e-05, | |
| "loss": 2.5045, | |
| "step": 24200 | |
| }, | |
| { | |
| "epoch": 2.377341877415252, | |
| "grad_norm": 23.201522827148438, | |
| "learning_rate": 1.1528844760110163e-05, | |
| "loss": 2.5402, | |
| "step": 24300 | |
| }, | |
| { | |
| "epoch": 2.3871251773223108, | |
| "grad_norm": 21.606412887573242, | |
| "learning_rate": 1.1347659081026236e-05, | |
| "loss": 2.4285, | |
| "step": 24400 | |
| }, | |
| { | |
| "epoch": 2.3969084772293696, | |
| "grad_norm": 16.318761825561523, | |
| "learning_rate": 1.116647340194231e-05, | |
| "loss": 2.5509, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 2.3969084772293696, | |
| "eval_runtime": 182.0431, | |
| "eval_samples_per_second": 112.292, | |
| "eval_steps_per_second": 14.041, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 2.406691777136428, | |
| "grad_norm": 17.779014587402344, | |
| "learning_rate": 1.0985287722858386e-05, | |
| "loss": 2.4245, | |
| "step": 24600 | |
| }, | |
| { | |
| "epoch": 2.416475077043487, | |
| "grad_norm": 18.44321060180664, | |
| "learning_rate": 1.080410204377446e-05, | |
| "loss": 2.5223, | |
| "step": 24700 | |
| }, | |
| { | |
| "epoch": 2.4262583769505452, | |
| "grad_norm": 24.017047882080078, | |
| "learning_rate": 1.0622916364690535e-05, | |
| "loss": 2.4846, | |
| "step": 24800 | |
| }, | |
| { | |
| "epoch": 2.436041676857604, | |
| "grad_norm": 14.89560604095459, | |
| "learning_rate": 1.044173068560661e-05, | |
| "loss": 2.5922, | |
| "step": 24900 | |
| }, | |
| { | |
| "epoch": 2.445824976764663, | |
| "grad_norm": 15.532561302185059, | |
| "learning_rate": 1.0260545006522685e-05, | |
| "loss": 2.3976, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 2.445824976764663, | |
| "eval_runtime": 182.1033, | |
| "eval_samples_per_second": 112.255, | |
| "eval_steps_per_second": 14.036, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 2.4556082766717213, | |
| "grad_norm": 18.041282653808594, | |
| "learning_rate": 1.007935932743876e-05, | |
| "loss": 2.4731, | |
| "step": 25100 | |
| }, | |
| { | |
| "epoch": 2.46539157657878, | |
| "grad_norm": 13.40858268737793, | |
| "learning_rate": 9.898173648354834e-06, | |
| "loss": 2.4838, | |
| "step": 25200 | |
| }, | |
| { | |
| "epoch": 2.4751748764858386, | |
| "grad_norm": 17.450841903686523, | |
| "learning_rate": 9.71698796927091e-06, | |
| "loss": 2.3999, | |
| "step": 25300 | |
| }, | |
| { | |
| "epoch": 2.4849581763928974, | |
| "grad_norm": 17.556467056274414, | |
| "learning_rate": 9.535802290186984e-06, | |
| "loss": 2.3867, | |
| "step": 25400 | |
| }, | |
| { | |
| "epoch": 2.494741476299956, | |
| "grad_norm": 18.578310012817383, | |
| "learning_rate": 9.354616611103059e-06, | |
| "loss": 2.4546, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 2.494741476299956, | |
| "eval_runtime": 182.0338, | |
| "eval_samples_per_second": 112.298, | |
| "eval_steps_per_second": 14.041, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 2.5045247762070146, | |
| "grad_norm": 14.936469078063965, | |
| "learning_rate": 9.173430932019134e-06, | |
| "loss": 2.5562, | |
| "step": 25600 | |
| }, | |
| { | |
| "epoch": 2.514308076114073, | |
| "grad_norm": 17.527040481567383, | |
| "learning_rate": 8.992245252935209e-06, | |
| "loss": 2.4008, | |
| "step": 25700 | |
| }, | |
| { | |
| "epoch": 2.524091376021132, | |
| "grad_norm": 12.91336727142334, | |
| "learning_rate": 8.811059573851283e-06, | |
| "loss": 2.4655, | |
| "step": 25800 | |
| }, | |
| { | |
| "epoch": 2.5338746759281907, | |
| "grad_norm": 15.168461799621582, | |
| "learning_rate": 8.629873894767358e-06, | |
| "loss": 2.4468, | |
| "step": 25900 | |
| }, | |
| { | |
| "epoch": 2.543657975835249, | |
| "grad_norm": 17.5390682220459, | |
| "learning_rate": 8.448688215683433e-06, | |
| "loss": 2.4836, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 2.543657975835249, | |
| "eval_runtime": 182.1148, | |
| "eval_samples_per_second": 112.248, | |
| "eval_steps_per_second": 14.035, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 2.553441275742308, | |
| "grad_norm": 15.126510620117188, | |
| "learning_rate": 8.267502536599508e-06, | |
| "loss": 2.387, | |
| "step": 26100 | |
| }, | |
| { | |
| "epoch": 2.5632245756493663, | |
| "grad_norm": 15.374293327331543, | |
| "learning_rate": 8.086316857515583e-06, | |
| "loss": 2.3652, | |
| "step": 26200 | |
| }, | |
| { | |
| "epoch": 2.573007875556425, | |
| "grad_norm": 15.498108863830566, | |
| "learning_rate": 7.905131178431657e-06, | |
| "loss": 2.4749, | |
| "step": 26300 | |
| }, | |
| { | |
| "epoch": 2.582791175463484, | |
| "grad_norm": 16.221315383911133, | |
| "learning_rate": 7.723945499347732e-06, | |
| "loss": 2.4567, | |
| "step": 26400 | |
| }, | |
| { | |
| "epoch": 2.5925744753705424, | |
| "grad_norm": 18.839122772216797, | |
| "learning_rate": 7.542759820263806e-06, | |
| "loss": 2.3554, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 2.5925744753705424, | |
| "eval_runtime": 181.9597, | |
| "eval_samples_per_second": 112.344, | |
| "eval_steps_per_second": 14.047, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 2.6023577752776013, | |
| "grad_norm": 22.626708984375, | |
| "learning_rate": 7.361574141179882e-06, | |
| "loss": 2.502, | |
| "step": 26600 | |
| }, | |
| { | |
| "epoch": 2.6121410751846597, | |
| "grad_norm": 16.519880294799805, | |
| "learning_rate": 7.180388462095957e-06, | |
| "loss": 2.5034, | |
| "step": 26700 | |
| }, | |
| { | |
| "epoch": 2.6219243750917185, | |
| "grad_norm": 27.421489715576172, | |
| "learning_rate": 6.999202783012031e-06, | |
| "loss": 2.5276, | |
| "step": 26800 | |
| }, | |
| { | |
| "epoch": 2.6317076749987773, | |
| "grad_norm": 15.274630546569824, | |
| "learning_rate": 6.8180171039281055e-06, | |
| "loss": 2.4121, | |
| "step": 26900 | |
| }, | |
| { | |
| "epoch": 2.6414909749058357, | |
| "grad_norm": 15.751582145690918, | |
| "learning_rate": 6.636831424844181e-06, | |
| "loss": 2.5799, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 2.6414909749058357, | |
| "eval_runtime": 182.0873, | |
| "eval_samples_per_second": 112.265, | |
| "eval_steps_per_second": 14.037, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 2.651274274812894, | |
| "grad_norm": 16.674850463867188, | |
| "learning_rate": 6.455645745760255e-06, | |
| "loss": 2.3872, | |
| "step": 27100 | |
| }, | |
| { | |
| "epoch": 2.661057574719953, | |
| "grad_norm": 12.62803840637207, | |
| "learning_rate": 6.27446006667633e-06, | |
| "loss": 2.4, | |
| "step": 27200 | |
| }, | |
| { | |
| "epoch": 2.670840874627012, | |
| "grad_norm": 18.055158615112305, | |
| "learning_rate": 6.093274387592405e-06, | |
| "loss": 2.4681, | |
| "step": 27300 | |
| }, | |
| { | |
| "epoch": 2.68062417453407, | |
| "grad_norm": 17.21278190612793, | |
| "learning_rate": 5.91208870850848e-06, | |
| "loss": 2.5441, | |
| "step": 27400 | |
| }, | |
| { | |
| "epoch": 2.690407474441129, | |
| "grad_norm": 20.945236206054688, | |
| "learning_rate": 5.7309030294245544e-06, | |
| "loss": 2.4388, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 2.690407474441129, | |
| "eval_runtime": 182.1279, | |
| "eval_samples_per_second": 112.24, | |
| "eval_steps_per_second": 14.034, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 2.7001907743481874, | |
| "grad_norm": 23.483661651611328, | |
| "learning_rate": 5.549717350340629e-06, | |
| "loss": 2.4589, | |
| "step": 27600 | |
| }, | |
| { | |
| "epoch": 2.7099740742552463, | |
| "grad_norm": 17.954036712646484, | |
| "learning_rate": 5.368531671256704e-06, | |
| "loss": 2.4477, | |
| "step": 27700 | |
| }, | |
| { | |
| "epoch": 2.719757374162305, | |
| "grad_norm": 16.187314987182617, | |
| "learning_rate": 5.187345992172779e-06, | |
| "loss": 2.4967, | |
| "step": 27800 | |
| }, | |
| { | |
| "epoch": 2.7295406740693635, | |
| "grad_norm": 14.324910163879395, | |
| "learning_rate": 5.006160313088854e-06, | |
| "loss": 2.3921, | |
| "step": 27900 | |
| }, | |
| { | |
| "epoch": 2.7393239739764224, | |
| "grad_norm": 20.81557846069336, | |
| "learning_rate": 4.8249746340049285e-06, | |
| "loss": 2.5201, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 2.7393239739764224, | |
| "eval_runtime": 182.146, | |
| "eval_samples_per_second": 112.229, | |
| "eval_steps_per_second": 14.033, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 2.7491072738834808, | |
| "grad_norm": 18.682844161987305, | |
| "learning_rate": 4.643788954921003e-06, | |
| "loss": 2.4325, | |
| "step": 28100 | |
| }, | |
| { | |
| "epoch": 2.7588905737905396, | |
| "grad_norm": 16.227272033691406, | |
| "learning_rate": 4.462603275837078e-06, | |
| "loss": 2.3864, | |
| "step": 28200 | |
| }, | |
| { | |
| "epoch": 2.7686738736975984, | |
| "grad_norm": 16.20302963256836, | |
| "learning_rate": 4.281417596753152e-06, | |
| "loss": 2.5296, | |
| "step": 28300 | |
| }, | |
| { | |
| "epoch": 2.778457173604657, | |
| "grad_norm": 18.634096145629883, | |
| "learning_rate": 4.100231917669228e-06, | |
| "loss": 2.4514, | |
| "step": 28400 | |
| }, | |
| { | |
| "epoch": 2.7882404735117157, | |
| "grad_norm": 13.040008544921875, | |
| "learning_rate": 3.919046238585303e-06, | |
| "loss": 2.3661, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 2.7882404735117157, | |
| "eval_runtime": 181.9164, | |
| "eval_samples_per_second": 112.37, | |
| "eval_steps_per_second": 14.05, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 2.798023773418774, | |
| "grad_norm": 14.142943382263184, | |
| "learning_rate": 3.737860559501377e-06, | |
| "loss": 2.5074, | |
| "step": 28600 | |
| }, | |
| { | |
| "epoch": 2.807807073325833, | |
| "grad_norm": 17.934324264526367, | |
| "learning_rate": 3.5566748804174523e-06, | |
| "loss": 2.4224, | |
| "step": 28700 | |
| }, | |
| { | |
| "epoch": 2.8175903732328913, | |
| "grad_norm": 14.450194358825684, | |
| "learning_rate": 3.3754892013335267e-06, | |
| "loss": 2.4949, | |
| "step": 28800 | |
| }, | |
| { | |
| "epoch": 2.82737367313995, | |
| "grad_norm": 17.746837615966797, | |
| "learning_rate": 3.194303522249602e-06, | |
| "loss": 2.4153, | |
| "step": 28900 | |
| }, | |
| { | |
| "epoch": 2.8371569730470085, | |
| "grad_norm": 13.962541580200195, | |
| "learning_rate": 3.0131178431656763e-06, | |
| "loss": 2.4804, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 2.8371569730470085, | |
| "eval_runtime": 182.0262, | |
| "eval_samples_per_second": 112.303, | |
| "eval_steps_per_second": 14.042, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 2.8469402729540674, | |
| "grad_norm": 16.669286727905273, | |
| "learning_rate": 2.831932164081751e-06, | |
| "loss": 2.5397, | |
| "step": 29100 | |
| }, | |
| { | |
| "epoch": 2.856723572861126, | |
| "grad_norm": 15.421733856201172, | |
| "learning_rate": 2.650746484997826e-06, | |
| "loss": 2.4175, | |
| "step": 29200 | |
| }, | |
| { | |
| "epoch": 2.8665068727681846, | |
| "grad_norm": 14.135702133178711, | |
| "learning_rate": 2.4695608059139007e-06, | |
| "loss": 2.5069, | |
| "step": 29300 | |
| }, | |
| { | |
| "epoch": 2.8762901726752435, | |
| "grad_norm": 17.41412925720215, | |
| "learning_rate": 2.2883751268299756e-06, | |
| "loss": 2.3997, | |
| "step": 29400 | |
| }, | |
| { | |
| "epoch": 2.886073472582302, | |
| "grad_norm": 14.824533462524414, | |
| "learning_rate": 2.1071894477460504e-06, | |
| "loss": 2.3945, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 2.886073472582302, | |
| "eval_runtime": 181.9299, | |
| "eval_samples_per_second": 112.362, | |
| "eval_steps_per_second": 14.049, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 2.8958567724893607, | |
| "grad_norm": 27.31865119934082, | |
| "learning_rate": 1.926003768662125e-06, | |
| "loss": 2.45, | |
| "step": 29600 | |
| }, | |
| { | |
| "epoch": 2.9056400723964195, | |
| "grad_norm": 18.966655731201172, | |
| "learning_rate": 1.7448180895781998e-06, | |
| "loss": 2.3916, | |
| "step": 29700 | |
| }, | |
| { | |
| "epoch": 2.915423372303478, | |
| "grad_norm": 18.538440704345703, | |
| "learning_rate": 1.5636324104942746e-06, | |
| "loss": 2.4625, | |
| "step": 29800 | |
| }, | |
| { | |
| "epoch": 2.9252066722105368, | |
| "grad_norm": 21.757272720336914, | |
| "learning_rate": 1.3824467314103494e-06, | |
| "loss": 2.3722, | |
| "step": 29900 | |
| }, | |
| { | |
| "epoch": 2.934989972117595, | |
| "grad_norm": 16.907358169555664, | |
| "learning_rate": 1.201261052326424e-06, | |
| "loss": 2.464, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 2.934989972117595, | |
| "eval_runtime": 181.9148, | |
| "eval_samples_per_second": 112.371, | |
| "eval_steps_per_second": 14.051, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 2.944773272024654, | |
| "grad_norm": 13.88399600982666, | |
| "learning_rate": 1.0200753732424989e-06, | |
| "loss": 2.5005, | |
| "step": 30100 | |
| }, | |
| { | |
| "epoch": 2.954556571931713, | |
| "grad_norm": 19.77507781982422, | |
| "learning_rate": 8.388896941585737e-07, | |
| "loss": 2.3829, | |
| "step": 30200 | |
| }, | |
| { | |
| "epoch": 2.9643398718387712, | |
| "grad_norm": 16.535932540893555, | |
| "learning_rate": 6.577040150746485e-07, | |
| "loss": 2.4788, | |
| "step": 30300 | |
| }, | |
| { | |
| "epoch": 2.9741231717458296, | |
| "grad_norm": 15.027000427246094, | |
| "learning_rate": 4.765183359907233e-07, | |
| "loss": 2.5007, | |
| "step": 30400 | |
| }, | |
| { | |
| "epoch": 2.9839064716528885, | |
| "grad_norm": 14.9392671585083, | |
| "learning_rate": 2.953326569067981e-07, | |
| "loss": 2.4847, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 2.9839064716528885, | |
| "eval_runtime": 181.9853, | |
| "eval_samples_per_second": 112.328, | |
| "eval_steps_per_second": 14.045, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 2.9936897715599473, | |
| "grad_norm": 15.128337860107422, | |
| "learning_rate": 1.1414697782287289e-07, | |
| "loss": 2.4209, | |
| "step": 30600 | |
| } | |
| ], | |
| "logging_steps": 100, | |
| "max_steps": 30663, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.0701267610290972e+16, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |