| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 9.0, |
| "eval_steps": 500, |
| "global_step": 2214, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.04065040650406504, |
| "grad_norm": 14.707608222961426, |
| "learning_rate": 9.000000000000001e-07, |
| "loss": 0.9613, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.08130081300813008, |
| "grad_norm": 12.035386085510254, |
| "learning_rate": 1.9000000000000002e-06, |
| "loss": 0.8509, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.12195121951219512, |
| "grad_norm": 18.44657325744629, |
| "learning_rate": 2.9e-06, |
| "loss": 0.9369, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.16260162601626016, |
| "grad_norm": 12.941998481750488, |
| "learning_rate": 3.900000000000001e-06, |
| "loss": 0.896, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.2032520325203252, |
| "grad_norm": 15.150432586669922, |
| "learning_rate": 4.9000000000000005e-06, |
| "loss": 0.9821, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.24390243902439024, |
| "grad_norm": 18.638032913208008, |
| "learning_rate": 5.9e-06, |
| "loss": 0.9666, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.2845528455284553, |
| "grad_norm": 15.954939842224121, |
| "learning_rate": 6.9e-06, |
| "loss": 0.9019, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.3252032520325203, |
| "grad_norm": 11.114273071289062, |
| "learning_rate": 7.9e-06, |
| "loss": 0.8321, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.36585365853658536, |
| "grad_norm": 15.285880088806152, |
| "learning_rate": 8.900000000000001e-06, |
| "loss": 0.7198, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.4065040650406504, |
| "grad_norm": 17.8586483001709, |
| "learning_rate": 9.9e-06, |
| "loss": 0.7806, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.44715447154471544, |
| "grad_norm": 11.217978477478027, |
| "learning_rate": 9.961864406779662e-06, |
| "loss": 1.0024, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.4878048780487805, |
| "grad_norm": 7.002477169036865, |
| "learning_rate": 9.91949152542373e-06, |
| "loss": 0.8086, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.5284552845528455, |
| "grad_norm": 13.211562156677246, |
| "learning_rate": 9.877118644067798e-06, |
| "loss": 0.8958, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.5691056910569106, |
| "grad_norm": 11.169763565063477, |
| "learning_rate": 9.834745762711865e-06, |
| "loss": 0.8316, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.6097560975609756, |
| "grad_norm": 15.369425773620605, |
| "learning_rate": 9.792372881355933e-06, |
| "loss": 0.675, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.6504065040650406, |
| "grad_norm": 11.003032684326172, |
| "learning_rate": 9.75e-06, |
| "loss": 0.7512, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.6910569105691057, |
| "grad_norm": 16.466957092285156, |
| "learning_rate": 9.707627118644068e-06, |
| "loss": 0.8546, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.7317073170731707, |
| "grad_norm": 13.067832946777344, |
| "learning_rate": 9.665254237288136e-06, |
| "loss": 0.7497, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.7723577235772358, |
| "grad_norm": 9.530882835388184, |
| "learning_rate": 9.622881355932205e-06, |
| "loss": 0.6136, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.8130081300813008, |
| "grad_norm": 15.063409805297852, |
| "learning_rate": 9.580508474576273e-06, |
| "loss": 0.6073, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.8536585365853658, |
| "grad_norm": 9.07528305053711, |
| "learning_rate": 9.53813559322034e-06, |
| "loss": 0.7295, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.8943089430894309, |
| "grad_norm": 14.199564933776855, |
| "learning_rate": 9.495762711864408e-06, |
| "loss": 0.596, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.9349593495934959, |
| "grad_norm": 16.03732681274414, |
| "learning_rate": 9.453389830508474e-06, |
| "loss": 0.6922, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.975609756097561, |
| "grad_norm": 17.533695220947266, |
| "learning_rate": 9.411016949152543e-06, |
| "loss": 0.7484, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_general_loss": 0.754059910774231, |
| "eval_general_runtime": 254.3068, |
| "eval_general_samples_per_second": 3.598, |
| "eval_general_steps_per_second": 0.9, |
| "step": 246 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_code_loss": 0.7234187722206116, |
| "eval_code_runtime": 296.8568, |
| "eval_code_samples_per_second": 3.092, |
| "eval_code_steps_per_second": 0.775, |
| "step": 246 |
| }, |
| { |
| "epoch": 1.0, |
| "eval_stem_loss": 0.7939386367797852, |
| "eval_stem_runtime": 250.6578, |
| "eval_stem_samples_per_second": 3.646, |
| "eval_stem_steps_per_second": 0.914, |
| "step": 246 |
| }, |
| { |
| "epoch": 1.016260162601626, |
| "grad_norm": 16.589147567749023, |
| "learning_rate": 9.368644067796611e-06, |
| "loss": 0.7309, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.056910569105691, |
| "grad_norm": 8.124390602111816, |
| "learning_rate": 9.32627118644068e-06, |
| "loss": 0.5937, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.0975609756097562, |
| "grad_norm": 13.223992347717285, |
| "learning_rate": 9.283898305084746e-06, |
| "loss": 0.6961, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.1382113821138211, |
| "grad_norm": 13.791741371154785, |
| "learning_rate": 9.241525423728814e-06, |
| "loss": 0.5783, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.1788617886178863, |
| "grad_norm": 16.130605697631836, |
| "learning_rate": 9.199152542372882e-06, |
| "loss": 0.6318, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.2195121951219512, |
| "grad_norm": 14.941628456115723, |
| "learning_rate": 9.15677966101695e-06, |
| "loss": 0.6163, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.2601626016260163, |
| "grad_norm": 6.407144546508789, |
| "learning_rate": 9.114406779661017e-06, |
| "loss": 0.5745, |
| "step": 310 |
| }, |
| { |
| "epoch": 1.3008130081300813, |
| "grad_norm": 7.032970905303955, |
| "learning_rate": 9.072033898305086e-06, |
| "loss": 0.474, |
| "step": 320 |
| }, |
| { |
| "epoch": 1.3414634146341464, |
| "grad_norm": 14.255295753479004, |
| "learning_rate": 9.029661016949152e-06, |
| "loss": 0.6833, |
| "step": 330 |
| }, |
| { |
| "epoch": 1.3821138211382114, |
| "grad_norm": 19.13692855834961, |
| "learning_rate": 8.98728813559322e-06, |
| "loss": 0.5369, |
| "step": 340 |
| }, |
| { |
| "epoch": 1.4227642276422765, |
| "grad_norm": 9.524161338806152, |
| "learning_rate": 8.944915254237289e-06, |
| "loss": 0.6925, |
| "step": 350 |
| }, |
| { |
| "epoch": 1.4634146341463414, |
| "grad_norm": 12.59819221496582, |
| "learning_rate": 8.902542372881357e-06, |
| "loss": 0.572, |
| "step": 360 |
| }, |
| { |
| "epoch": 1.5040650406504064, |
| "grad_norm": 7.560878753662109, |
| "learning_rate": 8.860169491525424e-06, |
| "loss": 0.5101, |
| "step": 370 |
| }, |
| { |
| "epoch": 1.5447154471544715, |
| "grad_norm": 10.911989212036133, |
| "learning_rate": 8.817796610169492e-06, |
| "loss": 0.7357, |
| "step": 380 |
| }, |
| { |
| "epoch": 1.5853658536585367, |
| "grad_norm": 13.019627571105957, |
| "learning_rate": 8.77542372881356e-06, |
| "loss": 0.7094, |
| "step": 390 |
| }, |
| { |
| "epoch": 1.6260162601626016, |
| "grad_norm": 5.221607685089111, |
| "learning_rate": 8.733050847457629e-06, |
| "loss": 0.5762, |
| "step": 400 |
| }, |
| { |
| "epoch": 1.6666666666666665, |
| "grad_norm": 9.154648780822754, |
| "learning_rate": 8.690677966101695e-06, |
| "loss": 0.6516, |
| "step": 410 |
| }, |
| { |
| "epoch": 1.7073170731707317, |
| "grad_norm": 8.04185962677002, |
| "learning_rate": 8.648305084745763e-06, |
| "loss": 0.5457, |
| "step": 420 |
| }, |
| { |
| "epoch": 1.7479674796747968, |
| "grad_norm": 12.428257942199707, |
| "learning_rate": 8.60593220338983e-06, |
| "loss": 0.6458, |
| "step": 430 |
| }, |
| { |
| "epoch": 1.7886178861788617, |
| "grad_norm": 6.699453353881836, |
| "learning_rate": 8.563559322033898e-06, |
| "loss": 0.6026, |
| "step": 440 |
| }, |
| { |
| "epoch": 1.8292682926829267, |
| "grad_norm": 5.6174774169921875, |
| "learning_rate": 8.521186440677967e-06, |
| "loss": 0.6756, |
| "step": 450 |
| }, |
| { |
| "epoch": 1.8699186991869918, |
| "grad_norm": 12.72862720489502, |
| "learning_rate": 8.478813559322035e-06, |
| "loss": 0.5932, |
| "step": 460 |
| }, |
| { |
| "epoch": 1.910569105691057, |
| "grad_norm": 4.651763439178467, |
| "learning_rate": 8.436440677966102e-06, |
| "loss": 0.418, |
| "step": 470 |
| }, |
| { |
| "epoch": 1.951219512195122, |
| "grad_norm": 16.044872283935547, |
| "learning_rate": 8.39406779661017e-06, |
| "loss": 0.5756, |
| "step": 480 |
| }, |
| { |
| "epoch": 1.9918699186991868, |
| "grad_norm": 10.388751029968262, |
| "learning_rate": 8.351694915254238e-06, |
| "loss": 0.6304, |
| "step": 490 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_general_loss": 0.6887519359588623, |
| "eval_general_runtime": 255.0576, |
| "eval_general_samples_per_second": 3.587, |
| "eval_general_steps_per_second": 0.898, |
| "step": 492 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_code_loss": 0.5912412405014038, |
| "eval_code_runtime": 297.8316, |
| "eval_code_samples_per_second": 3.082, |
| "eval_code_steps_per_second": 0.772, |
| "step": 492 |
| }, |
| { |
| "epoch": 2.0, |
| "eval_stem_loss": 0.7283160090446472, |
| "eval_stem_runtime": 251.4929, |
| "eval_stem_samples_per_second": 3.634, |
| "eval_stem_steps_per_second": 0.911, |
| "step": 492 |
| }, |
| { |
| "epoch": 2.032520325203252, |
| "grad_norm": 13.620870590209961, |
| "learning_rate": 8.309322033898306e-06, |
| "loss": 0.5292, |
| "step": 500 |
| }, |
| { |
| "epoch": 2.073170731707317, |
| "grad_norm": 11.176264762878418, |
| "learning_rate": 8.266949152542375e-06, |
| "loss": 0.4926, |
| "step": 510 |
| }, |
| { |
| "epoch": 2.113821138211382, |
| "grad_norm": 4.5993852615356445, |
| "learning_rate": 8.224576271186441e-06, |
| "loss": 0.476, |
| "step": 520 |
| }, |
| { |
| "epoch": 2.154471544715447, |
| "grad_norm": 16.03154182434082, |
| "learning_rate": 8.182203389830508e-06, |
| "loss": 0.492, |
| "step": 530 |
| }, |
| { |
| "epoch": 2.1951219512195124, |
| "grad_norm": 3.7860374450683594, |
| "learning_rate": 8.139830508474576e-06, |
| "loss": 0.4376, |
| "step": 540 |
| }, |
| { |
| "epoch": 2.2357723577235773, |
| "grad_norm": 14.141989707946777, |
| "learning_rate": 8.097457627118644e-06, |
| "loss": 0.5422, |
| "step": 550 |
| }, |
| { |
| "epoch": 2.2764227642276422, |
| "grad_norm": 11.716277122497559, |
| "learning_rate": 8.055084745762713e-06, |
| "loss": 0.4781, |
| "step": 560 |
| }, |
| { |
| "epoch": 2.317073170731707, |
| "grad_norm": 16.112192153930664, |
| "learning_rate": 8.012711864406781e-06, |
| "loss": 0.5243, |
| "step": 570 |
| }, |
| { |
| "epoch": 2.3577235772357725, |
| "grad_norm": 5.857080936431885, |
| "learning_rate": 7.970338983050848e-06, |
| "loss": 0.4588, |
| "step": 580 |
| }, |
| { |
| "epoch": 2.3983739837398375, |
| "grad_norm": 9.052570343017578, |
| "learning_rate": 7.927966101694916e-06, |
| "loss": 0.5717, |
| "step": 590 |
| }, |
| { |
| "epoch": 2.4390243902439024, |
| "grad_norm": 17.8330078125, |
| "learning_rate": 7.885593220338984e-06, |
| "loss": 0.4547, |
| "step": 600 |
| }, |
| { |
| "epoch": 2.4796747967479673, |
| "grad_norm": 13.779817581176758, |
| "learning_rate": 7.843220338983052e-06, |
| "loss": 0.4584, |
| "step": 610 |
| }, |
| { |
| "epoch": 2.5203252032520327, |
| "grad_norm": 18.924348831176758, |
| "learning_rate": 7.800847457627119e-06, |
| "loss": 0.4788, |
| "step": 620 |
| }, |
| { |
| "epoch": 2.5609756097560976, |
| "grad_norm": 12.157154083251953, |
| "learning_rate": 7.758474576271187e-06, |
| "loss": 0.4762, |
| "step": 630 |
| }, |
| { |
| "epoch": 2.6016260162601625, |
| "grad_norm": 14.143574714660645, |
| "learning_rate": 7.716101694915254e-06, |
| "loss": 0.7017, |
| "step": 640 |
| }, |
| { |
| "epoch": 2.642276422764228, |
| "grad_norm": 11.326656341552734, |
| "learning_rate": 7.673728813559322e-06, |
| "loss": 0.4524, |
| "step": 650 |
| }, |
| { |
| "epoch": 2.682926829268293, |
| "grad_norm": 5.236691474914551, |
| "learning_rate": 7.63135593220339e-06, |
| "loss": 0.5228, |
| "step": 660 |
| }, |
| { |
| "epoch": 2.7235772357723578, |
| "grad_norm": 17.48626708984375, |
| "learning_rate": 7.588983050847458e-06, |
| "loss": 0.4414, |
| "step": 670 |
| }, |
| { |
| "epoch": 2.7642276422764227, |
| "grad_norm": 13.257789611816406, |
| "learning_rate": 7.546610169491526e-06, |
| "loss": 0.5442, |
| "step": 680 |
| }, |
| { |
| "epoch": 2.8048780487804876, |
| "grad_norm": 20.655475616455078, |
| "learning_rate": 7.504237288135594e-06, |
| "loss": 0.5451, |
| "step": 690 |
| }, |
| { |
| "epoch": 2.845528455284553, |
| "grad_norm": 10.467672348022461, |
| "learning_rate": 7.461864406779662e-06, |
| "loss": 0.4601, |
| "step": 700 |
| }, |
| { |
| "epoch": 2.886178861788618, |
| "grad_norm": 13.194644927978516, |
| "learning_rate": 7.41949152542373e-06, |
| "loss": 0.502, |
| "step": 710 |
| }, |
| { |
| "epoch": 2.926829268292683, |
| "grad_norm": 9.131794929504395, |
| "learning_rate": 7.377118644067798e-06, |
| "loss": 0.5583, |
| "step": 720 |
| }, |
| { |
| "epoch": 2.9674796747967482, |
| "grad_norm": 8.837148666381836, |
| "learning_rate": 7.334745762711864e-06, |
| "loss": 0.4668, |
| "step": 730 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_general_loss": 0.6758893728256226, |
| "eval_general_runtime": 255.2921, |
| "eval_general_samples_per_second": 3.584, |
| "eval_general_steps_per_second": 0.897, |
| "step": 738 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_code_loss": 0.5598136782646179, |
| "eval_code_runtime": 297.7838, |
| "eval_code_samples_per_second": 3.083, |
| "eval_code_steps_per_second": 0.772, |
| "step": 738 |
| }, |
| { |
| "epoch": 3.0, |
| "eval_stem_loss": 0.7172386050224304, |
| "eval_stem_runtime": 251.2967, |
| "eval_stem_samples_per_second": 3.637, |
| "eval_stem_steps_per_second": 0.911, |
| "step": 738 |
| }, |
| { |
| "epoch": 3.008130081300813, |
| "grad_norm": 8.791101455688477, |
| "learning_rate": 7.292372881355933e-06, |
| "loss": 0.3545, |
| "step": 740 |
| }, |
| { |
| "epoch": 3.048780487804878, |
| "grad_norm": 7.602899074554443, |
| "learning_rate": 7.25e-06, |
| "loss": 0.4633, |
| "step": 750 |
| }, |
| { |
| "epoch": 3.089430894308943, |
| "grad_norm": 10.939055442810059, |
| "learning_rate": 7.207627118644068e-06, |
| "loss": 0.3989, |
| "step": 760 |
| }, |
| { |
| "epoch": 3.130081300813008, |
| "grad_norm": 9.168066024780273, |
| "learning_rate": 7.165254237288136e-06, |
| "loss": 0.5555, |
| "step": 770 |
| }, |
| { |
| "epoch": 3.1707317073170733, |
| "grad_norm": 17.25099754333496, |
| "learning_rate": 7.122881355932204e-06, |
| "loss": 0.5326, |
| "step": 780 |
| }, |
| { |
| "epoch": 3.2113821138211383, |
| "grad_norm": 13.0396146774292, |
| "learning_rate": 7.080508474576272e-06, |
| "loss": 0.3831, |
| "step": 790 |
| }, |
| { |
| "epoch": 3.252032520325203, |
| "grad_norm": 11.217447280883789, |
| "learning_rate": 7.03813559322034e-06, |
| "loss": 0.4142, |
| "step": 800 |
| }, |
| { |
| "epoch": 3.292682926829268, |
| "grad_norm": 8.102045059204102, |
| "learning_rate": 6.995762711864408e-06, |
| "loss": 0.4414, |
| "step": 810 |
| }, |
| { |
| "epoch": 3.3333333333333335, |
| "grad_norm": 11.761209487915039, |
| "learning_rate": 6.953389830508475e-06, |
| "loss": 0.501, |
| "step": 820 |
| }, |
| { |
| "epoch": 3.3739837398373984, |
| "grad_norm": 7.058698654174805, |
| "learning_rate": 6.911016949152542e-06, |
| "loss": 0.3771, |
| "step": 830 |
| }, |
| { |
| "epoch": 3.4146341463414633, |
| "grad_norm": 7.728908061981201, |
| "learning_rate": 6.86864406779661e-06, |
| "loss": 0.3867, |
| "step": 840 |
| }, |
| { |
| "epoch": 3.4552845528455283, |
| "grad_norm": 12.833827018737793, |
| "learning_rate": 6.826271186440679e-06, |
| "loss": 0.4132, |
| "step": 850 |
| }, |
| { |
| "epoch": 3.4959349593495936, |
| "grad_norm": 9.884450912475586, |
| "learning_rate": 6.783898305084746e-06, |
| "loss": 0.3674, |
| "step": 860 |
| }, |
| { |
| "epoch": 3.5365853658536586, |
| "grad_norm": 10.731056213378906, |
| "learning_rate": 6.741525423728814e-06, |
| "loss": 0.422, |
| "step": 870 |
| }, |
| { |
| "epoch": 3.5772357723577235, |
| "grad_norm": 6.745728015899658, |
| "learning_rate": 6.699152542372882e-06, |
| "loss": 0.499, |
| "step": 880 |
| }, |
| { |
| "epoch": 3.617886178861789, |
| "grad_norm": 9.808424949645996, |
| "learning_rate": 6.65677966101695e-06, |
| "loss": 0.2978, |
| "step": 890 |
| }, |
| { |
| "epoch": 3.658536585365854, |
| "grad_norm": 8.444202423095703, |
| "learning_rate": 6.614406779661018e-06, |
| "loss": 0.3698, |
| "step": 900 |
| }, |
| { |
| "epoch": 3.6991869918699187, |
| "grad_norm": 10.285717964172363, |
| "learning_rate": 6.572033898305086e-06, |
| "loss": 0.4677, |
| "step": 910 |
| }, |
| { |
| "epoch": 3.7398373983739837, |
| "grad_norm": 6.763285160064697, |
| "learning_rate": 6.5296610169491525e-06, |
| "loss": 0.4449, |
| "step": 920 |
| }, |
| { |
| "epoch": 3.7804878048780486, |
| "grad_norm": 20.608469009399414, |
| "learning_rate": 6.487288135593221e-06, |
| "loss": 0.5297, |
| "step": 930 |
| }, |
| { |
| "epoch": 3.821138211382114, |
| "grad_norm": 9.06677532196045, |
| "learning_rate": 6.444915254237288e-06, |
| "loss": 0.4292, |
| "step": 940 |
| }, |
| { |
| "epoch": 3.861788617886179, |
| "grad_norm": 6.93435525894165, |
| "learning_rate": 6.4025423728813565e-06, |
| "loss": 0.3877, |
| "step": 950 |
| }, |
| { |
| "epoch": 3.902439024390244, |
| "grad_norm": 11.75598430633545, |
| "learning_rate": 6.360169491525424e-06, |
| "loss": 0.4032, |
| "step": 960 |
| }, |
| { |
| "epoch": 3.943089430894309, |
| "grad_norm": 6.338623046875, |
| "learning_rate": 6.317796610169492e-06, |
| "loss": 0.4228, |
| "step": 970 |
| }, |
| { |
| "epoch": 3.983739837398374, |
| "grad_norm": 6.302221775054932, |
| "learning_rate": 6.27542372881356e-06, |
| "loss": 0.4054, |
| "step": 980 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_general_loss": 0.6872960329055786, |
| "eval_general_runtime": 254.991, |
| "eval_general_samples_per_second": 3.588, |
| "eval_general_steps_per_second": 0.898, |
| "step": 984 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_code_loss": 0.5443978309631348, |
| "eval_code_runtime": 297.6318, |
| "eval_code_samples_per_second": 3.084, |
| "eval_code_steps_per_second": 0.773, |
| "step": 984 |
| }, |
| { |
| "epoch": 4.0, |
| "eval_stem_loss": 0.7525980472564697, |
| "eval_stem_runtime": 251.3515, |
| "eval_stem_samples_per_second": 3.636, |
| "eval_stem_steps_per_second": 0.911, |
| "step": 984 |
| }, |
| { |
| "epoch": 4.024390243902439, |
| "grad_norm": 11.085492134094238, |
| "learning_rate": 6.233050847457628e-06, |
| "loss": 0.2924, |
| "step": 990 |
| }, |
| { |
| "epoch": 4.065040650406504, |
| "grad_norm": 9.018081665039062, |
| "learning_rate": 6.190677966101695e-06, |
| "loss": 0.3824, |
| "step": 1000 |
| }, |
| { |
| "epoch": 4.105691056910569, |
| "grad_norm": 5.635726451873779, |
| "learning_rate": 6.148305084745764e-06, |
| "loss": 0.3209, |
| "step": 1010 |
| }, |
| { |
| "epoch": 4.146341463414634, |
| "grad_norm": 17.105005264282227, |
| "learning_rate": 6.10593220338983e-06, |
| "loss": 0.3092, |
| "step": 1020 |
| }, |
| { |
| "epoch": 4.186991869918699, |
| "grad_norm": 9.317548751831055, |
| "learning_rate": 6.0635593220338986e-06, |
| "loss": 0.3398, |
| "step": 1030 |
| }, |
| { |
| "epoch": 4.227642276422764, |
| "grad_norm": 12.000333786010742, |
| "learning_rate": 6.021186440677966e-06, |
| "loss": 0.3214, |
| "step": 1040 |
| }, |
| { |
| "epoch": 4.2682926829268295, |
| "grad_norm": 18.374374389648438, |
| "learning_rate": 5.978813559322034e-06, |
| "loss": 0.4008, |
| "step": 1050 |
| }, |
| { |
| "epoch": 4.308943089430894, |
| "grad_norm": 8.767024993896484, |
| "learning_rate": 5.936440677966102e-06, |
| "loss": 0.4476, |
| "step": 1060 |
| }, |
| { |
| "epoch": 4.349593495934959, |
| "grad_norm": 12.842947006225586, |
| "learning_rate": 5.89406779661017e-06, |
| "loss": 0.3627, |
| "step": 1070 |
| }, |
| { |
| "epoch": 4.390243902439025, |
| "grad_norm": 11.248525619506836, |
| "learning_rate": 5.851694915254238e-06, |
| "loss": 0.3427, |
| "step": 1080 |
| }, |
| { |
| "epoch": 4.430894308943089, |
| "grad_norm": 18.474157333374023, |
| "learning_rate": 5.809322033898306e-06, |
| "loss": 0.2842, |
| "step": 1090 |
| }, |
| { |
| "epoch": 4.471544715447155, |
| "grad_norm": 10.508559226989746, |
| "learning_rate": 5.766949152542374e-06, |
| "loss": 0.4153, |
| "step": 1100 |
| }, |
| { |
| "epoch": 4.512195121951219, |
| "grad_norm": 20.686229705810547, |
| "learning_rate": 5.7245762711864415e-06, |
| "loss": 0.3661, |
| "step": 1110 |
| }, |
| { |
| "epoch": 4.5528455284552845, |
| "grad_norm": 7.763411998748779, |
| "learning_rate": 5.682203389830508e-06, |
| "loss": 0.3325, |
| "step": 1120 |
| }, |
| { |
| "epoch": 4.59349593495935, |
| "grad_norm": 6.85905647277832, |
| "learning_rate": 5.639830508474576e-06, |
| "loss": 0.4229, |
| "step": 1130 |
| }, |
| { |
| "epoch": 4.634146341463414, |
| "grad_norm": 17.94886589050293, |
| "learning_rate": 5.597457627118645e-06, |
| "loss": 0.3009, |
| "step": 1140 |
| }, |
| { |
| "epoch": 4.67479674796748, |
| "grad_norm": 20.208101272583008, |
| "learning_rate": 5.555084745762712e-06, |
| "loss": 0.3604, |
| "step": 1150 |
| }, |
| { |
| "epoch": 4.715447154471545, |
| "grad_norm": 8.471419334411621, |
| "learning_rate": 5.51271186440678e-06, |
| "loss": 0.3724, |
| "step": 1160 |
| }, |
| { |
| "epoch": 4.7560975609756095, |
| "grad_norm": 9.207096099853516, |
| "learning_rate": 5.470338983050848e-06, |
| "loss": 0.3979, |
| "step": 1170 |
| }, |
| { |
| "epoch": 4.796747967479675, |
| "grad_norm": 18.151241302490234, |
| "learning_rate": 5.427966101694916e-06, |
| "loss": 0.3412, |
| "step": 1180 |
| }, |
| { |
| "epoch": 4.83739837398374, |
| "grad_norm": 11.33778190612793, |
| "learning_rate": 5.3855932203389836e-06, |
| "loss": 0.3582, |
| "step": 1190 |
| }, |
| { |
| "epoch": 4.878048780487805, |
| "grad_norm": 3.6720833778381348, |
| "learning_rate": 5.343220338983052e-06, |
| "loss": 0.3269, |
| "step": 1200 |
| }, |
| { |
| "epoch": 4.91869918699187, |
| "grad_norm": 15.8824462890625, |
| "learning_rate": 5.300847457627119e-06, |
| "loss": 0.4141, |
| "step": 1210 |
| }, |
| { |
| "epoch": 4.959349593495935, |
| "grad_norm": 8.455586433410645, |
| "learning_rate": 5.258474576271187e-06, |
| "loss": 0.3515, |
| "step": 1220 |
| }, |
| { |
| "epoch": 5.0, |
| "grad_norm": 13.756336212158203, |
| "learning_rate": 5.216101694915254e-06, |
| "loss": 0.3621, |
| "step": 1230 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_general_loss": 0.7001935839653015, |
| "eval_general_runtime": 255.6611, |
| "eval_general_samples_per_second": 3.579, |
| "eval_general_steps_per_second": 0.896, |
| "step": 1230 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_code_loss": 0.5433691740036011, |
| "eval_code_runtime": 297.6204, |
| "eval_code_samples_per_second": 3.084, |
| "eval_code_steps_per_second": 0.773, |
| "step": 1230 |
| }, |
| { |
| "epoch": 5.0, |
| "eval_stem_loss": 0.8110136389732361, |
| "eval_stem_runtime": 251.3377, |
| "eval_stem_samples_per_second": 3.637, |
| "eval_stem_steps_per_second": 0.911, |
| "step": 1230 |
| }, |
| { |
| "epoch": 5.040650406504065, |
| "grad_norm": 8.500140190124512, |
| "learning_rate": 5.1737288135593225e-06, |
| "loss": 0.2453, |
| "step": 1240 |
| }, |
| { |
| "epoch": 5.08130081300813, |
| "grad_norm": 23.405906677246094, |
| "learning_rate": 5.13135593220339e-06, |
| "loss": 0.2729, |
| "step": 1250 |
| }, |
| { |
| "epoch": 5.121951219512195, |
| "grad_norm": 15.196890830993652, |
| "learning_rate": 5.088983050847458e-06, |
| "loss": 0.2626, |
| "step": 1260 |
| }, |
| { |
| "epoch": 5.16260162601626, |
| "grad_norm": 12.534261703491211, |
| "learning_rate": 5.046610169491526e-06, |
| "loss": 0.2891, |
| "step": 1270 |
| }, |
| { |
| "epoch": 5.203252032520325, |
| "grad_norm": 16.788536071777344, |
| "learning_rate": 5.004237288135594e-06, |
| "loss": 0.3668, |
| "step": 1280 |
| }, |
| { |
| "epoch": 5.2439024390243905, |
| "grad_norm": 7.071691513061523, |
| "learning_rate": 4.961864406779661e-06, |
| "loss": 0.2995, |
| "step": 1290 |
| }, |
| { |
| "epoch": 5.284552845528455, |
| "grad_norm": 23.835960388183594, |
| "learning_rate": 4.919491525423729e-06, |
| "loss": 0.4169, |
| "step": 1300 |
| }, |
| { |
| "epoch": 5.32520325203252, |
| "grad_norm": 6.12783145904541, |
| "learning_rate": 4.877118644067797e-06, |
| "loss": 0.3055, |
| "step": 1310 |
| }, |
| { |
| "epoch": 5.365853658536586, |
| "grad_norm": 11.301496505737305, |
| "learning_rate": 4.8347457627118645e-06, |
| "loss": 0.1926, |
| "step": 1320 |
| }, |
| { |
| "epoch": 5.40650406504065, |
| "grad_norm": 11.58179759979248, |
| "learning_rate": 4.792372881355933e-06, |
| "loss": 0.3254, |
| "step": 1330 |
| }, |
| { |
| "epoch": 5.4471544715447155, |
| "grad_norm": 18.70044708251953, |
| "learning_rate": 4.75e-06, |
| "loss": 0.3081, |
| "step": 1340 |
| }, |
| { |
| "epoch": 5.487804878048781, |
| "grad_norm": 22.241409301757812, |
| "learning_rate": 4.707627118644068e-06, |
| "loss": 0.2969, |
| "step": 1350 |
| }, |
| { |
| "epoch": 5.528455284552845, |
| "grad_norm": 17.672147750854492, |
| "learning_rate": 4.665254237288136e-06, |
| "loss": 0.3348, |
| "step": 1360 |
| }, |
| { |
| "epoch": 5.569105691056911, |
| "grad_norm": 9.32618522644043, |
| "learning_rate": 4.622881355932204e-06, |
| "loss": 0.2523, |
| "step": 1370 |
| }, |
| { |
| "epoch": 5.609756097560975, |
| "grad_norm": 6.630079746246338, |
| "learning_rate": 4.580508474576272e-06, |
| "loss": 0.2889, |
| "step": 1380 |
| }, |
| { |
| "epoch": 5.650406504065041, |
| "grad_norm": 21.26811408996582, |
| "learning_rate": 4.538135593220339e-06, |
| "loss": 0.3947, |
| "step": 1390 |
| }, |
| { |
| "epoch": 5.691056910569106, |
| "grad_norm": 20.458436965942383, |
| "learning_rate": 4.4957627118644075e-06, |
| "loss": 0.3509, |
| "step": 1400 |
| }, |
| { |
| "epoch": 5.7317073170731705, |
| "grad_norm": 8.210813522338867, |
| "learning_rate": 4.453389830508475e-06, |
| "loss": 0.2976, |
| "step": 1410 |
| }, |
| { |
| "epoch": 5.772357723577236, |
| "grad_norm": 16.646589279174805, |
| "learning_rate": 4.411016949152543e-06, |
| "loss": 0.2757, |
| "step": 1420 |
| }, |
| { |
| "epoch": 5.8130081300813, |
| "grad_norm": 14.104079246520996, |
| "learning_rate": 4.368644067796611e-06, |
| "loss": 0.3213, |
| "step": 1430 |
| }, |
| { |
| "epoch": 5.853658536585366, |
| "grad_norm": 10.447735786437988, |
| "learning_rate": 4.326271186440678e-06, |
| "loss": 0.2845, |
| "step": 1440 |
| }, |
| { |
| "epoch": 5.894308943089431, |
| "grad_norm": 22.94976043701172, |
| "learning_rate": 4.283898305084746e-06, |
| "loss": 0.3306, |
| "step": 1450 |
| }, |
| { |
| "epoch": 5.934959349593496, |
| "grad_norm": 8.621337890625, |
| "learning_rate": 4.241525423728814e-06, |
| "loss": 0.267, |
| "step": 1460 |
| }, |
| { |
| "epoch": 5.975609756097561, |
| "grad_norm": 20.209896087646484, |
| "learning_rate": 4.199152542372882e-06, |
| "loss": 0.2916, |
| "step": 1470 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_general_loss": 0.7351013422012329, |
| "eval_general_runtime": 254.9676, |
| "eval_general_samples_per_second": 3.589, |
| "eval_general_steps_per_second": 0.898, |
| "step": 1476 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_code_loss": 0.5539229512214661, |
| "eval_code_runtime": 297.3043, |
| "eval_code_samples_per_second": 3.088, |
| "eval_code_steps_per_second": 0.774, |
| "step": 1476 |
| }, |
| { |
| "epoch": 6.0, |
| "eval_stem_loss": 0.8781471252441406, |
| "eval_stem_runtime": 250.9123, |
| "eval_stem_samples_per_second": 3.643, |
| "eval_stem_steps_per_second": 0.913, |
| "step": 1476 |
| }, |
| { |
| "epoch": 6.016260162601626, |
| "grad_norm": 23.147737503051758, |
| "learning_rate": 4.1567796610169495e-06, |
| "loss": 0.298, |
| "step": 1480 |
| }, |
| { |
| "epoch": 6.056910569105691, |
| "grad_norm": 13.436469078063965, |
| "learning_rate": 4.114406779661017e-06, |
| "loss": 0.2302, |
| "step": 1490 |
| }, |
| { |
| "epoch": 6.097560975609756, |
| "grad_norm": 7.810673236846924, |
| "learning_rate": 4.072033898305085e-06, |
| "loss": 0.3332, |
| "step": 1500 |
| }, |
| { |
| "epoch": 6.138211382113822, |
| "grad_norm": 18.086008071899414, |
| "learning_rate": 4.029661016949153e-06, |
| "loss": 0.2208, |
| "step": 1510 |
| }, |
| { |
| "epoch": 6.178861788617886, |
| "grad_norm": 15.756752967834473, |
| "learning_rate": 3.987288135593221e-06, |
| "loss": 0.2687, |
| "step": 1520 |
| }, |
| { |
| "epoch": 6.219512195121951, |
| "grad_norm": 4.063956260681152, |
| "learning_rate": 3.9449152542372884e-06, |
| "loss": 0.1873, |
| "step": 1530 |
| }, |
| { |
| "epoch": 6.260162601626016, |
| "grad_norm": 22.55267333984375, |
| "learning_rate": 3.902542372881356e-06, |
| "loss": 0.309, |
| "step": 1540 |
| }, |
| { |
| "epoch": 6.300813008130081, |
| "grad_norm": 3.895887613296509, |
| "learning_rate": 3.860169491525424e-06, |
| "loss": 0.2664, |
| "step": 1550 |
| }, |
| { |
| "epoch": 6.341463414634147, |
| "grad_norm": 7.404182434082031, |
| "learning_rate": 3.817796610169492e-06, |
| "loss": 0.2889, |
| "step": 1560 |
| }, |
| { |
| "epoch": 6.382113821138211, |
| "grad_norm": 9.983993530273438, |
| "learning_rate": 3.77542372881356e-06, |
| "loss": 0.2803, |
| "step": 1570 |
| }, |
| { |
| "epoch": 6.4227642276422765, |
| "grad_norm": 7.133581638336182, |
| "learning_rate": 3.7330508474576277e-06, |
| "loss": 0.2078, |
| "step": 1580 |
| }, |
| { |
| "epoch": 6.463414634146342, |
| "grad_norm": 4.602142333984375, |
| "learning_rate": 3.690677966101695e-06, |
| "loss": 0.2432, |
| "step": 1590 |
| }, |
| { |
| "epoch": 6.504065040650406, |
| "grad_norm": 20.32489585876465, |
| "learning_rate": 3.648305084745763e-06, |
| "loss": 0.2643, |
| "step": 1600 |
| }, |
| { |
| "epoch": 6.544715447154472, |
| "grad_norm": 2.3672261238098145, |
| "learning_rate": 3.605932203389831e-06, |
| "loss": 0.2641, |
| "step": 1610 |
| }, |
| { |
| "epoch": 6.585365853658536, |
| "grad_norm": 3.9571468830108643, |
| "learning_rate": 3.5635593220338988e-06, |
| "loss": 0.2888, |
| "step": 1620 |
| }, |
| { |
| "epoch": 6.626016260162602, |
| "grad_norm": 7.1855902671813965, |
| "learning_rate": 3.5211864406779667e-06, |
| "loss": 0.2808, |
| "step": 1630 |
| }, |
| { |
| "epoch": 6.666666666666667, |
| "grad_norm": 17.62175750732422, |
| "learning_rate": 3.478813559322034e-06, |
| "loss": 0.1877, |
| "step": 1640 |
| }, |
| { |
| "epoch": 6.7073170731707314, |
| "grad_norm": 23.9554443359375, |
| "learning_rate": 3.436440677966102e-06, |
| "loss": 0.2536, |
| "step": 1650 |
| }, |
| { |
| "epoch": 6.747967479674797, |
| "grad_norm": 18.637380599975586, |
| "learning_rate": 3.39406779661017e-06, |
| "loss": 0.2285, |
| "step": 1660 |
| }, |
| { |
| "epoch": 6.788617886178862, |
| "grad_norm": 4.296135902404785, |
| "learning_rate": 3.3516949152542377e-06, |
| "loss": 0.2408, |
| "step": 1670 |
| }, |
| { |
| "epoch": 6.829268292682927, |
| "grad_norm": 13.094096183776855, |
| "learning_rate": 3.309322033898305e-06, |
| "loss": 0.2354, |
| "step": 1680 |
| }, |
| { |
| "epoch": 6.869918699186992, |
| "grad_norm": 21.35910987854004, |
| "learning_rate": 3.266949152542373e-06, |
| "loss": 0.2929, |
| "step": 1690 |
| }, |
| { |
| "epoch": 6.9105691056910565, |
| "grad_norm": 11.566039085388184, |
| "learning_rate": 3.224576271186441e-06, |
| "loss": 0.2473, |
| "step": 1700 |
| }, |
| { |
| "epoch": 6.951219512195122, |
| "grad_norm": 33.27814865112305, |
| "learning_rate": 3.1822033898305087e-06, |
| "loss": 0.2745, |
| "step": 1710 |
| }, |
| { |
| "epoch": 6.991869918699187, |
| "grad_norm": 28.20252799987793, |
| "learning_rate": 3.1398305084745766e-06, |
| "loss": 0.2758, |
| "step": 1720 |
| }, |
| { |
| "epoch": 7.0, |
| "eval_general_loss": 0.7735296487808228, |
| "eval_general_runtime": 254.8206, |
| "eval_general_samples_per_second": 3.591, |
| "eval_general_steps_per_second": 0.899, |
| "step": 1722 |
| }, |
| { |
| "epoch": 7.0, |
| "eval_code_loss": 0.5828394889831543, |
| "eval_code_runtime": 297.9571, |
| "eval_code_samples_per_second": 3.081, |
| "eval_code_steps_per_second": 0.772, |
| "step": 1722 |
| }, |
| { |
| "epoch": 7.0, |
| "eval_stem_loss": 0.9436448216438293, |
| "eval_stem_runtime": 251.491, |
| "eval_stem_samples_per_second": 3.634, |
| "eval_stem_steps_per_second": 0.911, |
| "step": 1722 |
| }, |
| { |
| "epoch": 7.032520325203252, |
| "grad_norm": 3.4257633686065674, |
| "learning_rate": 3.097457627118644e-06, |
| "loss": 0.2642, |
| "step": 1730 |
| }, |
| { |
| "epoch": 7.073170731707317, |
| "grad_norm": 5.06683349609375, |
| "learning_rate": 3.055084745762712e-06, |
| "loss": 0.1826, |
| "step": 1740 |
| }, |
| { |
| "epoch": 7.1138211382113825, |
| "grad_norm": 16.503429412841797, |
| "learning_rate": 3.0127118644067798e-06, |
| "loss": 0.2212, |
| "step": 1750 |
| }, |
| { |
| "epoch": 7.154471544715447, |
| "grad_norm": 8.533424377441406, |
| "learning_rate": 2.9703389830508476e-06, |
| "loss": 0.2552, |
| "step": 1760 |
| }, |
| { |
| "epoch": 7.195121951219512, |
| "grad_norm": 24.57413673400879, |
| "learning_rate": 2.9279661016949155e-06, |
| "loss": 0.2391, |
| "step": 1770 |
| }, |
| { |
| "epoch": 7.235772357723577, |
| "grad_norm": 9.451231002807617, |
| "learning_rate": 2.885593220338983e-06, |
| "loss": 0.2397, |
| "step": 1780 |
| }, |
| { |
| "epoch": 7.276422764227642, |
| "grad_norm": 4.1486382484436035, |
| "learning_rate": 2.843220338983051e-06, |
| "loss": 0.3058, |
| "step": 1790 |
| }, |
| { |
| "epoch": 7.317073170731708, |
| "grad_norm": 11.02759838104248, |
| "learning_rate": 2.8008474576271187e-06, |
| "loss": 0.186, |
| "step": 1800 |
| }, |
| { |
| "epoch": 7.357723577235772, |
| "grad_norm": 12.10676097869873, |
| "learning_rate": 2.758474576271187e-06, |
| "loss": 0.2678, |
| "step": 1810 |
| }, |
| { |
| "epoch": 7.3983739837398375, |
| "grad_norm": 5.671831130981445, |
| "learning_rate": 2.716101694915255e-06, |
| "loss": 0.2211, |
| "step": 1820 |
| }, |
| { |
| "epoch": 7.439024390243903, |
| "grad_norm": 7.359579086303711, |
| "learning_rate": 2.673728813559322e-06, |
| "loss": 0.1869, |
| "step": 1830 |
| }, |
| { |
| "epoch": 7.479674796747967, |
| "grad_norm": 9.749099731445312, |
| "learning_rate": 2.63135593220339e-06, |
| "loss": 0.2002, |
| "step": 1840 |
| }, |
| { |
| "epoch": 7.520325203252033, |
| "grad_norm": 10.702773094177246, |
| "learning_rate": 2.588983050847458e-06, |
| "loss": 0.2843, |
| "step": 1850 |
| }, |
| { |
| "epoch": 7.560975609756097, |
| "grad_norm": 2.2151198387145996, |
| "learning_rate": 2.546610169491526e-06, |
| "loss": 0.1893, |
| "step": 1860 |
| }, |
| { |
| "epoch": 7.6016260162601625, |
| "grad_norm": 4.354043483734131, |
| "learning_rate": 2.5042372881355937e-06, |
| "loss": 0.1663, |
| "step": 1870 |
| }, |
| { |
| "epoch": 7.642276422764228, |
| "grad_norm": 14.423757553100586, |
| "learning_rate": 2.461864406779661e-06, |
| "loss": 0.222, |
| "step": 1880 |
| }, |
| { |
| "epoch": 7.682926829268292, |
| "grad_norm": 5.424985885620117, |
| "learning_rate": 2.419491525423729e-06, |
| "loss": 0.193, |
| "step": 1890 |
| }, |
| { |
| "epoch": 7.723577235772358, |
| "grad_norm": 2.692128896713257, |
| "learning_rate": 2.377118644067797e-06, |
| "loss": 0.2082, |
| "step": 1900 |
| }, |
| { |
| "epoch": 7.764227642276423, |
| "grad_norm": 17.189456939697266, |
| "learning_rate": 2.3347457627118648e-06, |
| "loss": 0.1705, |
| "step": 1910 |
| }, |
| { |
| "epoch": 7.804878048780488, |
| "grad_norm": 8.40690803527832, |
| "learning_rate": 2.292372881355932e-06, |
| "loss": 0.2217, |
| "step": 1920 |
| }, |
| { |
| "epoch": 7.845528455284553, |
| "grad_norm": 38.97690200805664, |
| "learning_rate": 2.25e-06, |
| "loss": 0.1921, |
| "step": 1930 |
| }, |
| { |
| "epoch": 7.886178861788618, |
| "grad_norm": 5.772021293640137, |
| "learning_rate": 2.207627118644068e-06, |
| "loss": 0.2285, |
| "step": 1940 |
| }, |
| { |
| "epoch": 7.926829268292683, |
| "grad_norm": 15.428215026855469, |
| "learning_rate": 2.1652542372881358e-06, |
| "loss": 0.2625, |
| "step": 1950 |
| }, |
| { |
| "epoch": 7.967479674796748, |
| "grad_norm": 18.43503189086914, |
| "learning_rate": 2.1228813559322037e-06, |
| "loss": 0.1974, |
| "step": 1960 |
| }, |
| { |
| "epoch": 8.0, |
| "eval_general_loss": 0.8415325880050659, |
| "eval_general_runtime": 255.4935, |
| "eval_general_samples_per_second": 3.581, |
| "eval_general_steps_per_second": 0.896, |
| "step": 1968 |
| }, |
| { |
| "epoch": 8.0, |
| "eval_code_loss": 0.6173287034034729, |
| "eval_code_runtime": 297.722, |
| "eval_code_samples_per_second": 3.083, |
| "eval_code_steps_per_second": 0.773, |
| "step": 1968 |
| }, |
| { |
| "epoch": 8.0, |
| "eval_stem_loss": 1.0608537197113037, |
| "eval_stem_runtime": 251.5091, |
| "eval_stem_samples_per_second": 3.634, |
| "eval_stem_steps_per_second": 0.911, |
| "step": 1968 |
| }, |
| { |
| "epoch": 8.008130081300813, |
| "grad_norm": 17.22564125061035, |
| "learning_rate": 2.0805084745762715e-06, |
| "loss": 0.2504, |
| "step": 1970 |
| }, |
| { |
| "epoch": 8.048780487804878, |
| "grad_norm": 3.2470831871032715, |
| "learning_rate": 2.0381355932203394e-06, |
| "loss": 0.1798, |
| "step": 1980 |
| }, |
| { |
| "epoch": 8.089430894308943, |
| "grad_norm": 5.299134254455566, |
| "learning_rate": 1.995762711864407e-06, |
| "loss": 0.2662, |
| "step": 1990 |
| }, |
| { |
| "epoch": 8.130081300813009, |
| "grad_norm": 31.546850204467773, |
| "learning_rate": 1.9533898305084747e-06, |
| "loss": 0.1669, |
| "step": 2000 |
| }, |
| { |
| "epoch": 8.170731707317072, |
| "grad_norm": 4.480246543884277, |
| "learning_rate": 1.9110169491525426e-06, |
| "loss": 0.207, |
| "step": 2010 |
| }, |
| { |
| "epoch": 8.211382113821138, |
| "grad_norm": 7.999574184417725, |
| "learning_rate": 1.8686440677966102e-06, |
| "loss": 0.2235, |
| "step": 2020 |
| }, |
| { |
| "epoch": 8.252032520325203, |
| "grad_norm": 16.393817901611328, |
| "learning_rate": 1.8262711864406783e-06, |
| "loss": 0.216, |
| "step": 2030 |
| }, |
| { |
| "epoch": 8.292682926829269, |
| "grad_norm": 6.757204532623291, |
| "learning_rate": 1.7838983050847457e-06, |
| "loss": 0.1832, |
| "step": 2040 |
| }, |
| { |
| "epoch": 8.333333333333334, |
| "grad_norm": 5.856433391571045, |
| "learning_rate": 1.7415254237288138e-06, |
| "loss": 0.2552, |
| "step": 2050 |
| }, |
| { |
| "epoch": 8.373983739837398, |
| "grad_norm": 18.291067123413086, |
| "learning_rate": 1.6991525423728817e-06, |
| "loss": 0.1527, |
| "step": 2060 |
| }, |
| { |
| "epoch": 8.414634146341463, |
| "grad_norm": 13.822905540466309, |
| "learning_rate": 1.6567796610169493e-06, |
| "loss": 0.1749, |
| "step": 2070 |
| }, |
| { |
| "epoch": 8.455284552845528, |
| "grad_norm": 43.684326171875, |
| "learning_rate": 1.6144067796610172e-06, |
| "loss": 0.255, |
| "step": 2080 |
| }, |
| { |
| "epoch": 8.495934959349594, |
| "grad_norm": 4.421077728271484, |
| "learning_rate": 1.5720338983050848e-06, |
| "loss": 0.1881, |
| "step": 2090 |
| }, |
| { |
| "epoch": 8.536585365853659, |
| "grad_norm": 9.270566940307617, |
| "learning_rate": 1.5296610169491527e-06, |
| "loss": 0.188, |
| "step": 2100 |
| }, |
| { |
| "epoch": 8.577235772357724, |
| "grad_norm": 15.476509094238281, |
| "learning_rate": 1.4872881355932204e-06, |
| "loss": 0.2264, |
| "step": 2110 |
| }, |
| { |
| "epoch": 8.617886178861788, |
| "grad_norm": 11.165403366088867, |
| "learning_rate": 1.4449152542372882e-06, |
| "loss": 0.2861, |
| "step": 2120 |
| }, |
| { |
| "epoch": 8.658536585365853, |
| "grad_norm": 6.011041641235352, |
| "learning_rate": 1.402542372881356e-06, |
| "loss": 0.1759, |
| "step": 2130 |
| }, |
| { |
| "epoch": 8.699186991869919, |
| "grad_norm": 15.977751731872559, |
| "learning_rate": 1.3601694915254237e-06, |
| "loss": 0.1669, |
| "step": 2140 |
| }, |
| { |
| "epoch": 8.739837398373984, |
| "grad_norm": 8.859704971313477, |
| "learning_rate": 1.3177966101694916e-06, |
| "loss": 0.1772, |
| "step": 2150 |
| }, |
| { |
| "epoch": 8.78048780487805, |
| "grad_norm": 12.153793334960938, |
| "learning_rate": 1.2754237288135593e-06, |
| "loss": 0.1423, |
| "step": 2160 |
| }, |
| { |
| "epoch": 8.821138211382113, |
| "grad_norm": 14.467902183532715, |
| "learning_rate": 1.2330508474576271e-06, |
| "loss": 0.1438, |
| "step": 2170 |
| }, |
| { |
| "epoch": 8.861788617886178, |
| "grad_norm": 42.52398681640625, |
| "learning_rate": 1.190677966101695e-06, |
| "loss": 0.2745, |
| "step": 2180 |
| }, |
| { |
| "epoch": 8.902439024390244, |
| "grad_norm": 27.64867401123047, |
| "learning_rate": 1.1483050847457626e-06, |
| "loss": 0.2944, |
| "step": 2190 |
| }, |
| { |
| "epoch": 8.94308943089431, |
| "grad_norm": 25.19327163696289, |
| "learning_rate": 1.1059322033898307e-06, |
| "loss": 0.1636, |
| "step": 2200 |
| }, |
| { |
| "epoch": 8.983739837398375, |
| "grad_norm": 2.7334868907928467, |
| "learning_rate": 1.0635593220338984e-06, |
| "loss": 0.1786, |
| "step": 2210 |
| }, |
| { |
| "epoch": 9.0, |
| "eval_general_loss": 0.8703850507736206, |
| "eval_general_runtime": 254.9132, |
| "eval_general_samples_per_second": 3.589, |
| "eval_general_steps_per_second": 0.898, |
| "step": 2214 |
| }, |
| { |
| "epoch": 9.0, |
| "eval_code_loss": 0.6459083557128906, |
| "eval_code_runtime": 297.2179, |
| "eval_code_samples_per_second": 3.089, |
| "eval_code_steps_per_second": 0.774, |
| "step": 2214 |
| }, |
| { |
| "epoch": 9.0, |
| "eval_stem_loss": 1.1134814023971558, |
| "eval_stem_runtime": 250.7558, |
| "eval_stem_samples_per_second": 3.645, |
| "eval_stem_steps_per_second": 0.913, |
| "step": 2214 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 2460, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 10, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 5.485373470360732e+18, |
| "train_batch_size": 4, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|