{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 4.0, "eval_steps": 500, "global_step": 984, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.04065040650406504, "grad_norm": 14.707608222961426, "learning_rate": 9.000000000000001e-07, "loss": 0.9613, "step": 10 }, { "epoch": 0.08130081300813008, "grad_norm": 12.035386085510254, "learning_rate": 1.9000000000000002e-06, "loss": 0.8509, "step": 20 }, { "epoch": 0.12195121951219512, "grad_norm": 18.44657325744629, "learning_rate": 2.9e-06, "loss": 0.9369, "step": 30 }, { "epoch": 0.16260162601626016, "grad_norm": 12.941998481750488, "learning_rate": 3.900000000000001e-06, "loss": 0.896, "step": 40 }, { "epoch": 0.2032520325203252, "grad_norm": 15.150432586669922, "learning_rate": 4.9000000000000005e-06, "loss": 0.9821, "step": 50 }, { "epoch": 0.24390243902439024, "grad_norm": 18.638032913208008, "learning_rate": 5.9e-06, "loss": 0.9666, "step": 60 }, { "epoch": 0.2845528455284553, "grad_norm": 15.954939842224121, "learning_rate": 6.9e-06, "loss": 0.9019, "step": 70 }, { "epoch": 0.3252032520325203, "grad_norm": 11.114273071289062, "learning_rate": 7.9e-06, "loss": 0.8321, "step": 80 }, { "epoch": 0.36585365853658536, "grad_norm": 15.285880088806152, "learning_rate": 8.900000000000001e-06, "loss": 0.7198, "step": 90 }, { "epoch": 0.4065040650406504, "grad_norm": 17.8586483001709, "learning_rate": 9.9e-06, "loss": 0.7806, "step": 100 }, { "epoch": 0.44715447154471544, "grad_norm": 11.217978477478027, "learning_rate": 9.961864406779662e-06, "loss": 1.0024, "step": 110 }, { "epoch": 0.4878048780487805, "grad_norm": 7.002477169036865, "learning_rate": 9.91949152542373e-06, "loss": 0.8086, "step": 120 }, { "epoch": 0.5284552845528455, "grad_norm": 13.211562156677246, "learning_rate": 9.877118644067798e-06, "loss": 0.8958, "step": 130 }, { "epoch": 0.5691056910569106, "grad_norm": 11.169763565063477, "learning_rate": 9.834745762711865e-06, "loss": 0.8316, "step": 140 }, { "epoch": 0.6097560975609756, "grad_norm": 15.369425773620605, "learning_rate": 9.792372881355933e-06, "loss": 0.675, "step": 150 }, { "epoch": 0.6504065040650406, "grad_norm": 11.003032684326172, "learning_rate": 9.75e-06, "loss": 0.7512, "step": 160 }, { "epoch": 0.6910569105691057, "grad_norm": 16.466957092285156, "learning_rate": 9.707627118644068e-06, "loss": 0.8546, "step": 170 }, { "epoch": 0.7317073170731707, "grad_norm": 13.067832946777344, "learning_rate": 9.665254237288136e-06, "loss": 0.7497, "step": 180 }, { "epoch": 0.7723577235772358, "grad_norm": 9.530882835388184, "learning_rate": 9.622881355932205e-06, "loss": 0.6136, "step": 190 }, { "epoch": 0.8130081300813008, "grad_norm": 15.063409805297852, "learning_rate": 9.580508474576273e-06, "loss": 0.6073, "step": 200 }, { "epoch": 0.8536585365853658, "grad_norm": 9.07528305053711, "learning_rate": 9.53813559322034e-06, "loss": 0.7295, "step": 210 }, { "epoch": 0.8943089430894309, "grad_norm": 14.199564933776855, "learning_rate": 9.495762711864408e-06, "loss": 0.596, "step": 220 }, { "epoch": 0.9349593495934959, "grad_norm": 16.03732681274414, "learning_rate": 9.453389830508474e-06, "loss": 0.6922, "step": 230 }, { "epoch": 0.975609756097561, "grad_norm": 17.533695220947266, "learning_rate": 9.411016949152543e-06, "loss": 0.7484, "step": 240 }, { "epoch": 1.0, "eval_general_loss": 0.754059910774231, "eval_general_runtime": 254.3068, "eval_general_samples_per_second": 3.598, "eval_general_steps_per_second": 0.9, "step": 246 }, { "epoch": 1.0, "eval_code_loss": 0.7234187722206116, "eval_code_runtime": 296.8568, "eval_code_samples_per_second": 3.092, "eval_code_steps_per_second": 0.775, "step": 246 }, { "epoch": 1.0, "eval_stem_loss": 0.7939386367797852, "eval_stem_runtime": 250.6578, "eval_stem_samples_per_second": 3.646, "eval_stem_steps_per_second": 0.914, "step": 246 }, { "epoch": 1.016260162601626, "grad_norm": 16.589147567749023, "learning_rate": 9.368644067796611e-06, "loss": 0.7309, "step": 250 }, { "epoch": 1.056910569105691, "grad_norm": 8.124390602111816, "learning_rate": 9.32627118644068e-06, "loss": 0.5937, "step": 260 }, { "epoch": 1.0975609756097562, "grad_norm": 13.223992347717285, "learning_rate": 9.283898305084746e-06, "loss": 0.6961, "step": 270 }, { "epoch": 1.1382113821138211, "grad_norm": 13.791741371154785, "learning_rate": 9.241525423728814e-06, "loss": 0.5783, "step": 280 }, { "epoch": 1.1788617886178863, "grad_norm": 16.130605697631836, "learning_rate": 9.199152542372882e-06, "loss": 0.6318, "step": 290 }, { "epoch": 1.2195121951219512, "grad_norm": 14.941628456115723, "learning_rate": 9.15677966101695e-06, "loss": 0.6163, "step": 300 }, { "epoch": 1.2601626016260163, "grad_norm": 6.407144546508789, "learning_rate": 9.114406779661017e-06, "loss": 0.5745, "step": 310 }, { "epoch": 1.3008130081300813, "grad_norm": 7.032970905303955, "learning_rate": 9.072033898305086e-06, "loss": 0.474, "step": 320 }, { "epoch": 1.3414634146341464, "grad_norm": 14.255295753479004, "learning_rate": 9.029661016949152e-06, "loss": 0.6833, "step": 330 }, { "epoch": 1.3821138211382114, "grad_norm": 19.13692855834961, "learning_rate": 8.98728813559322e-06, "loss": 0.5369, "step": 340 }, { "epoch": 1.4227642276422765, "grad_norm": 9.524161338806152, "learning_rate": 8.944915254237289e-06, "loss": 0.6925, "step": 350 }, { "epoch": 1.4634146341463414, "grad_norm": 12.59819221496582, "learning_rate": 8.902542372881357e-06, "loss": 0.572, "step": 360 }, { "epoch": 1.5040650406504064, "grad_norm": 7.560878753662109, "learning_rate": 8.860169491525424e-06, "loss": 0.5101, "step": 370 }, { "epoch": 1.5447154471544715, "grad_norm": 10.911989212036133, "learning_rate": 8.817796610169492e-06, "loss": 0.7357, "step": 380 }, { "epoch": 1.5853658536585367, "grad_norm": 13.019627571105957, "learning_rate": 8.77542372881356e-06, "loss": 0.7094, "step": 390 }, { "epoch": 1.6260162601626016, "grad_norm": 5.221607685089111, "learning_rate": 8.733050847457629e-06, "loss": 0.5762, "step": 400 }, { "epoch": 1.6666666666666665, "grad_norm": 9.154648780822754, "learning_rate": 8.690677966101695e-06, "loss": 0.6516, "step": 410 }, { "epoch": 1.7073170731707317, "grad_norm": 8.04185962677002, "learning_rate": 8.648305084745763e-06, "loss": 0.5457, "step": 420 }, { "epoch": 1.7479674796747968, "grad_norm": 12.428257942199707, "learning_rate": 8.60593220338983e-06, "loss": 0.6458, "step": 430 }, { "epoch": 1.7886178861788617, "grad_norm": 6.699453353881836, "learning_rate": 8.563559322033898e-06, "loss": 0.6026, "step": 440 }, { "epoch": 1.8292682926829267, "grad_norm": 5.6174774169921875, "learning_rate": 8.521186440677967e-06, "loss": 0.6756, "step": 450 }, { "epoch": 1.8699186991869918, "grad_norm": 12.72862720489502, "learning_rate": 8.478813559322035e-06, "loss": 0.5932, "step": 460 }, { "epoch": 1.910569105691057, "grad_norm": 4.651763439178467, "learning_rate": 8.436440677966102e-06, "loss": 0.418, "step": 470 }, { "epoch": 1.951219512195122, "grad_norm": 16.044872283935547, "learning_rate": 8.39406779661017e-06, "loss": 0.5756, "step": 480 }, { "epoch": 1.9918699186991868, "grad_norm": 10.388751029968262, "learning_rate": 8.351694915254238e-06, "loss": 0.6304, "step": 490 }, { "epoch": 2.0, "eval_general_loss": 0.6887519359588623, "eval_general_runtime": 255.0576, "eval_general_samples_per_second": 3.587, "eval_general_steps_per_second": 0.898, "step": 492 }, { "epoch": 2.0, "eval_code_loss": 0.5912412405014038, "eval_code_runtime": 297.8316, "eval_code_samples_per_second": 3.082, "eval_code_steps_per_second": 0.772, "step": 492 }, { "epoch": 2.0, "eval_stem_loss": 0.7283160090446472, "eval_stem_runtime": 251.4929, "eval_stem_samples_per_second": 3.634, "eval_stem_steps_per_second": 0.911, "step": 492 }, { "epoch": 2.032520325203252, "grad_norm": 13.620870590209961, "learning_rate": 8.309322033898306e-06, "loss": 0.5292, "step": 500 }, { "epoch": 2.073170731707317, "grad_norm": 11.176264762878418, "learning_rate": 8.266949152542375e-06, "loss": 0.4926, "step": 510 }, { "epoch": 2.113821138211382, "grad_norm": 4.5993852615356445, "learning_rate": 8.224576271186441e-06, "loss": 0.476, "step": 520 }, { "epoch": 2.154471544715447, "grad_norm": 16.03154182434082, "learning_rate": 8.182203389830508e-06, "loss": 0.492, "step": 530 }, { "epoch": 2.1951219512195124, "grad_norm": 3.7860374450683594, "learning_rate": 8.139830508474576e-06, "loss": 0.4376, "step": 540 }, { "epoch": 2.2357723577235773, "grad_norm": 14.141989707946777, "learning_rate": 8.097457627118644e-06, "loss": 0.5422, "step": 550 }, { "epoch": 2.2764227642276422, "grad_norm": 11.716277122497559, "learning_rate": 8.055084745762713e-06, "loss": 0.4781, "step": 560 }, { "epoch": 2.317073170731707, "grad_norm": 16.112192153930664, "learning_rate": 8.012711864406781e-06, "loss": 0.5243, "step": 570 }, { "epoch": 2.3577235772357725, "grad_norm": 5.857080936431885, "learning_rate": 7.970338983050848e-06, "loss": 0.4588, "step": 580 }, { "epoch": 2.3983739837398375, "grad_norm": 9.052570343017578, "learning_rate": 7.927966101694916e-06, "loss": 0.5717, "step": 590 }, { "epoch": 2.4390243902439024, "grad_norm": 17.8330078125, "learning_rate": 7.885593220338984e-06, "loss": 0.4547, "step": 600 }, { "epoch": 2.4796747967479673, "grad_norm": 13.779817581176758, "learning_rate": 7.843220338983052e-06, "loss": 0.4584, "step": 610 }, { "epoch": 2.5203252032520327, "grad_norm": 18.924348831176758, "learning_rate": 7.800847457627119e-06, "loss": 0.4788, "step": 620 }, { "epoch": 2.5609756097560976, "grad_norm": 12.157154083251953, "learning_rate": 7.758474576271187e-06, "loss": 0.4762, "step": 630 }, { "epoch": 2.6016260162601625, "grad_norm": 14.143574714660645, "learning_rate": 7.716101694915254e-06, "loss": 0.7017, "step": 640 }, { "epoch": 2.642276422764228, "grad_norm": 11.326656341552734, "learning_rate": 7.673728813559322e-06, "loss": 0.4524, "step": 650 }, { "epoch": 2.682926829268293, "grad_norm": 5.236691474914551, "learning_rate": 7.63135593220339e-06, "loss": 0.5228, "step": 660 }, { "epoch": 2.7235772357723578, "grad_norm": 17.48626708984375, "learning_rate": 7.588983050847458e-06, "loss": 0.4414, "step": 670 }, { "epoch": 2.7642276422764227, "grad_norm": 13.257789611816406, "learning_rate": 7.546610169491526e-06, "loss": 0.5442, "step": 680 }, { "epoch": 2.8048780487804876, "grad_norm": 20.655475616455078, "learning_rate": 7.504237288135594e-06, "loss": 0.5451, "step": 690 }, { "epoch": 2.845528455284553, "grad_norm": 10.467672348022461, "learning_rate": 7.461864406779662e-06, "loss": 0.4601, "step": 700 }, { "epoch": 2.886178861788618, "grad_norm": 13.194644927978516, "learning_rate": 7.41949152542373e-06, "loss": 0.502, "step": 710 }, { "epoch": 2.926829268292683, "grad_norm": 9.131794929504395, "learning_rate": 7.377118644067798e-06, "loss": 0.5583, "step": 720 }, { "epoch": 2.9674796747967482, "grad_norm": 8.837148666381836, "learning_rate": 7.334745762711864e-06, "loss": 0.4668, "step": 730 }, { "epoch": 3.0, "eval_general_loss": 0.6758893728256226, "eval_general_runtime": 255.2921, "eval_general_samples_per_second": 3.584, "eval_general_steps_per_second": 0.897, "step": 738 }, { "epoch": 3.0, "eval_code_loss": 0.5598136782646179, "eval_code_runtime": 297.7838, "eval_code_samples_per_second": 3.083, "eval_code_steps_per_second": 0.772, "step": 738 }, { "epoch": 3.0, "eval_stem_loss": 0.7172386050224304, "eval_stem_runtime": 251.2967, "eval_stem_samples_per_second": 3.637, "eval_stem_steps_per_second": 0.911, "step": 738 }, { "epoch": 3.008130081300813, "grad_norm": 8.791101455688477, "learning_rate": 7.292372881355933e-06, "loss": 0.3545, "step": 740 }, { "epoch": 3.048780487804878, "grad_norm": 7.602899074554443, "learning_rate": 7.25e-06, "loss": 0.4633, "step": 750 }, { "epoch": 3.089430894308943, "grad_norm": 10.939055442810059, "learning_rate": 7.207627118644068e-06, "loss": 0.3989, "step": 760 }, { "epoch": 3.130081300813008, "grad_norm": 9.168066024780273, "learning_rate": 7.165254237288136e-06, "loss": 0.5555, "step": 770 }, { "epoch": 3.1707317073170733, "grad_norm": 17.25099754333496, "learning_rate": 7.122881355932204e-06, "loss": 0.5326, "step": 780 }, { "epoch": 3.2113821138211383, "grad_norm": 13.0396146774292, "learning_rate": 7.080508474576272e-06, "loss": 0.3831, "step": 790 }, { "epoch": 3.252032520325203, "grad_norm": 11.217447280883789, "learning_rate": 7.03813559322034e-06, "loss": 0.4142, "step": 800 }, { "epoch": 3.292682926829268, "grad_norm": 8.102045059204102, "learning_rate": 6.995762711864408e-06, "loss": 0.4414, "step": 810 }, { "epoch": 3.3333333333333335, "grad_norm": 11.761209487915039, "learning_rate": 6.953389830508475e-06, "loss": 0.501, "step": 820 }, { "epoch": 3.3739837398373984, "grad_norm": 7.058698654174805, "learning_rate": 6.911016949152542e-06, "loss": 0.3771, "step": 830 }, { "epoch": 3.4146341463414633, "grad_norm": 7.728908061981201, "learning_rate": 6.86864406779661e-06, "loss": 0.3867, "step": 840 }, { "epoch": 3.4552845528455283, "grad_norm": 12.833827018737793, "learning_rate": 6.826271186440679e-06, "loss": 0.4132, "step": 850 }, { "epoch": 3.4959349593495936, "grad_norm": 9.884450912475586, "learning_rate": 6.783898305084746e-06, "loss": 0.3674, "step": 860 }, { "epoch": 3.5365853658536586, "grad_norm": 10.731056213378906, "learning_rate": 6.741525423728814e-06, "loss": 0.422, "step": 870 }, { "epoch": 3.5772357723577235, "grad_norm": 6.745728015899658, "learning_rate": 6.699152542372882e-06, "loss": 0.499, "step": 880 }, { "epoch": 3.617886178861789, "grad_norm": 9.808424949645996, "learning_rate": 6.65677966101695e-06, "loss": 0.2978, "step": 890 }, { "epoch": 3.658536585365854, "grad_norm": 8.444202423095703, "learning_rate": 6.614406779661018e-06, "loss": 0.3698, "step": 900 }, { "epoch": 3.6991869918699187, "grad_norm": 10.285717964172363, "learning_rate": 6.572033898305086e-06, "loss": 0.4677, "step": 910 }, { "epoch": 3.7398373983739837, "grad_norm": 6.763285160064697, "learning_rate": 6.5296610169491525e-06, "loss": 0.4449, "step": 920 }, { "epoch": 3.7804878048780486, "grad_norm": 20.608469009399414, "learning_rate": 6.487288135593221e-06, "loss": 0.5297, "step": 930 }, { "epoch": 3.821138211382114, "grad_norm": 9.06677532196045, "learning_rate": 6.444915254237288e-06, "loss": 0.4292, "step": 940 }, { "epoch": 3.861788617886179, "grad_norm": 6.93435525894165, "learning_rate": 6.4025423728813565e-06, "loss": 0.3877, "step": 950 }, { "epoch": 3.902439024390244, "grad_norm": 11.75598430633545, "learning_rate": 6.360169491525424e-06, "loss": 0.4032, "step": 960 }, { "epoch": 3.943089430894309, "grad_norm": 6.338623046875, "learning_rate": 6.317796610169492e-06, "loss": 0.4228, "step": 970 }, { "epoch": 3.983739837398374, "grad_norm": 6.302221775054932, "learning_rate": 6.27542372881356e-06, "loss": 0.4054, "step": 980 }, { "epoch": 4.0, "eval_general_loss": 0.6872960329055786, "eval_general_runtime": 254.991, "eval_general_samples_per_second": 3.588, "eval_general_steps_per_second": 0.898, "step": 984 }, { "epoch": 4.0, "eval_code_loss": 0.5443978309631348, "eval_code_runtime": 297.6318, "eval_code_samples_per_second": 3.084, "eval_code_steps_per_second": 0.773, "step": 984 }, { "epoch": 4.0, "eval_stem_loss": 0.7525980472564697, "eval_stem_runtime": 251.3515, "eval_stem_samples_per_second": 3.636, "eval_stem_steps_per_second": 0.911, "step": 984 } ], "logging_steps": 10, "max_steps": 2460, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.4388046396105687e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }