{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.994106090373281, "eval_steps": 500, "global_step": 381, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.07858546168958742, "grad_norm": 5.295495045457736, "learning_rate": 2.564102564102564e-06, "loss": 1.1348, "num_input_tokens_seen": 285944, "step": 10 }, { "epoch": 0.15717092337917485, "grad_norm": 1.3364250221425176, "learning_rate": 5.128205128205128e-06, "loss": 0.9328, "num_input_tokens_seen": 553248, "step": 20 }, { "epoch": 0.2357563850687623, "grad_norm": 0.8492034246167627, "learning_rate": 7.692307692307694e-06, "loss": 0.8484, "num_input_tokens_seen": 822480, "step": 30 }, { "epoch": 0.3143418467583497, "grad_norm": 0.7234244773864981, "learning_rate": 9.999789047591563e-06, "loss": 0.758, "num_input_tokens_seen": 1085968, "step": 40 }, { "epoch": 0.3929273084479371, "grad_norm": 0.7012241271754472, "learning_rate": 9.974496289936769e-06, "loss": 0.7506, "num_input_tokens_seen": 1332664, "step": 50 }, { "epoch": 0.4715127701375246, "grad_norm": 0.6727005740081735, "learning_rate": 9.90725746626209e-06, "loss": 0.7311, "num_input_tokens_seen": 1605712, "step": 60 }, { "epoch": 0.550098231827112, "grad_norm": 0.5942119076165618, "learning_rate": 9.798639549376946e-06, "loss": 0.6815, "num_input_tokens_seen": 1896456, "step": 70 }, { "epoch": 0.6286836935166994, "grad_norm": 0.7307490789781241, "learning_rate": 9.64955842986544e-06, "loss": 0.7245, "num_input_tokens_seen": 2161224, "step": 80 }, { "epoch": 0.7072691552062869, "grad_norm": 0.5604571056904298, "learning_rate": 9.461271193091971e-06, "loss": 0.6255, "num_input_tokens_seen": 2467224, "step": 90 }, { "epoch": 0.7858546168958742, "grad_norm": 0.6954309102689223, "learning_rate": 9.23536551917611e-06, "loss": 0.6309, "num_input_tokens_seen": 2743848, "step": 100 }, { "epoch": 0.8644400785854617, "grad_norm": 0.6105791636001258, "learning_rate": 8.973746295318499e-06, "loss": 0.6661, "num_input_tokens_seen": 3015224, "step": 110 }, { "epoch": 0.9430255402750491, "grad_norm": 0.7555612161614201, "learning_rate": 8.67861955336566e-06, "loss": 0.6815, "num_input_tokens_seen": 3274408, "step": 120 }, { "epoch": 1.0216110019646365, "grad_norm": 0.736180742069797, "learning_rate": 8.352473868055746e-06, "loss": 0.6733, "num_input_tokens_seen": 3552384, "step": 130 }, { "epoch": 1.1001964636542239, "grad_norm": 0.8186129592841361, "learning_rate": 7.998059372799409e-06, "loss": 0.545, "num_input_tokens_seen": 3820496, "step": 140 }, { "epoch": 1.1787819253438114, "grad_norm": 0.7297158276181345, "learning_rate": 7.61836456993939e-06, "loss": 0.5028, "num_input_tokens_seen": 4086592, "step": 150 }, { "epoch": 1.2573673870333988, "grad_norm": 0.7309791930989968, "learning_rate": 7.2165911310299305e-06, "loss": 0.5087, "num_input_tokens_seen": 4355776, "step": 160 }, { "epoch": 1.3359528487229864, "grad_norm": 0.7564499779548668, "learning_rate": 6.796126899625688e-06, "loss": 0.49, "num_input_tokens_seen": 4634424, "step": 170 }, { "epoch": 1.4145383104125737, "grad_norm": 0.7980536006544474, "learning_rate": 6.360517324226676e-06, "loss": 0.4941, "num_input_tokens_seen": 4909728, "step": 180 }, { "epoch": 1.493123772102161, "grad_norm": 0.7591010900995777, "learning_rate": 5.913435562263036e-06, "loss": 0.4711, "num_input_tokens_seen": 5195200, "step": 190 }, { "epoch": 1.5717092337917484, "grad_norm": 0.7101983358318353, "learning_rate": 5.458651507209518e-06, "loss": 0.497, "num_input_tokens_seen": 5462608, "step": 200 }, { "epoch": 1.650294695481336, "grad_norm": 0.8796806516365058, "learning_rate": 5e-06, "loss": 0.482, "num_input_tokens_seen": 5730248, "step": 210 }, { "epoch": 1.7288801571709234, "grad_norm": 0.8085067044469503, "learning_rate": 4.541348492790482e-06, "loss": 0.5014, "num_input_tokens_seen": 5990552, "step": 220 }, { "epoch": 1.807465618860511, "grad_norm": 0.7890332198315568, "learning_rate": 4.0865644377369666e-06, "loss": 0.4757, "num_input_tokens_seen": 6273288, "step": 230 }, { "epoch": 1.8860510805500983, "grad_norm": 0.7795005553479045, "learning_rate": 3.639482675773324e-06, "loss": 0.4583, "num_input_tokens_seen": 6553400, "step": 240 }, { "epoch": 1.9646365422396856, "grad_norm": 0.7390457626908427, "learning_rate": 3.203873100374314e-06, "loss": 0.4617, "num_input_tokens_seen": 6836072, "step": 250 }, { "epoch": 2.043222003929273, "grad_norm": 0.8497968488938652, "learning_rate": 2.783408868970071e-06, "loss": 0.4566, "num_input_tokens_seen": 7093808, "step": 260 }, { "epoch": 2.1218074656188604, "grad_norm": 0.7260683682391704, "learning_rate": 2.381635430060611e-06, "loss": 0.3637, "num_input_tokens_seen": 7368800, "step": 270 }, { "epoch": 2.2003929273084477, "grad_norm": 0.8975430950896504, "learning_rate": 2.0019406272005913e-06, "loss": 0.3459, "num_input_tokens_seen": 7655808, "step": 280 }, { "epoch": 2.2789783889980355, "grad_norm": 0.7999294865618197, "learning_rate": 1.6475261319442553e-06, "loss": 0.3553, "num_input_tokens_seen": 7911600, "step": 290 }, { "epoch": 2.357563850687623, "grad_norm": 1.08697856574494, "learning_rate": 1.321380446634342e-06, "loss": 0.3451, "num_input_tokens_seen": 8192536, "step": 300 }, { "epoch": 2.43614931237721, "grad_norm": 0.9266005735156818, "learning_rate": 1.026253704681502e-06, "loss": 0.3735, "num_input_tokens_seen": 8465008, "step": 310 }, { "epoch": 2.5147347740667976, "grad_norm": 0.988052963356121, "learning_rate": 7.646344808238904e-07, "loss": 0.373, "num_input_tokens_seen": 8732552, "step": 320 }, { "epoch": 2.593320235756385, "grad_norm": 0.7077241690549176, "learning_rate": 5.387288069080298e-07, "loss": 0.3504, "num_input_tokens_seen": 9006928, "step": 330 }, { "epoch": 2.6719056974459727, "grad_norm": 0.8753220952555882, "learning_rate": 3.504415701345615e-07, "loss": 0.3255, "num_input_tokens_seen": 9272760, "step": 340 }, { "epoch": 2.75049115913556, "grad_norm": 0.8961502686499399, "learning_rate": 2.0136045062305543e-07, "loss": 0.337, "num_input_tokens_seen": 9539904, "step": 350 }, { "epoch": 2.8290766208251474, "grad_norm": 0.8575326308820346, "learning_rate": 9.274253373791064e-08, "loss": 0.381, "num_input_tokens_seen": 9813344, "step": 360 }, { "epoch": 2.907662082514735, "grad_norm": 0.729386140153712, "learning_rate": 2.55037100632316e-08, "loss": 0.3531, "num_input_tokens_seen": 10090560, "step": 370 }, { "epoch": 2.986247544204322, "grad_norm": 0.7450157172804136, "learning_rate": 2.1095240843815868e-10, "loss": 0.359, "num_input_tokens_seen": 10366800, "step": 380 }, { "epoch": 2.994106090373281, "num_input_tokens_seen": 10403944, "step": 381, "total_flos": 16608737034240.0, "train_loss": 0.5381014986926802, "train_runtime": 3056.3101, "train_samples_per_second": 1.996, "train_steps_per_second": 0.125 } ], "logging_steps": 10, "max_steps": 381, "num_input_tokens_seen": 10403944, "num_train_epochs": 3, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 16608737034240.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }