| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 1.0240655401945724, | |
| "eval_steps": 500, | |
| "global_step": 1500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.41249775886535645, | |
| "learning_rate": 0.00011363636363636365, | |
| "loss": 1.6991, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.4261733889579773, | |
| "learning_rate": 0.00019999118973067944, | |
| "loss": 1.463, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.46912631392478943, | |
| "learning_rate": 0.00019976490347667713, | |
| "loss": 1.4387, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.4052003026008606, | |
| "learning_rate": 0.00019923349759697913, | |
| "loss": 1.4552, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.41492465138435364, | |
| "learning_rate": 0.00019839859733610985, | |
| "loss": 1.432, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.4481758177280426, | |
| "learning_rate": 0.0001972627561416867, | |
| "loss": 1.4365, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.5057295560836792, | |
| "learning_rate": 0.00019582944785499073, | |
| "loss": 1.4398, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.4403720796108246, | |
| "learning_rate": 0.00019410305608661744, | |
| "loss": 1.4236, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.4727613627910614, | |
| "learning_rate": 0.00019208886080970187, | |
| "loss": 1.4109, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.5325965285301208, | |
| "learning_rate": 0.0001897930222117203, | |
| "loss": 1.4556, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.5341053605079651, | |
| "learning_rate": 0.00018722256185425655, | |
| "loss": 1.426, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.6165050268173218, | |
| "learning_rate": 0.00018438534119835363, | |
| "loss": 1.3777, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.455497682094574, | |
| "learning_rate": 0.0001812900375611279, | |
| "loss": 1.4352, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.4459982216358185, | |
| "learning_rate": 0.00017794611757718015, | |
| "loss": 1.3917, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.488208144903183, | |
| "learning_rate": 0.00017436380824596915, | |
| "loss": 1.3801, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 0.6411360502243042, | |
| "learning_rate": 0.00017055406565369532, | |
| "loss": 1.4015, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 0.6391683220863342, | |
| "learning_rate": 0.0001665285414653555, | |
| "loss": 1.3578, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 0.4634678363800049, | |
| "learning_rate": 0.00016229954728944895, | |
| "loss": 1.3958, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 0.5685352087020874, | |
| "learning_rate": 0.00015788001702432132, | |
| "loss": 1.3824, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 0.48762017488479614, | |
| "learning_rate": 0.0001532834673013053, | |
| "loss": 1.3651, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 0.4447900354862213, | |
| "learning_rate": 0.00014852395614563935, | |
| "loss": 1.3928, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 0.5502452850341797, | |
| "learning_rate": 0.00014361603998159388, | |
| "loss": 1.3901, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 0.5311463475227356, | |
| "learning_rate": 0.00013857472911330131, | |
| "loss": 1.3506, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 0.5501254796981812, | |
| "learning_rate": 0.0001334154418174456, | |
| "loss": 1.3611, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 0.6120589375495911, | |
| "learning_rate": 0.00012815395718821415, | |
| "loss": 1.3879, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 0.6142286658287048, | |
| "learning_rate": 0.0001228063668787309, | |
| "loss": 1.3633, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 0.610168993473053, | |
| "learning_rate": 0.00011738902588656242, | |
| "loss": 1.3522, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 0.5812483429908752, | |
| "learning_rate": 0.00011191850253381601, | |
| "loss": 1.3938, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 0.6008601784706116, | |
| "learning_rate": 0.00010641152779480806, | |
| "loss": 1.3396, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "grad_norm": 0.6404789686203003, | |
| "learning_rate": 0.00010088494412627968, | |
| "loss": 1.3007, | |
| "step": 1500 | |
| } | |
| ], | |
| "logging_steps": 50, | |
| "max_steps": 2928, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 2, | |
| "save_steps": 500, | |
| "total_flos": 3.646597932955238e+16, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |