{ "best_global_step": 3261, "best_metric": 3.45804762840271, "best_model_checkpoint": "sindhibert_session6r/checkpoint-3261", "epoch": 1.0, "eval_steps": 3261, "global_step": 3261, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.030672494440610383, "grad_norm": 5.580650806427002, "learning_rate": 1.2659846547314581e-06, "loss": 14.62664794921875, "step": 100 }, { "epoch": 0.061344988881220766, "grad_norm": 5.809163570404053, "learning_rate": 2.544757033248082e-06, "loss": 14.634923095703124, "step": 200 }, { "epoch": 0.09201748332183114, "grad_norm": 5.849374294281006, "learning_rate": 3.8235294117647055e-06, "loss": 14.635355224609375, "step": 300 }, { "epoch": 0.12268997776244153, "grad_norm": 5.751659870147705, "learning_rate": 4.999978994815915e-06, "loss": 14.59581298828125, "step": 400 }, { "epoch": 0.1533624722030519, "grad_norm": 5.633106708526611, "learning_rate": 4.996172776747675e-06, "loss": 14.598475341796876, "step": 500 }, { "epoch": 0.18403496664366228, "grad_norm": 5.498394012451172, "learning_rate": 4.9858139123513936e-06, "loss": 14.56541259765625, "step": 600 }, { "epoch": 0.2147074610842727, "grad_norm": 5.918303489685059, "learning_rate": 4.968929594446166e-06, "loss": 14.57242919921875, "step": 700 }, { "epoch": 0.24537995552488306, "grad_norm": 5.2229814529418945, "learning_rate": 4.945564145670534e-06, "loss": 14.53744873046875, "step": 800 }, { "epoch": 0.27605244996549344, "grad_norm": 5.516529560089111, "learning_rate": 4.9157789021321375e-06, "loss": 14.531802978515625, "step": 900 }, { "epoch": 0.3067249444061038, "grad_norm": 5.460958003997803, "learning_rate": 4.879652052395696e-06, "loss": 14.5186328125, "step": 1000 }, { "epoch": 0.3373974388467142, "grad_norm": 5.609668731689453, "learning_rate": 4.837278432231982e-06, "loss": 14.51820068359375, "step": 1100 }, { "epoch": 0.36806993328732457, "grad_norm": 5.466991424560547, "learning_rate": 4.788769275666605e-06, "loss": 14.477803955078125, "step": 1200 }, { "epoch": 0.398742427727935, "grad_norm": 5.586793899536133, "learning_rate": 4.7342519229820996e-06, "loss": 14.506885986328125, "step": 1300 }, { "epoch": 0.4294149221685454, "grad_norm": 5.488733768463135, "learning_rate": 4.673869486439846e-06, "loss": 14.464423828125, "step": 1400 }, { "epoch": 0.46008741660915575, "grad_norm": 5.8738017082214355, "learning_rate": 4.60778047459934e-06, "loss": 14.470579833984376, "step": 1500 }, { "epoch": 0.49075991104976613, "grad_norm": 5.592001438140869, "learning_rate": 4.536158376220971e-06, "loss": 14.42775146484375, "step": 1600 }, { "epoch": 0.5214324054903765, "grad_norm": 5.235401153564453, "learning_rate": 4.459191204844634e-06, "loss": 14.42791259765625, "step": 1700 }, { "epoch": 0.5521048999309869, "grad_norm": 5.959557056427002, "learning_rate": 4.377081005239665e-06, "loss": 14.444658203125, "step": 1800 }, { "epoch": 0.5827773943715973, "grad_norm": 5.362053394317627, "learning_rate": 4.2900433230217155e-06, "loss": 14.463519287109374, "step": 1900 }, { "epoch": 0.6134498888122076, "grad_norm": 5.445187568664551, "learning_rate": 4.198306638828863e-06, "loss": 14.472896728515625, "step": 2000 }, { "epoch": 0.644122383252818, "grad_norm": 5.700764179229736, "learning_rate": 4.102111768542294e-06, "loss": 14.44870849609375, "step": 2100 }, { "epoch": 0.6747948776934284, "grad_norm": 5.197051525115967, "learning_rate": 4.001711231126012e-06, "loss": 14.42983642578125, "step": 2200 }, { "epoch": 0.7054673721340388, "grad_norm": 5.519675254821777, "learning_rate": 3.897368585745059e-06, "loss": 14.4178271484375, "step": 2300 }, { "epoch": 0.7361398665746491, "grad_norm": 5.8170647621154785, "learning_rate": 3.789357739902342e-06, "loss": 14.38358642578125, "step": 2400 }, { "epoch": 0.7668123610152595, "grad_norm": 5.513352394104004, "learning_rate": 3.6779622304102957e-06, "loss": 14.41911376953125, "step": 2500 }, { "epoch": 0.79748485545587, "grad_norm": 5.644667625427246, "learning_rate": 3.5634744790848436e-06, "loss": 14.40310791015625, "step": 2600 }, { "epoch": 0.8281573498964804, "grad_norm": 5.676363945007324, "learning_rate": 3.4461950251155457e-06, "loss": 14.40308349609375, "step": 2700 }, { "epoch": 0.8588298443370908, "grad_norm": 5.251427173614502, "learning_rate": 3.326431736127017e-06, "loss": 14.39386474609375, "step": 2800 }, { "epoch": 0.8895023387777011, "grad_norm": 5.631777286529541, "learning_rate": 3.2044990000026366e-06, "loss": 14.374044189453125, "step": 2900 }, { "epoch": 0.9201748332183115, "grad_norm": 5.5782670974731445, "learning_rate": 3.080716899592065e-06, "loss": 14.37978515625, "step": 3000 }, { "epoch": 0.9508473276589219, "grad_norm": 5.784708023071289, "learning_rate": 2.9554103724690526e-06, "loss": 14.3670703125, "step": 3100 }, { "epoch": 0.9815198220995323, "grad_norm": 5.501222610473633, "learning_rate": 2.8289083579452043e-06, "loss": 14.374014892578124, "step": 3200 }, { "epoch": 1.0, "eval_loss": 3.45804762840271, "eval_runtime": 13.1436, "eval_samples_per_second": 641.376, "eval_steps_per_second": 10.043, "step": 3261 } ], "logging_steps": 100, "max_steps": 6522, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 3261, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 2.1967584480474624e+17, "train_batch_size": 64, "trial_name": null, "trial_params": null }