{ "best_global_step": 6522, "best_metric": 3.4418444633483887, "best_model_checkpoint": "sindhibert_session6r/checkpoint-6522", "epoch": 2.0, "eval_steps": 3261, "global_step": 6522, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.030672494440610383, "grad_norm": 5.580650806427002, "learning_rate": 1.2659846547314581e-06, "loss": 14.62664794921875, "step": 100 }, { "epoch": 0.061344988881220766, "grad_norm": 5.809163570404053, "learning_rate": 2.544757033248082e-06, "loss": 14.634923095703124, "step": 200 }, { "epoch": 0.09201748332183114, "grad_norm": 5.849374294281006, "learning_rate": 3.8235294117647055e-06, "loss": 14.635355224609375, "step": 300 }, { "epoch": 0.12268997776244153, "grad_norm": 5.751659870147705, "learning_rate": 4.999978994815915e-06, "loss": 14.59581298828125, "step": 400 }, { "epoch": 0.1533624722030519, "grad_norm": 5.633106708526611, "learning_rate": 4.996172776747675e-06, "loss": 14.598475341796876, "step": 500 }, { "epoch": 0.18403496664366228, "grad_norm": 5.498394012451172, "learning_rate": 4.9858139123513936e-06, "loss": 14.56541259765625, "step": 600 }, { "epoch": 0.2147074610842727, "grad_norm": 5.918303489685059, "learning_rate": 4.968929594446166e-06, "loss": 14.57242919921875, "step": 700 }, { "epoch": 0.24537995552488306, "grad_norm": 5.2229814529418945, "learning_rate": 4.945564145670534e-06, "loss": 14.53744873046875, "step": 800 }, { "epoch": 0.27605244996549344, "grad_norm": 5.516529560089111, "learning_rate": 4.9157789021321375e-06, "loss": 14.531802978515625, "step": 900 }, { "epoch": 0.3067249444061038, "grad_norm": 5.460958003997803, "learning_rate": 4.879652052395696e-06, "loss": 14.5186328125, "step": 1000 }, { "epoch": 0.3373974388467142, "grad_norm": 5.609668731689453, "learning_rate": 4.837278432231982e-06, "loss": 14.51820068359375, "step": 1100 }, { "epoch": 0.36806993328732457, "grad_norm": 5.466991424560547, "learning_rate": 4.788769275666605e-06, "loss": 14.477803955078125, "step": 1200 }, { "epoch": 0.398742427727935, "grad_norm": 5.586793899536133, "learning_rate": 4.7342519229820996e-06, "loss": 14.506885986328125, "step": 1300 }, { "epoch": 0.4294149221685454, "grad_norm": 5.488733768463135, "learning_rate": 4.673869486439846e-06, "loss": 14.464423828125, "step": 1400 }, { "epoch": 0.46008741660915575, "grad_norm": 5.8738017082214355, "learning_rate": 4.60778047459934e-06, "loss": 14.470579833984376, "step": 1500 }, { "epoch": 0.49075991104976613, "grad_norm": 5.592001438140869, "learning_rate": 4.536158376220971e-06, "loss": 14.42775146484375, "step": 1600 }, { "epoch": 0.5214324054903765, "grad_norm": 5.235401153564453, "learning_rate": 4.459191204844634e-06, "loss": 14.42791259765625, "step": 1700 }, { "epoch": 0.5521048999309869, "grad_norm": 5.959557056427002, "learning_rate": 4.377081005239665e-06, "loss": 14.444658203125, "step": 1800 }, { "epoch": 0.5827773943715973, "grad_norm": 5.362053394317627, "learning_rate": 4.2900433230217155e-06, "loss": 14.463519287109374, "step": 1900 }, { "epoch": 0.6134498888122076, "grad_norm": 5.445187568664551, "learning_rate": 4.198306638828863e-06, "loss": 14.472896728515625, "step": 2000 }, { "epoch": 0.644122383252818, "grad_norm": 5.700764179229736, "learning_rate": 4.102111768542294e-06, "loss": 14.44870849609375, "step": 2100 }, { "epoch": 0.6747948776934284, "grad_norm": 5.197051525115967, "learning_rate": 4.001711231126012e-06, "loss": 14.42983642578125, "step": 2200 }, { "epoch": 0.7054673721340388, "grad_norm": 5.519675254821777, "learning_rate": 3.897368585745059e-06, "loss": 14.4178271484375, "step": 2300 }, { "epoch": 0.7361398665746491, "grad_norm": 5.8170647621154785, "learning_rate": 3.789357739902342e-06, "loss": 14.38358642578125, "step": 2400 }, { "epoch": 0.7668123610152595, "grad_norm": 5.513352394104004, "learning_rate": 3.6779622304102957e-06, "loss": 14.41911376953125, "step": 2500 }, { "epoch": 0.79748485545587, "grad_norm": 5.644667625427246, "learning_rate": 3.5634744790848436e-06, "loss": 14.40310791015625, "step": 2600 }, { "epoch": 0.8281573498964804, "grad_norm": 5.676363945007324, "learning_rate": 3.4461950251155457e-06, "loss": 14.40308349609375, "step": 2700 }, { "epoch": 0.8588298443370908, "grad_norm": 5.251427173614502, "learning_rate": 3.326431736127017e-06, "loss": 14.39386474609375, "step": 2800 }, { "epoch": 0.8895023387777011, "grad_norm": 5.631777286529541, "learning_rate": 3.2044990000026366e-06, "loss": 14.374044189453125, "step": 2900 }, { "epoch": 0.9201748332183115, "grad_norm": 5.5782670974731445, "learning_rate": 3.080716899592065e-06, "loss": 14.37978515625, "step": 3000 }, { "epoch": 0.9508473276589219, "grad_norm": 5.784708023071289, "learning_rate": 2.9554103724690526e-06, "loss": 14.3670703125, "step": 3100 }, { "epoch": 0.9815198220995323, "grad_norm": 5.501222610473633, "learning_rate": 2.8289083579452043e-06, "loss": 14.374014892578124, "step": 3200 }, { "epoch": 1.0, "eval_loss": 3.45804762840271, "eval_runtime": 13.1436, "eval_samples_per_second": 641.376, "eval_steps_per_second": 10.043, "step": 3261 }, { "epoch": 1.011962272831838, "grad_norm": 5.622377872467041, "learning_rate": 2.7015429335788844e-06, "loss": 14.245347900390625, "step": 3300 }, { "epoch": 1.0426347672724485, "grad_norm": 5.506319522857666, "learning_rate": 2.57364844344598e-06, "loss": 14.36246826171875, "step": 3400 }, { "epoch": 1.0733072617130588, "grad_norm": 5.29400634765625, "learning_rate": 2.4455606204608764e-06, "loss": 14.3725537109375, "step": 3500 }, { "epoch": 1.1039797561536693, "grad_norm": 5.597475528717041, "learning_rate": 2.3176157050516264e-06, "loss": 14.357423095703124, "step": 3600 }, { "epoch": 1.1346522505942795, "grad_norm": 5.25772762298584, "learning_rate": 2.1901495625028606e-06, "loss": 14.315875244140624, "step": 3700 }, { "epoch": 1.16532474503489, "grad_norm": 5.585110187530518, "learning_rate": 2.063496801283472e-06, "loss": 14.327174072265626, "step": 3800 }, { "epoch": 1.1959972394755003, "grad_norm": 5.402249813079834, "learning_rate": 1.9379898946735452e-06, "loss": 14.34656494140625, "step": 3900 }, { "epoch": 1.2266697339161108, "grad_norm": 5.390650749206543, "learning_rate": 1.8139583079963144e-06, "loss": 14.36371826171875, "step": 4000 }, { "epoch": 1.257342228356721, "grad_norm": 5.242497444152832, "learning_rate": 1.6917276337462466e-06, "loss": 14.323082275390625, "step": 4100 }, { "epoch": 1.2880147227973315, "grad_norm": 5.234506130218506, "learning_rate": 1.5716187368835971e-06, "loss": 14.337457275390625, "step": 4200 }, { "epoch": 1.3186872172379418, "grad_norm": 5.626885890960693, "learning_rate": 1.4539469125391031e-06, "loss": 14.354864501953125, "step": 4300 }, { "epoch": 1.3493597116785523, "grad_norm": 5.311978816986084, "learning_rate": 1.3390210583399215e-06, "loss": 14.32729248046875, "step": 4400 }, { "epoch": 1.3800322061191626, "grad_norm": 5.375194072723389, "learning_rate": 1.2271428635294922e-06, "loss": 14.337626953125, "step": 4500 }, { "epoch": 1.410704700559773, "grad_norm": 5.57642936706543, "learning_rate": 1.118606017009978e-06, "loss": 14.334080810546874, "step": 4600 }, { "epoch": 1.4413771950003835, "grad_norm": 5.510963439941406, "learning_rate": 1.0136954363862126e-06, "loss": 14.33389404296875, "step": 4700 }, { "epoch": 1.4720496894409938, "grad_norm": 5.4500041007995605, "learning_rate": 9.126865200349847e-07, "loss": 14.31758056640625, "step": 4800 }, { "epoch": 1.502722183881604, "grad_norm": 5.399460792541504, "learning_rate": 8.158444241630245e-07, "loss": 14.280606689453125, "step": 4900 }, { "epoch": 1.5333946783222145, "grad_norm": 5.639800071716309, "learning_rate": 7.234233667514868e-07, "loss": 14.342408447265624, "step": 5000 }, { "epoch": 1.564067172762825, "grad_norm": 5.429558277130127, "learning_rate": 6.356659602141116e-07, "loss": 14.314246826171875, "step": 5100 }, { "epoch": 1.5947396672034353, "grad_norm": 5.131410598754883, "learning_rate": 5.528025745209039e-07, "loss": 14.32427978515625, "step": 5200 }, { "epoch": 1.6254121616440456, "grad_norm": 5.480471611022949, "learning_rate": 4.7505073245916774e-07, "loss": 14.32928466796875, "step": 5300 }, { "epoch": 1.656084656084656, "grad_norm": 5.693525791168213, "learning_rate": 4.026145386193914e-07, "loss": 14.34359375, "step": 5400 }, { "epoch": 1.6867571505252665, "grad_norm": 5.723335266113281, "learning_rate": 3.3568414360493075e-07, "loss": 14.31550537109375, "step": 5500 }, { "epoch": 1.7174296449658768, "grad_norm": 5.319007873535156, "learning_rate": 2.7443524487199333e-07, "loss": 14.32018310546875, "step": 5600 }, { "epoch": 1.7481021394064873, "grad_norm": 5.462660312652588, "learning_rate": 2.1902862551024112e-07, "loss": 14.333775634765624, "step": 5700 }, { "epoch": 1.7787746338470978, "grad_norm": 5.3777546882629395, "learning_rate": 1.6960973217476778e-07, "loss": 14.32505126953125, "step": 5800 }, { "epoch": 1.809447128287708, "grad_norm": 5.701692581176758, "learning_rate": 1.2630829327738987e-07, "loss": 14.33883056640625, "step": 5900 }, { "epoch": 1.8401196227283183, "grad_norm": 5.285277843475342, "learning_rate": 8.92379784395514e-08, "loss": 14.3292626953125, "step": 6000 }, { "epoch": 1.8707921171689288, "grad_norm": 5.558449745178223, "learning_rate": 5.849610010078499e-08, "loss": 14.3488134765625, "step": 6100 }, { "epoch": 1.9014646116095393, "grad_norm": 5.662689685821533, "learning_rate": 3.416335806604443e-08, "loss": 14.28878173828125, "step": 6200 }, { "epoch": 1.9321371060501495, "grad_norm": 5.507394790649414, "learning_rate": 1.630362766248256e-08, "loss": 14.3447216796875, "step": 6300 }, { "epoch": 1.9628096004907598, "grad_norm": 5.682081699371338, "learning_rate": 4.963792061784822e-09, "loss": 14.32853515625, "step": 6400 }, { "epoch": 1.9934820949313703, "grad_norm": 5.3612589836120605, "learning_rate": 1.736192082227306e-10, "loss": 14.27859130859375, "step": 6500 }, { "epoch": 2.0, "eval_loss": 3.4418444633483887, "eval_runtime": 13.1653, "eval_samples_per_second": 640.317, "eval_steps_per_second": 10.026, "step": 6522 } ], "logging_steps": 100, "max_steps": 6522, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 3261, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.393516896094925e+17, "train_batch_size": 64, "trial_name": null, "trial_params": null }