{ "best_global_step": 1962, "best_metric": 3.396049737930298, "best_model_checkpoint": "sindhibert_session5/checkpoint-1962", "epoch": 1.0, "eval_steps": 1962, "global_step": 1962, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.05098139179199592, "grad_norm": 5.245308876037598, "learning_rate": 2.1063829787234044e-06, "loss": 14.123677978515625, "step": 100 }, { "epoch": 0.10196278358399184, "grad_norm": 5.646098613739014, "learning_rate": 4.23404255319149e-06, "loss": 14.066470947265625, "step": 200 }, { "epoch": 0.15294417537598776, "grad_norm": 5.477426528930664, "learning_rate": 4.99628768445363e-06, "loss": 14.01585693359375, "step": 300 }, { "epoch": 0.20392556716798368, "grad_norm": 5.2961344718933105, "learning_rate": 4.975657001043815e-06, "loss": 13.965966796875, "step": 400 }, { "epoch": 0.2549069589599796, "grad_norm": 5.533503532409668, "learning_rate": 4.937082682936824e-06, "loss": 13.97388671875, "step": 500 }, { "epoch": 0.3058883507519755, "grad_norm": 5.314619064331055, "learning_rate": 4.8808443179239025e-06, "loss": 13.95988525390625, "step": 600 }, { "epoch": 0.35686974254397147, "grad_norm": 5.3324384689331055, "learning_rate": 4.807349523327375e-06, "loss": 13.953388671875, "step": 700 }, { "epoch": 0.40785113433596737, "grad_norm": 5.467737674713135, "learning_rate": 4.717130991578312e-06, "loss": 13.9411962890625, "step": 800 }, { "epoch": 0.45883252612796327, "grad_norm": 5.45510721206665, "learning_rate": 4.610842629246138e-06, "loss": 13.934332275390625, "step": 900 }, { "epoch": 0.5098139179199592, "grad_norm": 5.6341962814331055, "learning_rate": 4.489254817504595e-06, "loss": 13.926766357421876, "step": 1000 }, { "epoch": 0.5607953097119551, "grad_norm": 5.584266662597656, "learning_rate": 4.353248828386359e-06, "loss": 13.91165283203125, "step": 1100 }, { "epoch": 0.611776701503951, "grad_norm": 5.35775899887085, "learning_rate": 4.203810437297447e-06, "loss": 13.915504150390625, "step": 1200 }, { "epoch": 0.6627580932959469, "grad_norm": 5.549437999725342, "learning_rate": 4.042022778088111e-06, "loss": 13.945462646484375, "step": 1300 }, { "epoch": 0.7137394850879429, "grad_norm": 5.455720901489258, "learning_rate": 3.869058492466847e-06, "loss": 13.89906494140625, "step": 1400 }, { "epoch": 0.7647208768799388, "grad_norm": 5.856232166290283, "learning_rate": 3.6861712306588434e-06, "loss": 13.905211181640626, "step": 1500 }, { "epoch": 0.8157022686719347, "grad_norm": 5.481319904327393, "learning_rate": 3.4946865649123025e-06, "loss": 13.896219482421875, "step": 1600 }, { "epoch": 0.8666836604639306, "grad_norm": 5.278250217437744, "learning_rate": 3.2959923817118296e-06, "loss": 13.906019287109375, "step": 1700 }, { "epoch": 0.9176650522559265, "grad_norm": 5.768893718719482, "learning_rate": 3.091528822336405e-06, "loss": 13.93161376953125, "step": 1800 }, { "epoch": 0.9686464440479226, "grad_norm": 5.366046905517578, "learning_rate": 2.8827778446730764e-06, "loss": 13.906478271484374, "step": 1900 }, { "epoch": 1.0, "eval_loss": 3.396049737930298, "eval_runtime": 7.952, "eval_samples_per_second": 637.827, "eval_steps_per_second": 10.06, "step": 1962 } ], "logging_steps": 100, "max_steps": 3924, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 1962, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 0 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.321661037009623e+17, "train_batch_size": 64, "trial_name": null, "trial_params": null }