{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500.0, "global_step": 619, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0016155088852988692, "grad_norm": 17.625, "learning_rate": 1.0526315789473685e-06, "loss": 1.1847093105316162, "step": 1, "token_acc": 0.7405614714424008 }, { "epoch": 0.01615508885298869, "grad_norm": 7.3125, "learning_rate": 1.0526315789473684e-05, "loss": 0.9536339441935221, "step": 10, "token_acc": 0.7746876697447039 }, { "epoch": 0.03231017770597738, "grad_norm": 5.21875, "learning_rate": 1.999986292247427e-05, "loss": 0.7083949089050293, "step": 20, "token_acc": 0.817535314174379 }, { "epoch": 0.048465266558966075, "grad_norm": 5.09375, "learning_rate": 1.9983418166140286e-05, "loss": 0.6189894676208496, "step": 30, "token_acc": 0.8316753668866345 }, { "epoch": 0.06462035541195477, "grad_norm": 5.25, "learning_rate": 1.99396095545518e-05, "loss": 0.595418643951416, "step": 40, "token_acc": 0.8332687463669831 }, { "epoch": 0.08077544426494346, "grad_norm": 4.96875, "learning_rate": 1.9868557164068073e-05, "loss": 0.5784546852111816, "step": 50, "token_acc": 0.8330070477682068 }, { "epoch": 0.09693053311793215, "grad_norm": 4.78125, "learning_rate": 1.977045574435264e-05, "loss": 0.5632569313049316, "step": 60, "token_acc": 0.8422178988326848 }, { "epoch": 0.11308562197092084, "grad_norm": 4.53125, "learning_rate": 1.9645574184577982e-05, "loss": 0.5411516189575195, "step": 70, "token_acc": 0.8414407360281884 }, { "epoch": 0.12924071082390953, "grad_norm": 5.125, "learning_rate": 1.949425477641904e-05, "loss": 0.5536646366119384, "step": 80, "token_acc": 0.8417469117790098 }, { "epoch": 0.14539579967689822, "grad_norm": 5.375, "learning_rate": 1.931691227585549e-05, "loss": 0.552952766418457, "step": 90, "token_acc": 0.8431372549019608 }, { "epoch": 0.16155088852988692, "grad_norm": 4.875, "learning_rate": 1.9114032766354453e-05, "loss": 0.5462624549865722, "step": 100, "token_acc": 0.843832148768377 }, { "epoch": 0.1777059773828756, "grad_norm": 4.4375, "learning_rate": 1.888617232654949e-05, "loss": 0.5441291809082032, "step": 110, "token_acc": 0.8419827335338054 }, { "epoch": 0.1938610662358643, "grad_norm": 4.3125, "learning_rate": 1.8633955506067717e-05, "loss": 0.5480209350585937, "step": 120, "token_acc": 0.8493486139680674 }, { "epoch": 0.210016155088853, "grad_norm": 5.28125, "learning_rate": 1.8358073613682705e-05, "loss": 0.553197193145752, "step": 130, "token_acc": 0.8403582205782147 }, { "epoch": 0.22617124394184168, "grad_norm": 5.4375, "learning_rate": 1.805928282248516e-05, "loss": 0.558505630493164, "step": 140, "token_acc": 0.8416950803701899 }, { "epoch": 0.24232633279483037, "grad_norm": 4.5625, "learning_rate": 1.7738402097265063e-05, "loss": 0.5198162078857422, "step": 150, "token_acc": 0.8472963789923308 }, { "epoch": 0.25848142164781907, "grad_norm": 4.875, "learning_rate": 1.73963109497861e-05, "loss": 0.5393078804016114, "step": 160, "token_acc": 0.844424985405721 }, { "epoch": 0.27463651050080773, "grad_norm": 4.90625, "learning_rate": 1.703394702810504e-05, "loss": 0.5661784172058105, "step": 170, "token_acc": 0.8364490554957424 }, { "epoch": 0.29079159935379645, "grad_norm": 4.5625, "learning_rate": 1.665230354654361e-05, "loss": 0.5104843139648437, "step": 180, "token_acc": 0.8500578927055191 }, { "epoch": 0.3069466882067851, "grad_norm": 4.71875, "learning_rate": 1.6252426563357054e-05, "loss": 0.5394854545593262, "step": 190, "token_acc": 0.8471082454083626 }, { "epoch": 0.32310177705977383, "grad_norm": 4.46875, "learning_rate": 1.5835412113561176e-05, "loss": 0.5530763149261475, "step": 200, "token_acc": 0.8418517790446236 }, { "epoch": 0.3392568659127625, "grad_norm": 4.40625, "learning_rate": 1.5402403204776552e-05, "loss": 0.5297245979309082, "step": 210, "token_acc": 0.8469646691391762 }, { "epoch": 0.3554119547657512, "grad_norm": 4.75, "learning_rate": 1.4954586684324077e-05, "loss": 0.5240192413330078, "step": 220, "token_acc": 0.8480382588327152 }, { "epoch": 0.3715670436187399, "grad_norm": 4.625, "learning_rate": 1.4493189986158966e-05, "loss": 0.5166411399841309, "step": 230, "token_acc": 0.8447706065318819 }, { "epoch": 0.3877221324717286, "grad_norm": 4.375, "learning_rate": 1.4019477766559604e-05, "loss": 0.5292513847351075, "step": 240, "token_acc": 0.8480617127233668 }, { "epoch": 0.40387722132471726, "grad_norm": 5.0, "learning_rate": 1.3534748437792573e-05, "loss": 0.5284510135650635, "step": 250, "token_acc": 0.8485742379547689 }, { "epoch": 0.420032310177706, "grad_norm": 4.28125, "learning_rate": 1.3040330609254903e-05, "loss": 0.5196167945861816, "step": 260, "token_acc": 0.8470174753490188 }, { "epoch": 0.43618739903069464, "grad_norm": 4.9375, "learning_rate": 1.2537579445848058e-05, "loss": 0.5236739635467529, "step": 270, "token_acc": 0.8470025385666862 }, { "epoch": 0.45234248788368336, "grad_norm": 4.625, "learning_rate": 1.2027872953565125e-05, "loss": 0.5147537231445313, "step": 280, "token_acc": 0.8483849063924725 }, { "epoch": 0.46849757673667203, "grad_norm": 4.53125, "learning_rate": 1.1512608202472195e-05, "loss": 0.5182638168334961, "step": 290, "token_acc": 0.8484908071969325 }, { "epoch": 0.48465266558966075, "grad_norm": 4.5625, "learning_rate": 1.0993197497436392e-05, "loss": 0.5263832092285157, "step": 300, "token_acc": 0.848084030344291 }, { "epoch": 0.5008077544426495, "grad_norm": 4.28125, "learning_rate": 1.0471064507096427e-05, "loss": 0.5192934036254883, "step": 310, "token_acc": 0.8522849069472863 }, { "epoch": 0.5169628432956381, "grad_norm": 4.46875, "learning_rate": 9.947640361685805e-06, "loss": 0.49528074264526367, "step": 320, "token_acc": 0.853359489946052 }, { "epoch": 0.5331179321486268, "grad_norm": 4.625, "learning_rate": 9.424359730404329e-06, "loss": 0.5119996070861816, "step": 330, "token_acc": 0.8492686234621718 }, { "epoch": 0.5492730210016155, "grad_norm": 4.5625, "learning_rate": 8.902656889089548e-06, "loss": 0.5261263847351074, "step": 340, "token_acc": 0.8463559404495475 }, { "epoch": 0.5654281098546042, "grad_norm": 4.375, "learning_rate": 8.38396178896639e-06, "loss": 0.5155754566192627, "step": 350, "token_acc": 0.853113410111263 }, { "epoch": 0.5815831987075929, "grad_norm": 4.59375, "learning_rate": 7.869696137250235e-06, "loss": 0.5171246528625488, "step": 360, "token_acc": 0.8516091166976426 }, { "epoch": 0.5977382875605816, "grad_norm": 4.6875, "learning_rate": 7.361269500346274e-06, "loss": 0.5315141677856445, "step": 370, "token_acc": 0.8478004713275726 }, { "epoch": 0.6138933764135702, "grad_norm": 5.0625, "learning_rate": 6.860075440325951e-06, "loss": 0.5140278816223145, "step": 380, "token_acc": 0.8488246287006983 }, { "epoch": 0.630048465266559, "grad_norm": 4.78125, "learning_rate": 6.367487695270218e-06, "loss": 0.5044659614562989, "step": 390, "token_acc": 0.8520192591136877 }, { "epoch": 0.6462035541195477, "grad_norm": 5.0, "learning_rate": 5.884856413948913e-06, "loss": 0.5070418357849121, "step": 400, "token_acc": 0.8493284017909286 }, { "epoch": 0.6623586429725363, "grad_norm": 4.59375, "learning_rate": 5.413504455156855e-06, "loss": 0.5049197196960449, "step": 410, "token_acc": 0.8500828540793449 }, { "epoch": 0.678513731825525, "grad_norm": 4.4375, "learning_rate": 4.954723761849809e-06, "loss": 0.4989957332611084, "step": 420, "token_acc": 0.8527764127764128 }, { "epoch": 0.6946688206785138, "grad_norm": 5.125, "learning_rate": 4.509771820018682e-06, "loss": 0.49854435920715334, "step": 430, "token_acc": 0.8521560574948666 }, { "epoch": 0.7108239095315024, "grad_norm": 4.46875, "learning_rate": 4.0798682120078046e-06, "loss": 0.5228716373443604, "step": 440, "token_acc": 0.8458528951486698 }, { "epoch": 0.7269789983844911, "grad_norm": 5.25, "learning_rate": 3.6661912737244996e-06, "loss": 0.5174414157867432, "step": 450, "token_acc": 0.8437163814180929 }, { "epoch": 0.7431340872374798, "grad_norm": 5.28125, "learning_rate": 3.2698748649022693e-06, "loss": 0.4988471508026123, "step": 460, "token_acc": 0.8513395031660984 }, { "epoch": 0.7592891760904685, "grad_norm": 4.59375, "learning_rate": 2.8920052612700755e-06, "loss": 0.5026087760925293, "step": 470, "token_acc": 0.8537626066718387 }, { "epoch": 0.7754442649434572, "grad_norm": 4.40625, "learning_rate": 2.5336181771460877e-06, "loss": 0.49767255783081055, "step": 480, "token_acc": 0.8517356344381252 }, { "epoch": 0.7915993537964459, "grad_norm": 4.8125, "learning_rate": 2.195695926616702e-06, "loss": 0.5065960884094238, "step": 490, "token_acc": 0.8521747648902821 }, { "epoch": 0.8077544426494345, "grad_norm": 4.4375, "learning_rate": 1.8791647310819371e-06, "loss": 0.5075524330139161, "step": 500, "token_acc": 0.8482975174173291 }, { "epoch": 0.8239095315024233, "grad_norm": 5.25, "learning_rate": 1.5848921805469396e-06, "loss": 0.5103363990783691, "step": 510, "token_acc": 0.8537348455220962 }, { "epoch": 0.840064620355412, "grad_norm": 4.9375, "learning_rate": 1.3136848556180893e-06, "loss": 0.527727460861206, "step": 520, "token_acc": 0.8470945020543925 }, { "epoch": 0.8562197092084006, "grad_norm": 4.375, "learning_rate": 1.0662861167216243e-06, "loss": 0.4919741630554199, "step": 530, "token_acc": 0.8542659791808541 }, { "epoch": 0.8723747980613893, "grad_norm": 6.90625, "learning_rate": 8.433740666043899e-07, "loss": 0.5399604797363281, "step": 540, "token_acc": 0.8415647921760391 }, { "epoch": 0.8885298869143781, "grad_norm": 4.40625, "learning_rate": 6.455596917013274e-07, "loss": 0.5133064270019532, "step": 550, "token_acc": 0.8489405331510594 }, { "epoch": 0.9046849757673667, "grad_norm": 4.875, "learning_rate": 4.733851874641382e-07, "loss": 0.5087760925292969, "step": 560, "token_acc": 0.8505905511811024 }, { "epoch": 0.9208400646203554, "grad_norm": 4.625, "learning_rate": 3.273224722412327e-07, "loss": 0.4963532447814941, "step": 570, "token_acc": 0.8557692307692307 }, { "epoch": 0.9369951534733441, "grad_norm": 4.65625, "learning_rate": 2.077718937823414e-07, "loss": 0.5168875694274903, "step": 580, "token_acc": 0.8515441751368257 }, { "epoch": 0.9531502423263328, "grad_norm": 4.09375, "learning_rate": 1.1506113191316447e-07, "loss": 0.5301695823669433, "step": 590, "token_acc": 0.8447769953051644 }, { "epoch": 0.9693053311793215, "grad_norm": 4.53125, "learning_rate": 4.944430038773762e-08, "loss": 0.507291316986084, "step": 600, "token_acc": 0.8520109332292073 }, { "epoch": 0.9854604200323102, "grad_norm": 4.28125, "learning_rate": 1.1101250380300965e-08, "loss": 0.48858890533447263, "step": 610, "token_acc": 0.8529469355786176 } ], "logging_steps": 10, "max_steps": 619, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.618575333317673e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }