{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 10.0, "eval_steps": 500.0, "global_step": 30630, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.1632519794302506, "grad_norm": 0.42205774784088135, "learning_rate": 0.0001996, "loss": 9.0337958984375, "step": 500 }, { "epoch": 0.3265039588605012, "grad_norm": 0.4745626747608185, "learning_rate": 0.00019668768669100565, "loss": 7.18489794921875, "step": 1000 }, { "epoch": 0.48975593829075176, "grad_norm": 0.5245969295501709, "learning_rate": 0.00019336873547958847, "loss": 6.76750439453125, "step": 1500 }, { "epoch": 0.6530079177210024, "grad_norm": 0.5147088170051575, "learning_rate": 0.00019004978426817126, "loss": 6.49613671875, "step": 2000 }, { "epoch": 0.8162598971512529, "grad_norm": 0.5029264688491821, "learning_rate": 0.00018673083305675408, "loss": 6.30743798828125, "step": 2500 }, { "epoch": 0.9795118765815035, "grad_norm": 0.52044677734375, "learning_rate": 0.00018341188184533688, "loss": 6.20166357421875, "step": 3000 }, { "epoch": 1.142682230022039, "grad_norm": 0.5113121867179871, "learning_rate": 0.0001800929306339197, "loss": 5.97177001953125, "step": 3500 }, { "epoch": 1.3059342094522897, "grad_norm": 0.5511940121650696, "learning_rate": 0.0001767739794225025, "loss": 5.921890625, "step": 4000 }, { "epoch": 1.4691861888825402, "grad_norm": 0.4959582984447479, "learning_rate": 0.0001734550282110853, "loss": 5.78601318359375, "step": 4500 }, { "epoch": 1.6324381683127909, "grad_norm": 0.527199387550354, "learning_rate": 0.0001701360769996681, "loss": 5.76173681640625, "step": 5000 }, { "epoch": 1.7956901477430414, "grad_norm": 0.49316641688346863, "learning_rate": 0.00016681712578825092, "loss": 5.67507861328125, "step": 5500 }, { "epoch": 1.9589421271732919, "grad_norm": 0.4699156880378723, "learning_rate": 0.00016349817457683371, "loss": 5.567580078125, "step": 6000 }, { "epoch": 2.1221124806138274, "grad_norm": 0.5287707448005676, "learning_rate": 0.00016017922336541653, "loss": 5.44504248046875, "step": 6500 }, { "epoch": 2.285364460044078, "grad_norm": 1.1574363708496094, "learning_rate": 0.00015686027215399935, "loss": 5.31701904296875, "step": 7000 }, { "epoch": 2.4486164394743284, "grad_norm": 0.45680922269821167, "learning_rate": 0.00015354132094258215, "loss": 5.35315771484375, "step": 7500 }, { "epoch": 2.6118684189045793, "grad_norm": 0.5091222524642944, "learning_rate": 0.00015022236973116497, "loss": 5.32297998046875, "step": 8000 }, { "epoch": 2.77512039833483, "grad_norm": 0.5736968517303467, "learning_rate": 0.00014690341851974776, "loss": 5.2701904296875, "step": 8500 }, { "epoch": 2.9383723777650803, "grad_norm": 0.5231903791427612, "learning_rate": 0.00014358446730833058, "loss": 5.30401953125, "step": 9000 }, { "epoch": 3.101542731205616, "grad_norm": 0.5694177746772766, "learning_rate": 0.00014026551609691337, "loss": 5.1482998046875, "step": 9500 }, { "epoch": 3.2647947106358663, "grad_norm": 0.5935769081115723, "learning_rate": 0.0001369465648854962, "loss": 5.08368701171875, "step": 10000 }, { "epoch": 3.4280466900661173, "grad_norm": 0.6495661735534668, "learning_rate": 0.000133627613674079, "loss": 5.0398076171875, "step": 10500 }, { "epoch": 3.591298669496368, "grad_norm": 0.5465214252471924, "learning_rate": 0.0001303086624626618, "loss": 5.07168017578125, "step": 11000 }, { "epoch": 3.7545506489266183, "grad_norm": 0.5718339681625366, "learning_rate": 0.00012698971125124463, "loss": 5.0271943359375, "step": 11500 }, { "epoch": 3.9178026283568688, "grad_norm": 0.607941746711731, "learning_rate": 0.00012367076003982742, "loss": 5.05531640625, "step": 12000 }, { "epoch": 4.080972981797404, "grad_norm": 0.5361756682395935, "learning_rate": 0.00012035180882841021, "loss": 4.88087744140625, "step": 12500 }, { "epoch": 4.244224961227655, "grad_norm": 0.5884597301483154, "learning_rate": 0.00011703285761699303, "loss": 4.826615234375, "step": 13000 }, { "epoch": 4.407476940657905, "grad_norm": 0.6183493137359619, "learning_rate": 0.00011371390640557584, "loss": 4.8345087890625, "step": 13500 }, { "epoch": 4.570728920088156, "grad_norm": 0.4895070493221283, "learning_rate": 0.00011039495519415866, "loss": 4.808126953125, "step": 14000 }, { "epoch": 4.733980899518406, "grad_norm": 0.5796904563903809, "learning_rate": 0.00010707600398274147, "loss": 4.8475546875, "step": 14500 }, { "epoch": 4.897232878948657, "grad_norm": 0.5432486534118652, "learning_rate": 0.00010375705277132426, "loss": 4.81570703125, "step": 15000 }, { "epoch": 5.060403232389192, "grad_norm": 0.6771642565727234, "learning_rate": 0.00010043810155990707, "loss": 4.76427978515625, "step": 15500 }, { "epoch": 5.223655211819444, "grad_norm": 0.6445333957672119, "learning_rate": 9.711915034848989e-05, "loss": 4.65180908203125, "step": 16000 }, { "epoch": 5.386907191249694, "grad_norm": 0.599590539932251, "learning_rate": 9.380019913707268e-05, "loss": 4.70791796875, "step": 16500 }, { "epoch": 5.550159170679945, "grad_norm": 0.6182076334953308, "learning_rate": 9.04812479256555e-05, "loss": 4.666, "step": 17000 }, { "epoch": 5.713411150110195, "grad_norm": 0.601693332195282, "learning_rate": 8.71622967142383e-05, "loss": 4.6578359375, "step": 17500 }, { "epoch": 5.876663129540446, "grad_norm": 0.6478536128997803, "learning_rate": 8.384334550282111e-05, "loss": 4.64422509765625, "step": 18000 }, { "epoch": 6.039833482980981, "grad_norm": 0.6848897337913513, "learning_rate": 8.052439429140392e-05, "loss": 4.63602099609375, "step": 18500 }, { "epoch": 6.203085462411232, "grad_norm": 0.599690854549408, "learning_rate": 7.720544307998673e-05, "loss": 4.57735595703125, "step": 19000 }, { "epoch": 6.366337441841482, "grad_norm": 0.567436158657074, "learning_rate": 7.388649186856955e-05, "loss": 4.51408740234375, "step": 19500 }, { "epoch": 6.529589421271733, "grad_norm": 0.9956502914428711, "learning_rate": 7.056754065715234e-05, "loss": 4.52962158203125, "step": 20000 }, { "epoch": 6.692841400701983, "grad_norm": 0.6528770923614502, "learning_rate": 6.724858944573516e-05, "loss": 4.47805078125, "step": 20500 }, { "epoch": 6.856093380132235, "grad_norm": 0.5711560249328613, "learning_rate": 6.392963823431795e-05, "loss": 4.5910791015625, "step": 21000 }, { "epoch": 7.019263733572769, "grad_norm": 0.5683479309082031, "learning_rate": 6.061068702290077e-05, "loss": 4.54874267578125, "step": 21500 }, { "epoch": 7.182515713003021, "grad_norm": 0.6191678643226624, "learning_rate": 5.729173581148357e-05, "loss": 4.418267578125, "step": 22000 }, { "epoch": 7.345767692433271, "grad_norm": 0.6200758814811707, "learning_rate": 5.3972784600066386e-05, "loss": 4.46817431640625, "step": 22500 }, { "epoch": 7.509019671863522, "grad_norm": 0.5810480713844299, "learning_rate": 5.0653833388649185e-05, "loss": 4.4288740234375, "step": 23000 }, { "epoch": 7.672271651293772, "grad_norm": 0.6930559277534485, "learning_rate": 4.7334882177232e-05, "loss": 4.42653369140625, "step": 23500 }, { "epoch": 7.835523630724023, "grad_norm": 0.6396164298057556, "learning_rate": 4.4015930965814805e-05, "loss": 4.49230419921875, "step": 24000 }, { "epoch": 7.998775610154273, "grad_norm": 0.5532464981079102, "learning_rate": 4.069697975439761e-05, "loss": 4.39004931640625, "step": 24500 }, { "epoch": 8.161945963594809, "grad_norm": 0.6885871291160583, "learning_rate": 3.737802854298042e-05, "loss": 4.382765625, "step": 25000 }, { "epoch": 8.32519794302506, "grad_norm": 0.6912867426872253, "learning_rate": 3.4059077331563225e-05, "loss": 4.396109375, "step": 25500 }, { "epoch": 8.48844992245531, "grad_norm": 0.6827392578125, "learning_rate": 3.074012612014603e-05, "loss": 4.40128125, "step": 26000 }, { "epoch": 8.651701901885561, "grad_norm": 0.6782070398330688, "learning_rate": 2.742117490872884e-05, "loss": 4.35358203125, "step": 26500 }, { "epoch": 8.81495388131581, "grad_norm": 0.7143053412437439, "learning_rate": 2.410222369731165e-05, "loss": 4.339625, "step": 27000 }, { "epoch": 8.978205860746062, "grad_norm": 0.684012234210968, "learning_rate": 2.0783272485894458e-05, "loss": 4.3393515625, "step": 27500 }, { "epoch": 9.141376214186597, "grad_norm": 0.6413611769676208, "learning_rate": 1.7464321274477265e-05, "loss": 4.32941748046875, "step": 28000 }, { "epoch": 9.304628193616848, "grad_norm": 0.6120012402534485, "learning_rate": 1.4145370063060073e-05, "loss": 4.31831884765625, "step": 28500 }, { "epoch": 9.467880173047098, "grad_norm": 0.6299709677696228, "learning_rate": 1.0826418851642881e-05, "loss": 4.28150341796875, "step": 29000 }, { "epoch": 9.631132152477349, "grad_norm": 0.6235489845275879, "learning_rate": 7.5074676402256894e-06, "loss": 4.31871484375, "step": 29500 }, { "epoch": 9.794384131907599, "grad_norm": 0.7088222503662109, "learning_rate": 4.188516428808497e-06, "loss": 4.35441064453125, "step": 30000 }, { "epoch": 9.95763611133785, "grad_norm": 0.636341392993927, "learning_rate": 8.695652173913044e-07, "loss": 4.2993779296875, "step": 30500 }, { "epoch": 10.0, "step": 30630, "total_flos": 1.2426966463650202e+17, "train_loss": 5.026907345643823, "train_runtime": 19932.1348, "train_samples_per_second": 49.168, "train_steps_per_second": 1.537 } ], "logging_steps": 500, "max_steps": 30630, "num_input_tokens_seen": 0, "num_train_epochs": 10, "save_steps": 2000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.2426966463650202e+17, "train_batch_size": 8, "trial_name": null, "trial_params": null }