{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 105, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.047619047619047616, "grad_norm": 46.73658663770524, "learning_rate": 4.5454545454545455e-06, "loss": 11.3797, "step": 1 }, { "epoch": 0.09523809523809523, "grad_norm": 44.523666214134316, "learning_rate": 9.090909090909091e-06, "loss": 11.5191, "step": 2 }, { "epoch": 0.14285714285714285, "grad_norm": 52.06639933238829, "learning_rate": 1.3636363636363637e-05, "loss": 10.8448, "step": 3 }, { "epoch": 0.19047619047619047, "grad_norm": 69.75134205890501, "learning_rate": 1.8181818181818182e-05, "loss": 9.5655, "step": 4 }, { "epoch": 0.23809523809523808, "grad_norm": 50.314368468329306, "learning_rate": 2.272727272727273e-05, "loss": 4.1848, "step": 5 }, { "epoch": 0.2857142857142857, "grad_norm": 7.357015429387816, "learning_rate": 2.7272727272727273e-05, "loss": 1.9862, "step": 6 }, { "epoch": 0.3333333333333333, "grad_norm": 7.401615020276034, "learning_rate": 3.181818181818182e-05, "loss": 1.9584, "step": 7 }, { "epoch": 0.38095238095238093, "grad_norm": 3.149164302427893, "learning_rate": 3.6363636363636364e-05, "loss": 1.5975, "step": 8 }, { "epoch": 0.42857142857142855, "grad_norm": 1.9054072928164427, "learning_rate": 4.0909090909090915e-05, "loss": 1.4777, "step": 9 }, { "epoch": 0.47619047619047616, "grad_norm": 2.09522071898415, "learning_rate": 4.545454545454546e-05, "loss": 1.4076, "step": 10 }, { "epoch": 0.5238095238095238, "grad_norm": 1.4886426113733002, "learning_rate": 5e-05, "loss": 1.3226, "step": 11 }, { "epoch": 0.5714285714285714, "grad_norm": 1.594161802836668, "learning_rate": 4.946808510638298e-05, "loss": 1.2825, "step": 12 }, { "epoch": 0.6190476190476191, "grad_norm": 1.5132463664090081, "learning_rate": 4.893617021276596e-05, "loss": 1.3821, "step": 13 }, { "epoch": 0.6666666666666666, "grad_norm": 1.4916977417841104, "learning_rate": 4.840425531914894e-05, "loss": 1.2297, "step": 14 }, { "epoch": 0.7142857142857143, "grad_norm": 1.07234396111405, "learning_rate": 4.787234042553192e-05, "loss": 1.2448, "step": 15 }, { "epoch": 0.7619047619047619, "grad_norm": 0.9261313011360983, "learning_rate": 4.734042553191489e-05, "loss": 1.2062, "step": 16 }, { "epoch": 0.8095238095238095, "grad_norm": 1.3648296835839082, "learning_rate": 4.680851063829788e-05, "loss": 1.2428, "step": 17 }, { "epoch": 0.8571428571428571, "grad_norm": 0.9239559683166827, "learning_rate": 4.627659574468085e-05, "loss": 1.2616, "step": 18 }, { "epoch": 0.9047619047619048, "grad_norm": 0.7209850490291895, "learning_rate": 4.574468085106383e-05, "loss": 1.0656, "step": 19 }, { "epoch": 0.9523809523809523, "grad_norm": 0.9769698482701535, "learning_rate": 4.5212765957446815e-05, "loss": 1.0884, "step": 20 }, { "epoch": 1.0, "grad_norm": 0.6106942050998588, "learning_rate": 4.468085106382979e-05, "loss": 1.0672, "step": 21 }, { "epoch": 1.0476190476190477, "grad_norm": 0.801040183534444, "learning_rate": 4.414893617021277e-05, "loss": 1.0478, "step": 22 }, { "epoch": 1.0952380952380953, "grad_norm": 0.7146659615588544, "learning_rate": 4.3617021276595746e-05, "loss": 1.091, "step": 23 }, { "epoch": 1.1428571428571428, "grad_norm": 0.5905709163842993, "learning_rate": 4.3085106382978725e-05, "loss": 1.1275, "step": 24 }, { "epoch": 1.1904761904761905, "grad_norm": 0.5826060194579513, "learning_rate": 4.2553191489361704e-05, "loss": 1.0079, "step": 25 }, { "epoch": 1.2380952380952381, "grad_norm": 0.6151603075503278, "learning_rate": 4.2021276595744684e-05, "loss": 1.0802, "step": 26 }, { "epoch": 1.2857142857142856, "grad_norm": 0.5264186877249164, "learning_rate": 4.148936170212766e-05, "loss": 1.0406, "step": 27 }, { "epoch": 1.3333333333333333, "grad_norm": 0.5461664948214178, "learning_rate": 4.095744680851064e-05, "loss": 1.0072, "step": 28 }, { "epoch": 1.380952380952381, "grad_norm": 0.4822129939360476, "learning_rate": 4.0425531914893614e-05, "loss": 1.0367, "step": 29 }, { "epoch": 1.4285714285714286, "grad_norm": 0.49848757883873196, "learning_rate": 3.9893617021276594e-05, "loss": 0.9754, "step": 30 }, { "epoch": 1.4761904761904763, "grad_norm": 0.5766625117987682, "learning_rate": 3.936170212765958e-05, "loss": 1.0631, "step": 31 }, { "epoch": 1.5238095238095237, "grad_norm": 0.4859811195017287, "learning_rate": 3.882978723404255e-05, "loss": 1.0003, "step": 32 }, { "epoch": 1.5714285714285714, "grad_norm": 0.5617547786833981, "learning_rate": 3.829787234042553e-05, "loss": 1.0312, "step": 33 }, { "epoch": 1.619047619047619, "grad_norm": 0.4760704205398701, "learning_rate": 3.776595744680852e-05, "loss": 0.9957, "step": 34 }, { "epoch": 1.6666666666666665, "grad_norm": 0.5525097760212518, "learning_rate": 3.723404255319149e-05, "loss": 1.0427, "step": 35 }, { "epoch": 1.7142857142857144, "grad_norm": 0.48722284467868165, "learning_rate": 3.670212765957447e-05, "loss": 0.9141, "step": 36 }, { "epoch": 1.7619047619047619, "grad_norm": 0.48286908931413486, "learning_rate": 3.617021276595745e-05, "loss": 0.9701, "step": 37 }, { "epoch": 1.8095238095238095, "grad_norm": 0.509607069236827, "learning_rate": 3.563829787234043e-05, "loss": 0.9234, "step": 38 }, { "epoch": 1.8571428571428572, "grad_norm": 0.48664273824178367, "learning_rate": 3.5106382978723407e-05, "loss": 1.0605, "step": 39 }, { "epoch": 1.9047619047619047, "grad_norm": 0.4951502177756041, "learning_rate": 3.4574468085106386e-05, "loss": 1.0418, "step": 40 }, { "epoch": 1.9523809523809523, "grad_norm": 0.5599844810380401, "learning_rate": 3.4042553191489365e-05, "loss": 1.0207, "step": 41 }, { "epoch": 2.0, "grad_norm": 0.47167217001219114, "learning_rate": 3.3510638297872344e-05, "loss": 0.8667, "step": 42 }, { "epoch": 2.0476190476190474, "grad_norm": 0.5766076030131434, "learning_rate": 3.2978723404255317e-05, "loss": 0.9305, "step": 43 }, { "epoch": 2.0952380952380953, "grad_norm": 0.5223243597728271, "learning_rate": 3.2446808510638296e-05, "loss": 0.8412, "step": 44 }, { "epoch": 2.142857142857143, "grad_norm": 0.46907426786479, "learning_rate": 3.191489361702128e-05, "loss": 0.8973, "step": 45 }, { "epoch": 2.1904761904761907, "grad_norm": 0.4701126779049834, "learning_rate": 3.1382978723404254e-05, "loss": 0.8509, "step": 46 }, { "epoch": 2.238095238095238, "grad_norm": 0.48418735511330596, "learning_rate": 3.085106382978723e-05, "loss": 0.8757, "step": 47 }, { "epoch": 2.2857142857142856, "grad_norm": 0.6556800657324943, "learning_rate": 3.0319148936170216e-05, "loss": 0.9124, "step": 48 }, { "epoch": 2.3333333333333335, "grad_norm": 0.6014700280019579, "learning_rate": 2.9787234042553192e-05, "loss": 0.8755, "step": 49 }, { "epoch": 2.380952380952381, "grad_norm": 0.5915165889418732, "learning_rate": 2.925531914893617e-05, "loss": 0.8925, "step": 50 }, { "epoch": 2.4285714285714284, "grad_norm": 0.6317703078383352, "learning_rate": 2.8723404255319154e-05, "loss": 0.8663, "step": 51 }, { "epoch": 2.4761904761904763, "grad_norm": 0.6429525648150871, "learning_rate": 2.819148936170213e-05, "loss": 0.8198, "step": 52 }, { "epoch": 2.5238095238095237, "grad_norm": 0.5187394614056622, "learning_rate": 2.765957446808511e-05, "loss": 0.7651, "step": 53 }, { "epoch": 2.571428571428571, "grad_norm": 0.4845593782104512, "learning_rate": 2.7127659574468084e-05, "loss": 0.8805, "step": 54 }, { "epoch": 2.619047619047619, "grad_norm": 0.47561465960794413, "learning_rate": 2.6595744680851064e-05, "loss": 0.9561, "step": 55 }, { "epoch": 2.6666666666666665, "grad_norm": 0.48208401639762166, "learning_rate": 2.6063829787234046e-05, "loss": 0.8638, "step": 56 }, { "epoch": 2.7142857142857144, "grad_norm": 0.5193206572882076, "learning_rate": 2.5531914893617022e-05, "loss": 0.8062, "step": 57 }, { "epoch": 2.761904761904762, "grad_norm": 0.5299016279110512, "learning_rate": 2.5e-05, "loss": 0.8236, "step": 58 }, { "epoch": 2.8095238095238093, "grad_norm": 0.47247578172991544, "learning_rate": 2.446808510638298e-05, "loss": 0.786, "step": 59 }, { "epoch": 2.857142857142857, "grad_norm": 0.4887209838922498, "learning_rate": 2.393617021276596e-05, "loss": 0.8275, "step": 60 }, { "epoch": 2.9047619047619047, "grad_norm": 0.47820835963456026, "learning_rate": 2.340425531914894e-05, "loss": 0.8581, "step": 61 }, { "epoch": 2.9523809523809526, "grad_norm": 0.46592054309550573, "learning_rate": 2.2872340425531915e-05, "loss": 0.8726, "step": 62 }, { "epoch": 3.0, "grad_norm": 0.5063226147267863, "learning_rate": 2.2340425531914894e-05, "loss": 0.8522, "step": 63 }, { "epoch": 3.0476190476190474, "grad_norm": 0.4349456733161818, "learning_rate": 2.1808510638297873e-05, "loss": 0.7678, "step": 64 }, { "epoch": 3.0952380952380953, "grad_norm": 0.43322041549114737, "learning_rate": 2.1276595744680852e-05, "loss": 0.7291, "step": 65 }, { "epoch": 3.142857142857143, "grad_norm": 0.3891658154700991, "learning_rate": 2.074468085106383e-05, "loss": 0.7098, "step": 66 }, { "epoch": 3.1904761904761907, "grad_norm": 0.48090626334768466, "learning_rate": 2.0212765957446807e-05, "loss": 0.7075, "step": 67 }, { "epoch": 3.238095238095238, "grad_norm": 0.4192109123095055, "learning_rate": 1.968085106382979e-05, "loss": 0.7135, "step": 68 }, { "epoch": 3.2857142857142856, "grad_norm": 0.4538447830174013, "learning_rate": 1.9148936170212766e-05, "loss": 0.7293, "step": 69 }, { "epoch": 3.3333333333333335, "grad_norm": 0.400428247237971, "learning_rate": 1.8617021276595745e-05, "loss": 0.7182, "step": 70 }, { "epoch": 3.380952380952381, "grad_norm": 0.41259082028208505, "learning_rate": 1.8085106382978724e-05, "loss": 0.7111, "step": 71 }, { "epoch": 3.4285714285714284, "grad_norm": 0.42210992879504494, "learning_rate": 1.7553191489361703e-05, "loss": 0.7358, "step": 72 }, { "epoch": 3.4761904761904763, "grad_norm": 0.4805171232293903, "learning_rate": 1.7021276595744682e-05, "loss": 0.7693, "step": 73 }, { "epoch": 3.5238095238095237, "grad_norm": 0.4290739751199803, "learning_rate": 1.6489361702127658e-05, "loss": 0.7538, "step": 74 }, { "epoch": 3.571428571428571, "grad_norm": 0.46386967645771476, "learning_rate": 1.595744680851064e-05, "loss": 0.7483, "step": 75 }, { "epoch": 3.619047619047619, "grad_norm": 0.4600000815160992, "learning_rate": 1.5425531914893617e-05, "loss": 0.6961, "step": 76 }, { "epoch": 3.6666666666666665, "grad_norm": 0.48967690124437346, "learning_rate": 1.4893617021276596e-05, "loss": 0.7759, "step": 77 }, { "epoch": 3.7142857142857144, "grad_norm": 0.43799987175594934, "learning_rate": 1.4361702127659577e-05, "loss": 0.6973, "step": 78 }, { "epoch": 3.761904761904762, "grad_norm": 0.4170501746303817, "learning_rate": 1.3829787234042554e-05, "loss": 0.7713, "step": 79 }, { "epoch": 3.8095238095238093, "grad_norm": 0.44313691170789904, "learning_rate": 1.3297872340425532e-05, "loss": 0.7287, "step": 80 }, { "epoch": 3.857142857142857, "grad_norm": 0.3626173621626196, "learning_rate": 1.2765957446808511e-05, "loss": 0.7152, "step": 81 }, { "epoch": 3.9047619047619047, "grad_norm": 0.40980441661962524, "learning_rate": 1.223404255319149e-05, "loss": 0.7388, "step": 82 }, { "epoch": 3.9523809523809526, "grad_norm": 0.3476970829860624, "learning_rate": 1.170212765957447e-05, "loss": 0.6835, "step": 83 }, { "epoch": 4.0, "grad_norm": 0.34534853622038764, "learning_rate": 1.1170212765957447e-05, "loss": 0.6936, "step": 84 }, { "epoch": 4.0476190476190474, "grad_norm": 0.34717149811141956, "learning_rate": 1.0638297872340426e-05, "loss": 0.6412, "step": 85 }, { "epoch": 4.095238095238095, "grad_norm": 0.39121639106624856, "learning_rate": 1.0106382978723404e-05, "loss": 0.6447, "step": 86 }, { "epoch": 4.142857142857143, "grad_norm": 0.3853359648131134, "learning_rate": 9.574468085106383e-06, "loss": 0.67, "step": 87 }, { "epoch": 4.190476190476191, "grad_norm": 0.333219732679469, "learning_rate": 9.042553191489362e-06, "loss": 0.6951, "step": 88 }, { "epoch": 4.238095238095238, "grad_norm": 0.4113440938373367, "learning_rate": 8.510638297872341e-06, "loss": 0.6427, "step": 89 }, { "epoch": 4.285714285714286, "grad_norm": 0.4292570295550588, "learning_rate": 7.97872340425532e-06, "loss": 0.6563, "step": 90 }, { "epoch": 4.333333333333333, "grad_norm": 0.3325469730092871, "learning_rate": 7.446808510638298e-06, "loss": 0.6215, "step": 91 }, { "epoch": 4.380952380952381, "grad_norm": 0.37390136106392785, "learning_rate": 6.914893617021277e-06, "loss": 0.6194, "step": 92 }, { "epoch": 4.428571428571429, "grad_norm": 0.4143640454410716, "learning_rate": 6.3829787234042555e-06, "loss": 0.6207, "step": 93 }, { "epoch": 4.476190476190476, "grad_norm": 0.3648244167487935, "learning_rate": 5.851063829787235e-06, "loss": 0.5998, "step": 94 }, { "epoch": 4.523809523809524, "grad_norm": 0.3512446333779434, "learning_rate": 5.319148936170213e-06, "loss": 0.6351, "step": 95 }, { "epoch": 4.571428571428571, "grad_norm": 0.339436273470992, "learning_rate": 4.787234042553191e-06, "loss": 0.599, "step": 96 }, { "epoch": 4.619047619047619, "grad_norm": 0.4440038854080703, "learning_rate": 4.255319148936171e-06, "loss": 0.584, "step": 97 }, { "epoch": 4.666666666666667, "grad_norm": 0.36581621694617583, "learning_rate": 3.723404255319149e-06, "loss": 0.6203, "step": 98 }, { "epoch": 4.714285714285714, "grad_norm": 0.3509331585886843, "learning_rate": 3.1914893617021277e-06, "loss": 0.5981, "step": 99 }, { "epoch": 4.761904761904762, "grad_norm": 0.3403976301106126, "learning_rate": 2.6595744680851065e-06, "loss": 0.6415, "step": 100 }, { "epoch": 4.809523809523809, "grad_norm": 0.30854856192103863, "learning_rate": 2.1276595744680853e-06, "loss": 0.6054, "step": 101 }, { "epoch": 4.857142857142857, "grad_norm": 0.2984920819758894, "learning_rate": 1.5957446808510639e-06, "loss": 0.6145, "step": 102 }, { "epoch": 4.904761904761905, "grad_norm": 0.30989076720361786, "learning_rate": 1.0638297872340427e-06, "loss": 0.6317, "step": 103 }, { "epoch": 4.9523809523809526, "grad_norm": 0.3438826262811614, "learning_rate": 5.319148936170213e-07, "loss": 0.6193, "step": 104 }, { "epoch": 5.0, "grad_norm": 0.32644346890582215, "learning_rate": 0.0, "loss": 0.6277, "step": 105 }, { "epoch": 5.0, "step": 105, "total_flos": 8.363833783064986e+16, "train_loss": 1.3066247133981614, "train_runtime": 6202.6685, "train_samples_per_second": 0.266, "train_steps_per_second": 0.017 } ], "logging_steps": 1, "max_steps": 105, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 8.363833783064986e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }