| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 5.0, | |
| "eval_steps": 500, | |
| "global_step": 45500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.054945054945054944, | |
| "grad_norm": 1.780437707901001, | |
| "learning_rate": 0.0001978021978021978, | |
| "loss": 1.5906, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.10989010989010989, | |
| "grad_norm": 2.4972989559173584, | |
| "learning_rate": 0.00019560439560439562, | |
| "loss": 1.5616, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.16483516483516483, | |
| "grad_norm": 1.6102287769317627, | |
| "learning_rate": 0.00019340659340659342, | |
| "loss": 1.5549, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.21978021978021978, | |
| "grad_norm": 1.6323072910308838, | |
| "learning_rate": 0.00019120879120879122, | |
| "loss": 1.5446, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.27472527472527475, | |
| "grad_norm": 1.6228586435317993, | |
| "learning_rate": 0.00018901098901098903, | |
| "loss": 1.5549, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.32967032967032966, | |
| "grad_norm": 1.433160662651062, | |
| "learning_rate": 0.00018681318681318683, | |
| "loss": 1.5521, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.38461538461538464, | |
| "grad_norm": 1.4724050760269165, | |
| "learning_rate": 0.00018461538461538463, | |
| "loss": 1.5335, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.43956043956043955, | |
| "grad_norm": 1.2267966270446777, | |
| "learning_rate": 0.0001824175824175824, | |
| "loss": 1.5359, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.4945054945054945, | |
| "grad_norm": 1.3739681243896484, | |
| "learning_rate": 0.00018021978021978024, | |
| "loss": 1.5362, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.5494505494505495, | |
| "grad_norm": 1.3636854887008667, | |
| "learning_rate": 0.00017802197802197802, | |
| "loss": 1.5257, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.6043956043956044, | |
| "grad_norm": 1.9858155250549316, | |
| "learning_rate": 0.00017582417582417582, | |
| "loss": 1.5378, | |
| "step": 5500 | |
| }, | |
| { | |
| "epoch": 0.6593406593406593, | |
| "grad_norm": 1.6531416177749634, | |
| "learning_rate": 0.00017362637362637365, | |
| "loss": 1.5239, | |
| "step": 6000 | |
| }, | |
| { | |
| "epoch": 0.7142857142857143, | |
| "grad_norm": 1.638098120689392, | |
| "learning_rate": 0.00017142857142857143, | |
| "loss": 1.5331, | |
| "step": 6500 | |
| }, | |
| { | |
| "epoch": 0.7692307692307693, | |
| "grad_norm": 1.8147164583206177, | |
| "learning_rate": 0.00016923076923076923, | |
| "loss": 1.5291, | |
| "step": 7000 | |
| }, | |
| { | |
| "epoch": 0.8241758241758241, | |
| "grad_norm": 1.4330575466156006, | |
| "learning_rate": 0.00016703296703296706, | |
| "loss": 1.5328, | |
| "step": 7500 | |
| }, | |
| { | |
| "epoch": 0.8791208791208791, | |
| "grad_norm": 1.3865350484848022, | |
| "learning_rate": 0.00016483516483516484, | |
| "loss": 1.5266, | |
| "step": 8000 | |
| }, | |
| { | |
| "epoch": 0.9340659340659341, | |
| "grad_norm": 1.8921843767166138, | |
| "learning_rate": 0.00016263736263736264, | |
| "loss": 1.5323, | |
| "step": 8500 | |
| }, | |
| { | |
| "epoch": 0.989010989010989, | |
| "grad_norm": 2.120455026626587, | |
| "learning_rate": 0.00016043956043956044, | |
| "loss": 1.5254, | |
| "step": 9000 | |
| }, | |
| { | |
| "epoch": 1.043956043956044, | |
| "grad_norm": 1.6023536920547485, | |
| "learning_rate": 0.00015824175824175824, | |
| "loss": 1.4525, | |
| "step": 9500 | |
| }, | |
| { | |
| "epoch": 1.098901098901099, | |
| "grad_norm": 1.4707773923873901, | |
| "learning_rate": 0.00015604395604395605, | |
| "loss": 1.4538, | |
| "step": 10000 | |
| }, | |
| { | |
| "epoch": 1.1538461538461537, | |
| "grad_norm": 1.7984991073608398, | |
| "learning_rate": 0.00015384615384615385, | |
| "loss": 1.4421, | |
| "step": 10500 | |
| }, | |
| { | |
| "epoch": 1.2087912087912087, | |
| "grad_norm": 1.70187509059906, | |
| "learning_rate": 0.00015164835164835165, | |
| "loss": 1.4454, | |
| "step": 11000 | |
| }, | |
| { | |
| "epoch": 1.2637362637362637, | |
| "grad_norm": 2.5541532039642334, | |
| "learning_rate": 0.00014945054945054946, | |
| "loss": 1.4512, | |
| "step": 11500 | |
| }, | |
| { | |
| "epoch": 1.3186813186813187, | |
| "grad_norm": 1.5920054912567139, | |
| "learning_rate": 0.00014725274725274726, | |
| "loss": 1.4557, | |
| "step": 12000 | |
| }, | |
| { | |
| "epoch": 1.3736263736263736, | |
| "grad_norm": 2.0435898303985596, | |
| "learning_rate": 0.00014505494505494506, | |
| "loss": 1.4496, | |
| "step": 12500 | |
| }, | |
| { | |
| "epoch": 1.4285714285714286, | |
| "grad_norm": 2.3308217525482178, | |
| "learning_rate": 0.00014285714285714287, | |
| "loss": 1.437, | |
| "step": 13000 | |
| }, | |
| { | |
| "epoch": 1.4835164835164836, | |
| "grad_norm": 2.7433712482452393, | |
| "learning_rate": 0.00014065934065934067, | |
| "loss": 1.437, | |
| "step": 13500 | |
| }, | |
| { | |
| "epoch": 1.5384615384615383, | |
| "grad_norm": 2.6499412059783936, | |
| "learning_rate": 0.00013846153846153847, | |
| "loss": 1.4419, | |
| "step": 14000 | |
| }, | |
| { | |
| "epoch": 1.5934065934065935, | |
| "grad_norm": 2.6177027225494385, | |
| "learning_rate": 0.00013626373626373628, | |
| "loss": 1.4408, | |
| "step": 14500 | |
| }, | |
| { | |
| "epoch": 1.6483516483516483, | |
| "grad_norm": 1.9484443664550781, | |
| "learning_rate": 0.00013406593406593405, | |
| "loss": 1.4617, | |
| "step": 15000 | |
| }, | |
| { | |
| "epoch": 1.7032967032967035, | |
| "grad_norm": 2.649055242538452, | |
| "learning_rate": 0.00013186813186813188, | |
| "loss": 1.4482, | |
| "step": 15500 | |
| }, | |
| { | |
| "epoch": 1.7582417582417582, | |
| "grad_norm": 1.9473294019699097, | |
| "learning_rate": 0.0001296703296703297, | |
| "loss": 1.4358, | |
| "step": 16000 | |
| }, | |
| { | |
| "epoch": 1.8131868131868132, | |
| "grad_norm": 2.2678744792938232, | |
| "learning_rate": 0.00012747252747252746, | |
| "loss": 1.4333, | |
| "step": 16500 | |
| }, | |
| { | |
| "epoch": 1.8681318681318682, | |
| "grad_norm": 1.629854440689087, | |
| "learning_rate": 0.00012527472527472527, | |
| "loss": 1.4305, | |
| "step": 17000 | |
| }, | |
| { | |
| "epoch": 1.9230769230769231, | |
| "grad_norm": 2.318514823913574, | |
| "learning_rate": 0.0001230769230769231, | |
| "loss": 1.4292, | |
| "step": 17500 | |
| }, | |
| { | |
| "epoch": 1.978021978021978, | |
| "grad_norm": 2.1276426315307617, | |
| "learning_rate": 0.00012087912087912087, | |
| "loss": 1.435, | |
| "step": 18000 | |
| }, | |
| { | |
| "epoch": 2.032967032967033, | |
| "grad_norm": 2.3333852291107178, | |
| "learning_rate": 0.00011868131868131869, | |
| "loss": 1.3761, | |
| "step": 18500 | |
| }, | |
| { | |
| "epoch": 2.087912087912088, | |
| "grad_norm": 3.305957078933716, | |
| "learning_rate": 0.0001164835164835165, | |
| "loss": 1.3318, | |
| "step": 19000 | |
| }, | |
| { | |
| "epoch": 2.142857142857143, | |
| "grad_norm": 5.382218837738037, | |
| "learning_rate": 0.00011428571428571428, | |
| "loss": 1.3463, | |
| "step": 19500 | |
| }, | |
| { | |
| "epoch": 2.197802197802198, | |
| "grad_norm": 3.35638427734375, | |
| "learning_rate": 0.0001120879120879121, | |
| "loss": 1.3418, | |
| "step": 20000 | |
| }, | |
| { | |
| "epoch": 2.2527472527472527, | |
| "grad_norm": 3.04825496673584, | |
| "learning_rate": 0.0001098901098901099, | |
| "loss": 1.3568, | |
| "step": 20500 | |
| }, | |
| { | |
| "epoch": 2.3076923076923075, | |
| "grad_norm": 2.9136993885040283, | |
| "learning_rate": 0.0001076923076923077, | |
| "loss": 1.3507, | |
| "step": 21000 | |
| }, | |
| { | |
| "epoch": 2.3626373626373627, | |
| "grad_norm": 2.5451395511627197, | |
| "learning_rate": 0.0001054945054945055, | |
| "loss": 1.353, | |
| "step": 21500 | |
| }, | |
| { | |
| "epoch": 2.4175824175824174, | |
| "grad_norm": 2.779684066772461, | |
| "learning_rate": 0.00010329670329670331, | |
| "loss": 1.3456, | |
| "step": 22000 | |
| }, | |
| { | |
| "epoch": 2.4725274725274726, | |
| "grad_norm": 3.8211276531219482, | |
| "learning_rate": 0.0001010989010989011, | |
| "loss": 1.3519, | |
| "step": 22500 | |
| }, | |
| { | |
| "epoch": 2.5274725274725274, | |
| "grad_norm": 2.877223253250122, | |
| "learning_rate": 9.89010989010989e-05, | |
| "loss": 1.3506, | |
| "step": 23000 | |
| }, | |
| { | |
| "epoch": 2.5824175824175826, | |
| "grad_norm": 2.53200101852417, | |
| "learning_rate": 9.670329670329671e-05, | |
| "loss": 1.3564, | |
| "step": 23500 | |
| }, | |
| { | |
| "epoch": 2.6373626373626373, | |
| "grad_norm": 3.249028205871582, | |
| "learning_rate": 9.450549450549451e-05, | |
| "loss": 1.3393, | |
| "step": 24000 | |
| }, | |
| { | |
| "epoch": 2.6923076923076925, | |
| "grad_norm": 2.819939613342285, | |
| "learning_rate": 9.230769230769232e-05, | |
| "loss": 1.3413, | |
| "step": 24500 | |
| }, | |
| { | |
| "epoch": 2.7472527472527473, | |
| "grad_norm": 3.4391469955444336, | |
| "learning_rate": 9.010989010989012e-05, | |
| "loss": 1.3517, | |
| "step": 25000 | |
| }, | |
| { | |
| "epoch": 2.802197802197802, | |
| "grad_norm": 2.9319329261779785, | |
| "learning_rate": 8.791208791208791e-05, | |
| "loss": 1.3362, | |
| "step": 25500 | |
| }, | |
| { | |
| "epoch": 2.857142857142857, | |
| "grad_norm": 2.257335901260376, | |
| "learning_rate": 8.571428571428571e-05, | |
| "loss": 1.3402, | |
| "step": 26000 | |
| }, | |
| { | |
| "epoch": 2.912087912087912, | |
| "grad_norm": 2.6121294498443604, | |
| "learning_rate": 8.351648351648353e-05, | |
| "loss": 1.3615, | |
| "step": 26500 | |
| }, | |
| { | |
| "epoch": 2.967032967032967, | |
| "grad_norm": 2.716127395629883, | |
| "learning_rate": 8.131868131868132e-05, | |
| "loss": 1.3418, | |
| "step": 27000 | |
| }, | |
| { | |
| "epoch": 3.021978021978022, | |
| "grad_norm": 2.227154493331909, | |
| "learning_rate": 7.912087912087912e-05, | |
| "loss": 1.3199, | |
| "step": 27500 | |
| }, | |
| { | |
| "epoch": 3.076923076923077, | |
| "grad_norm": 3.0606658458709717, | |
| "learning_rate": 7.692307692307693e-05, | |
| "loss": 1.2553, | |
| "step": 28000 | |
| }, | |
| { | |
| "epoch": 3.131868131868132, | |
| "grad_norm": 2.8261947631835938, | |
| "learning_rate": 7.472527472527473e-05, | |
| "loss": 1.2688, | |
| "step": 28500 | |
| }, | |
| { | |
| "epoch": 3.186813186813187, | |
| "grad_norm": 3.738393783569336, | |
| "learning_rate": 7.252747252747253e-05, | |
| "loss": 1.2513, | |
| "step": 29000 | |
| }, | |
| { | |
| "epoch": 3.241758241758242, | |
| "grad_norm": 2.5947108268737793, | |
| "learning_rate": 7.032967032967034e-05, | |
| "loss": 1.2757, | |
| "step": 29500 | |
| }, | |
| { | |
| "epoch": 3.2967032967032965, | |
| "grad_norm": 2.707387924194336, | |
| "learning_rate": 6.813186813186814e-05, | |
| "loss": 1.274, | |
| "step": 30000 | |
| }, | |
| { | |
| "epoch": 3.3516483516483517, | |
| "grad_norm": 3.4624383449554443, | |
| "learning_rate": 6.593406593406594e-05, | |
| "loss": 1.2702, | |
| "step": 30500 | |
| }, | |
| { | |
| "epoch": 3.4065934065934065, | |
| "grad_norm": 3.7795355319976807, | |
| "learning_rate": 6.373626373626373e-05, | |
| "loss": 1.2409, | |
| "step": 31000 | |
| }, | |
| { | |
| "epoch": 3.4615384615384617, | |
| "grad_norm": 3.521811008453369, | |
| "learning_rate": 6.153846153846155e-05, | |
| "loss": 1.267, | |
| "step": 31500 | |
| }, | |
| { | |
| "epoch": 3.5164835164835164, | |
| "grad_norm": 2.7402503490448, | |
| "learning_rate": 5.9340659340659345e-05, | |
| "loss": 1.2641, | |
| "step": 32000 | |
| }, | |
| { | |
| "epoch": 3.571428571428571, | |
| "grad_norm": 2.937671661376953, | |
| "learning_rate": 5.714285714285714e-05, | |
| "loss": 1.2626, | |
| "step": 32500 | |
| }, | |
| { | |
| "epoch": 3.6263736263736264, | |
| "grad_norm": 2.886845350265503, | |
| "learning_rate": 5.494505494505495e-05, | |
| "loss": 1.269, | |
| "step": 33000 | |
| }, | |
| { | |
| "epoch": 3.6813186813186816, | |
| "grad_norm": 3.9152579307556152, | |
| "learning_rate": 5.274725274725275e-05, | |
| "loss": 1.2659, | |
| "step": 33500 | |
| }, | |
| { | |
| "epoch": 3.7362637362637363, | |
| "grad_norm": 2.7805376052856445, | |
| "learning_rate": 5.054945054945055e-05, | |
| "loss": 1.2523, | |
| "step": 34000 | |
| }, | |
| { | |
| "epoch": 3.791208791208791, | |
| "grad_norm": 3.7305562496185303, | |
| "learning_rate": 4.8351648351648355e-05, | |
| "loss": 1.2528, | |
| "step": 34500 | |
| }, | |
| { | |
| "epoch": 3.8461538461538463, | |
| "grad_norm": 3.0859031677246094, | |
| "learning_rate": 4.615384615384616e-05, | |
| "loss": 1.273, | |
| "step": 35000 | |
| }, | |
| { | |
| "epoch": 3.901098901098901, | |
| "grad_norm": 4.31486177444458, | |
| "learning_rate": 4.3956043956043955e-05, | |
| "loss": 1.2739, | |
| "step": 35500 | |
| }, | |
| { | |
| "epoch": 3.956043956043956, | |
| "grad_norm": 2.46838116645813, | |
| "learning_rate": 4.1758241758241765e-05, | |
| "loss": 1.2555, | |
| "step": 36000 | |
| }, | |
| { | |
| "epoch": 4.010989010989011, | |
| "grad_norm": 3.3113248348236084, | |
| "learning_rate": 3.956043956043956e-05, | |
| "loss": 1.2388, | |
| "step": 36500 | |
| }, | |
| { | |
| "epoch": 4.065934065934066, | |
| "grad_norm": 5.830134868621826, | |
| "learning_rate": 3.7362637362637365e-05, | |
| "loss": 1.1952, | |
| "step": 37000 | |
| }, | |
| { | |
| "epoch": 4.1208791208791204, | |
| "grad_norm": 2.97209095954895, | |
| "learning_rate": 3.516483516483517e-05, | |
| "loss": 1.1902, | |
| "step": 37500 | |
| }, | |
| { | |
| "epoch": 4.175824175824176, | |
| "grad_norm": 3.9511055946350098, | |
| "learning_rate": 3.296703296703297e-05, | |
| "loss": 1.1837, | |
| "step": 38000 | |
| }, | |
| { | |
| "epoch": 4.230769230769231, | |
| "grad_norm": 6.446171760559082, | |
| "learning_rate": 3.0769230769230774e-05, | |
| "loss": 1.1888, | |
| "step": 38500 | |
| }, | |
| { | |
| "epoch": 4.285714285714286, | |
| "grad_norm": 4.0232744216918945, | |
| "learning_rate": 2.857142857142857e-05, | |
| "loss": 1.1752, | |
| "step": 39000 | |
| }, | |
| { | |
| "epoch": 4.34065934065934, | |
| "grad_norm": 4.632090091705322, | |
| "learning_rate": 2.6373626373626374e-05, | |
| "loss": 1.1964, | |
| "step": 39500 | |
| }, | |
| { | |
| "epoch": 4.395604395604396, | |
| "grad_norm": 3.9423835277557373, | |
| "learning_rate": 2.4175824175824177e-05, | |
| "loss": 1.1861, | |
| "step": 40000 | |
| }, | |
| { | |
| "epoch": 4.450549450549451, | |
| "grad_norm": 4.528479099273682, | |
| "learning_rate": 2.1978021978021977e-05, | |
| "loss": 1.1999, | |
| "step": 40500 | |
| }, | |
| { | |
| "epoch": 4.5054945054945055, | |
| "grad_norm": 4.554440021514893, | |
| "learning_rate": 1.978021978021978e-05, | |
| "loss": 1.2045, | |
| "step": 41000 | |
| }, | |
| { | |
| "epoch": 4.56043956043956, | |
| "grad_norm": 4.417350769042969, | |
| "learning_rate": 1.7582417582417584e-05, | |
| "loss": 1.2074, | |
| "step": 41500 | |
| }, | |
| { | |
| "epoch": 4.615384615384615, | |
| "grad_norm": 4.048484802246094, | |
| "learning_rate": 1.5384615384615387e-05, | |
| "loss": 1.203, | |
| "step": 42000 | |
| }, | |
| { | |
| "epoch": 4.670329670329671, | |
| "grad_norm": 4.659012317657471, | |
| "learning_rate": 1.3186813186813187e-05, | |
| "loss": 1.1917, | |
| "step": 42500 | |
| }, | |
| { | |
| "epoch": 4.725274725274725, | |
| "grad_norm": 3.7058045864105225, | |
| "learning_rate": 1.0989010989010989e-05, | |
| "loss": 1.1981, | |
| "step": 43000 | |
| }, | |
| { | |
| "epoch": 4.78021978021978, | |
| "grad_norm": 3.214338779449463, | |
| "learning_rate": 8.791208791208792e-06, | |
| "loss": 1.1868, | |
| "step": 43500 | |
| }, | |
| { | |
| "epoch": 4.835164835164835, | |
| "grad_norm": 3.2473630905151367, | |
| "learning_rate": 6.5934065934065935e-06, | |
| "loss": 1.1979, | |
| "step": 44000 | |
| }, | |
| { | |
| "epoch": 4.8901098901098905, | |
| "grad_norm": 3.6867451667785645, | |
| "learning_rate": 4.395604395604396e-06, | |
| "loss": 1.1873, | |
| "step": 44500 | |
| }, | |
| { | |
| "epoch": 4.945054945054945, | |
| "grad_norm": 4.085856914520264, | |
| "learning_rate": 2.197802197802198e-06, | |
| "loss": 1.1997, | |
| "step": 45000 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 3.7809948921203613, | |
| "learning_rate": 0.0, | |
| "loss": 1.1797, | |
| "step": 45500 | |
| } | |
| ], | |
| "logging_steps": 500, | |
| "max_steps": 45500, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 3.6063170257526784e+18, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |