| { | |
| "best_metric": 0.7447552447552448, | |
| "best_model_checkpoint": "wav2vec2-5Class-train-test-finetune-Medium/checkpoint-2640", | |
| "epoch": 178.98305084745763, | |
| "eval_steps": 500, | |
| "global_step": 2640, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 0.8927629590034485, | |
| "learning_rate": 1.2000000000000002e-06, | |
| "loss": 1.6852, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "eval_accuracy": 0.34265734265734266, | |
| "eval_loss": 1.5986738204956055, | |
| "eval_runtime": 4.6543, | |
| "eval_samples_per_second": 61.448, | |
| "eval_steps_per_second": 3.867, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "grad_norm": 0.6501776576042175, | |
| "learning_rate": 2.4857142857142858e-06, | |
| "loss": 1.5721, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "eval_accuracy": 0.34265734265734266, | |
| "eval_loss": 1.5975812673568726, | |
| "eval_runtime": 4.4155, | |
| "eval_samples_per_second": 64.772, | |
| "eval_steps_per_second": 4.077, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "grad_norm": 0.6221457719802856, | |
| "learning_rate": 3.771428571428572e-06, | |
| "loss": 1.5696, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "eval_accuracy": 0.34265734265734266, | |
| "eval_loss": 1.5957201719284058, | |
| "eval_runtime": 4.5928, | |
| "eval_samples_per_second": 62.272, | |
| "eval_steps_per_second": 3.919, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 1.563844919204712, | |
| "learning_rate": 5.057142857142857e-06, | |
| "loss": 1.5671, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_accuracy": 0.3356643356643357, | |
| "eval_loss": 1.5932137966156006, | |
| "eval_runtime": 4.5483, | |
| "eval_samples_per_second": 62.881, | |
| "eval_steps_per_second": 3.958, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "grad_norm": 0.7324579358100891, | |
| "learning_rate": 6.257142857142858e-06, | |
| "loss": 1.6757, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "eval_accuracy": 0.32517482517482516, | |
| "eval_loss": 1.5901715755462646, | |
| "eval_runtime": 4.9986, | |
| "eval_samples_per_second": 57.216, | |
| "eval_steps_per_second": 3.601, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 5.97, | |
| "grad_norm": 0.878238320350647, | |
| "learning_rate": 7.542857142857144e-06, | |
| "loss": 1.5595, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 5.97, | |
| "eval_accuracy": 0.32167832167832167, | |
| "eval_loss": 1.5863642692565918, | |
| "eval_runtime": 4.9711, | |
| "eval_samples_per_second": 57.532, | |
| "eval_steps_per_second": 3.621, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 6.98, | |
| "grad_norm": 1.4210667610168457, | |
| "learning_rate": 8.828571428571429e-06, | |
| "loss": 1.5536, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 6.98, | |
| "eval_accuracy": 0.3181818181818182, | |
| "eval_loss": 1.5817956924438477, | |
| "eval_runtime": 5.5676, | |
| "eval_samples_per_second": 51.368, | |
| "eval_steps_per_second": 3.233, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "grad_norm": 0.49261531233787537, | |
| "learning_rate": 1.0114285714285715e-05, | |
| "loss": 1.5484, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_accuracy": 0.3111888111888112, | |
| "eval_loss": 1.576475739479065, | |
| "eval_runtime": 5.5008, | |
| "eval_samples_per_second": 51.992, | |
| "eval_steps_per_second": 3.272, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 8.95, | |
| "grad_norm": 0.9810589551925659, | |
| "learning_rate": 1.1314285714285715e-05, | |
| "loss": 1.6506, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 8.95, | |
| "eval_accuracy": 0.3076923076923077, | |
| "eval_loss": 1.570804238319397, | |
| "eval_runtime": 4.0121, | |
| "eval_samples_per_second": 71.285, | |
| "eval_steps_per_second": 4.486, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 9.97, | |
| "grad_norm": 0.6308433413505554, | |
| "learning_rate": 1.26e-05, | |
| "loss": 1.5317, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 9.97, | |
| "eval_accuracy": 0.3006993006993007, | |
| "eval_loss": 1.5640877485275269, | |
| "eval_runtime": 5.5443, | |
| "eval_samples_per_second": 51.585, | |
| "eval_steps_per_second": 3.247, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 10.98, | |
| "grad_norm": 1.381785273551941, | |
| "learning_rate": 1.3885714285714286e-05, | |
| "loss": 1.5226, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 10.98, | |
| "eval_accuracy": 0.2867132867132867, | |
| "eval_loss": 1.5564289093017578, | |
| "eval_runtime": 5.0151, | |
| "eval_samples_per_second": 57.027, | |
| "eval_steps_per_second": 3.589, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "grad_norm": 0.552884578704834, | |
| "learning_rate": 1.517142857142857e-05, | |
| "loss": 1.5116, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 12.0, | |
| "eval_accuracy": 0.2692307692307692, | |
| "eval_loss": 1.5478310585021973, | |
| "eval_runtime": 5.4183, | |
| "eval_samples_per_second": 52.784, | |
| "eval_steps_per_second": 3.322, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 12.95, | |
| "grad_norm": 0.8657445907592773, | |
| "learning_rate": 1.6371428571428572e-05, | |
| "loss": 1.6046, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 12.95, | |
| "eval_accuracy": 0.26223776223776224, | |
| "eval_loss": 1.5384888648986816, | |
| "eval_runtime": 5.1185, | |
| "eval_samples_per_second": 55.875, | |
| "eval_steps_per_second": 3.517, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 13.97, | |
| "grad_norm": 0.8375154733657837, | |
| "learning_rate": 1.7657142857142857e-05, | |
| "loss": 1.4822, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 13.97, | |
| "eval_accuracy": 0.25524475524475526, | |
| "eval_loss": 1.5284570455551147, | |
| "eval_runtime": 5.5939, | |
| "eval_samples_per_second": 51.127, | |
| "eval_steps_per_second": 3.218, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 14.98, | |
| "grad_norm": 0.7776028513908386, | |
| "learning_rate": 1.8942857142857145e-05, | |
| "loss": 1.4614, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 14.98, | |
| "eval_accuracy": 0.25524475524475526, | |
| "eval_loss": 1.5183566808700562, | |
| "eval_runtime": 5.2849, | |
| "eval_samples_per_second": 54.116, | |
| "eval_steps_per_second": 3.406, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "grad_norm": 0.3529152572154999, | |
| "learning_rate": 2.022857142857143e-05, | |
| "loss": 1.4396, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 16.0, | |
| "eval_accuracy": 0.2517482517482518, | |
| "eval_loss": 1.510148048400879, | |
| "eval_runtime": 6.3848, | |
| "eval_samples_per_second": 44.794, | |
| "eval_steps_per_second": 2.819, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 16.95, | |
| "grad_norm": 0.5623555779457092, | |
| "learning_rate": 2.1428571428571428e-05, | |
| "loss": 1.5047, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 16.95, | |
| "eval_accuracy": 0.24475524475524477, | |
| "eval_loss": 1.510407567024231, | |
| "eval_runtime": 5.762, | |
| "eval_samples_per_second": 49.635, | |
| "eval_steps_per_second": 3.124, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 17.97, | |
| "grad_norm": 0.889561116695404, | |
| "learning_rate": 2.2714285714285713e-05, | |
| "loss": 1.3741, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 17.97, | |
| "eval_accuracy": 0.2517482517482518, | |
| "eval_loss": 1.524996280670166, | |
| "eval_runtime": 5.3825, | |
| "eval_samples_per_second": 53.135, | |
| "eval_steps_per_second": 3.344, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 18.98, | |
| "grad_norm": 0.837506890296936, | |
| "learning_rate": 2.4e-05, | |
| "loss": 1.3512, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 18.98, | |
| "eval_accuracy": 0.26573426573426573, | |
| "eval_loss": 1.5328779220581055, | |
| "eval_runtime": 4.8496, | |
| "eval_samples_per_second": 58.974, | |
| "eval_steps_per_second": 3.712, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 0.6774041056632996, | |
| "learning_rate": 2.5285714285714285e-05, | |
| "loss": 1.3286, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "eval_accuracy": 0.32517482517482516, | |
| "eval_loss": 1.509504795074463, | |
| "eval_runtime": 4.7142, | |
| "eval_samples_per_second": 60.667, | |
| "eval_steps_per_second": 3.818, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 20.95, | |
| "grad_norm": 0.5629450678825378, | |
| "learning_rate": 2.6485714285714287e-05, | |
| "loss": 1.3967, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 20.95, | |
| "eval_accuracy": 0.34965034965034963, | |
| "eval_loss": 1.482858419418335, | |
| "eval_runtime": 5.0574, | |
| "eval_samples_per_second": 56.551, | |
| "eval_steps_per_second": 3.559, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 21.97, | |
| "grad_norm": 0.5265232920646667, | |
| "learning_rate": 2.7771428571428572e-05, | |
| "loss": 1.2779, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 21.97, | |
| "eval_accuracy": 0.38461538461538464, | |
| "eval_loss": 1.4598273038864136, | |
| "eval_runtime": 5.2206, | |
| "eval_samples_per_second": 54.783, | |
| "eval_steps_per_second": 3.448, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 22.98, | |
| "grad_norm": 0.4762378931045532, | |
| "learning_rate": 2.9057142857142856e-05, | |
| "loss": 1.2449, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 22.98, | |
| "eval_accuracy": 0.4160839160839161, | |
| "eval_loss": 1.4284018278121948, | |
| "eval_runtime": 5.3164, | |
| "eval_samples_per_second": 53.796, | |
| "eval_steps_per_second": 3.386, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "grad_norm": 0.598441481590271, | |
| "learning_rate": 2.996190476190476e-05, | |
| "loss": 1.2118, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 24.0, | |
| "eval_accuracy": 0.4230769230769231, | |
| "eval_loss": 1.4161577224731445, | |
| "eval_runtime": 4.9587, | |
| "eval_samples_per_second": 57.677, | |
| "eval_steps_per_second": 3.63, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 24.95, | |
| "grad_norm": 1.1748360395431519, | |
| "learning_rate": 2.982857142857143e-05, | |
| "loss": 1.2521, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 24.95, | |
| "eval_accuracy": 0.44755244755244755, | |
| "eval_loss": 1.3797944784164429, | |
| "eval_runtime": 7.057, | |
| "eval_samples_per_second": 40.527, | |
| "eval_steps_per_second": 2.551, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 25.97, | |
| "grad_norm": 0.7063133716583252, | |
| "learning_rate": 2.9685714285714284e-05, | |
| "loss": 1.1183, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 25.97, | |
| "eval_accuracy": 0.479020979020979, | |
| "eval_loss": 1.324568271636963, | |
| "eval_runtime": 5.5598, | |
| "eval_samples_per_second": 51.44, | |
| "eval_steps_per_second": 3.238, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 26.98, | |
| "grad_norm": 0.91399085521698, | |
| "learning_rate": 2.9542857142857142e-05, | |
| "loss": 1.0778, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 26.98, | |
| "eval_accuracy": 0.493006993006993, | |
| "eval_loss": 1.276139259338379, | |
| "eval_runtime": 4.8059, | |
| "eval_samples_per_second": 59.51, | |
| "eval_steps_per_second": 3.745, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "grad_norm": 1.1466203927993774, | |
| "learning_rate": 2.94e-05, | |
| "loss": 1.0306, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 28.0, | |
| "eval_accuracy": 0.5104895104895105, | |
| "eval_loss": 1.2299922704696655, | |
| "eval_runtime": 6.0199, | |
| "eval_samples_per_second": 47.509, | |
| "eval_steps_per_second": 2.99, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 28.95, | |
| "grad_norm": 1.1782015562057495, | |
| "learning_rate": 2.9266666666666665e-05, | |
| "loss": 1.0808, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 28.95, | |
| "eval_accuracy": 0.513986013986014, | |
| "eval_loss": 1.2064085006713867, | |
| "eval_runtime": 6.0946, | |
| "eval_samples_per_second": 46.927, | |
| "eval_steps_per_second": 2.953, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 29.97, | |
| "grad_norm": 0.9410634636878967, | |
| "learning_rate": 2.9123809523809523e-05, | |
| "loss": 0.9767, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 29.97, | |
| "eval_accuracy": 0.5524475524475524, | |
| "eval_loss": 1.1669567823410034, | |
| "eval_runtime": 4.8236, | |
| "eval_samples_per_second": 59.292, | |
| "eval_steps_per_second": 3.732, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 30.98, | |
| "grad_norm": 1.2911593914031982, | |
| "learning_rate": 2.898095238095238e-05, | |
| "loss": 0.9589, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 30.98, | |
| "eval_accuracy": 0.5734265734265734, | |
| "eval_loss": 1.126379132270813, | |
| "eval_runtime": 5.0702, | |
| "eval_samples_per_second": 56.408, | |
| "eval_steps_per_second": 3.55, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "grad_norm": 1.2354660034179688, | |
| "learning_rate": 2.883809523809524e-05, | |
| "loss": 0.9193, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 32.0, | |
| "eval_accuracy": 0.5874125874125874, | |
| "eval_loss": 1.1012728214263916, | |
| "eval_runtime": 5.9275, | |
| "eval_samples_per_second": 48.249, | |
| "eval_steps_per_second": 3.037, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 32.95, | |
| "grad_norm": 0.8961493968963623, | |
| "learning_rate": 2.8704761904761905e-05, | |
| "loss": 0.9462, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 32.95, | |
| "eval_accuracy": 0.6048951048951049, | |
| "eval_loss": 1.0736197233200073, | |
| "eval_runtime": 5.7691, | |
| "eval_samples_per_second": 49.575, | |
| "eval_steps_per_second": 3.12, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 33.97, | |
| "grad_norm": 0.9503061175346375, | |
| "learning_rate": 2.8561904761904763e-05, | |
| "loss": 0.85, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 33.97, | |
| "eval_accuracy": 0.6048951048951049, | |
| "eval_loss": 1.0628284215927124, | |
| "eval_runtime": 6.1054, | |
| "eval_samples_per_second": 46.844, | |
| "eval_steps_per_second": 2.948, | |
| "step": 501 | |
| }, | |
| { | |
| "epoch": 34.98, | |
| "grad_norm": 1.0706520080566406, | |
| "learning_rate": 2.841904761904762e-05, | |
| "loss": 0.8294, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 34.98, | |
| "eval_accuracy": 0.6188811188811189, | |
| "eval_loss": 1.0472813844680786, | |
| "eval_runtime": 4.7734, | |
| "eval_samples_per_second": 59.916, | |
| "eval_steps_per_second": 3.771, | |
| "step": 516 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "grad_norm": 0.9784532189369202, | |
| "learning_rate": 2.827619047619048e-05, | |
| "loss": 0.8025, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 36.0, | |
| "eval_accuracy": 0.6328671328671329, | |
| "eval_loss": 1.0030184984207153, | |
| "eval_runtime": 5.5291, | |
| "eval_samples_per_second": 51.726, | |
| "eval_steps_per_second": 3.255, | |
| "step": 531 | |
| }, | |
| { | |
| "epoch": 36.95, | |
| "grad_norm": 0.8138810396194458, | |
| "learning_rate": 2.8142857142857145e-05, | |
| "loss": 0.8206, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 36.95, | |
| "eval_accuracy": 0.6398601398601399, | |
| "eval_loss": 0.9963561296463013, | |
| "eval_runtime": 5.8221, | |
| "eval_samples_per_second": 49.123, | |
| "eval_steps_per_second": 3.092, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 37.97, | |
| "grad_norm": 0.8114917278289795, | |
| "learning_rate": 2.8e-05, | |
| "loss": 0.7541, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 37.97, | |
| "eval_accuracy": 0.6608391608391608, | |
| "eval_loss": 0.9604987502098083, | |
| "eval_runtime": 4.9473, | |
| "eval_samples_per_second": 57.809, | |
| "eval_steps_per_second": 3.638, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 38.98, | |
| "grad_norm": 0.9221667051315308, | |
| "learning_rate": 2.7857142857142858e-05, | |
| "loss": 0.7413, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 38.98, | |
| "eval_accuracy": 0.6643356643356644, | |
| "eval_loss": 0.946722686290741, | |
| "eval_runtime": 4.7955, | |
| "eval_samples_per_second": 59.64, | |
| "eval_steps_per_second": 3.754, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "grad_norm": 1.2665307521820068, | |
| "learning_rate": 2.7714285714285716e-05, | |
| "loss": 0.709, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "eval_accuracy": 0.6678321678321678, | |
| "eval_loss": 0.9348079562187195, | |
| "eval_runtime": 5.3252, | |
| "eval_samples_per_second": 53.707, | |
| "eval_steps_per_second": 3.38, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 40.95, | |
| "grad_norm": 1.0893254280090332, | |
| "learning_rate": 2.758095238095238e-05, | |
| "loss": 0.7817, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 40.95, | |
| "eval_accuracy": 0.6678321678321678, | |
| "eval_loss": 0.9366168975830078, | |
| "eval_runtime": 6.4862, | |
| "eval_samples_per_second": 44.094, | |
| "eval_steps_per_second": 2.775, | |
| "step": 604 | |
| }, | |
| { | |
| "epoch": 41.97, | |
| "grad_norm": 0.7136903405189514, | |
| "learning_rate": 2.743809523809524e-05, | |
| "loss": 0.7034, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 41.97, | |
| "eval_accuracy": 0.6818181818181818, | |
| "eval_loss": 0.9109411835670471, | |
| "eval_runtime": 6.1724, | |
| "eval_samples_per_second": 46.335, | |
| "eval_steps_per_second": 2.916, | |
| "step": 619 | |
| }, | |
| { | |
| "epoch": 42.98, | |
| "grad_norm": 1.142288327217102, | |
| "learning_rate": 2.7295238095238097e-05, | |
| "loss": 0.6856, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 42.98, | |
| "eval_accuracy": 0.6573426573426573, | |
| "eval_loss": 0.9276965856552124, | |
| "eval_runtime": 4.6805, | |
| "eval_samples_per_second": 61.105, | |
| "eval_steps_per_second": 3.846, | |
| "step": 634 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "grad_norm": 1.3661341667175293, | |
| "learning_rate": 2.7152380952380952e-05, | |
| "loss": 0.6625, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 44.0, | |
| "eval_accuracy": 0.6783216783216783, | |
| "eval_loss": 0.8979520201683044, | |
| "eval_runtime": 5.8084, | |
| "eval_samples_per_second": 49.239, | |
| "eval_steps_per_second": 3.099, | |
| "step": 649 | |
| }, | |
| { | |
| "epoch": 44.95, | |
| "grad_norm": 0.7071289420127869, | |
| "learning_rate": 2.701904761904762e-05, | |
| "loss": 0.7207, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 44.95, | |
| "eval_accuracy": 0.6713286713286714, | |
| "eval_loss": 0.9049975275993347, | |
| "eval_runtime": 5.5071, | |
| "eval_samples_per_second": 51.933, | |
| "eval_steps_per_second": 3.268, | |
| "step": 663 | |
| }, | |
| { | |
| "epoch": 45.97, | |
| "grad_norm": 0.9930222034454346, | |
| "learning_rate": 2.6876190476190476e-05, | |
| "loss": 0.6684, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 45.97, | |
| "eval_accuracy": 0.6748251748251748, | |
| "eval_loss": 0.8973459005355835, | |
| "eval_runtime": 4.8744, | |
| "eval_samples_per_second": 58.674, | |
| "eval_steps_per_second": 3.693, | |
| "step": 678 | |
| }, | |
| { | |
| "epoch": 46.98, | |
| "grad_norm": 1.2905802726745605, | |
| "learning_rate": 2.6733333333333334e-05, | |
| "loss": 0.6651, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 46.98, | |
| "eval_accuracy": 0.6783216783216783, | |
| "eval_loss": 0.8934686779975891, | |
| "eval_runtime": 4.9699, | |
| "eval_samples_per_second": 57.546, | |
| "eval_steps_per_second": 3.622, | |
| "step": 693 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "grad_norm": 1.5243291854858398, | |
| "learning_rate": 2.6590476190476192e-05, | |
| "loss": 0.6451, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 48.0, | |
| "eval_accuracy": 0.6748251748251748, | |
| "eval_loss": 0.8944538831710815, | |
| "eval_runtime": 4.9181, | |
| "eval_samples_per_second": 58.153, | |
| "eval_steps_per_second": 3.66, | |
| "step": 708 | |
| }, | |
| { | |
| "epoch": 48.95, | |
| "grad_norm": 1.4421076774597168, | |
| "learning_rate": 2.6457142857142857e-05, | |
| "loss": 0.6774, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 48.95, | |
| "eval_accuracy": 0.6818181818181818, | |
| "eval_loss": 0.8879284858703613, | |
| "eval_runtime": 5.1997, | |
| "eval_samples_per_second": 55.003, | |
| "eval_steps_per_second": 3.462, | |
| "step": 722 | |
| }, | |
| { | |
| "epoch": 49.97, | |
| "grad_norm": 1.3786752223968506, | |
| "learning_rate": 2.6314285714285715e-05, | |
| "loss": 0.6308, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 49.97, | |
| "eval_accuracy": 0.6818181818181818, | |
| "eval_loss": 0.8882840275764465, | |
| "eval_runtime": 5.1707, | |
| "eval_samples_per_second": 55.312, | |
| "eval_steps_per_second": 3.481, | |
| "step": 737 | |
| }, | |
| { | |
| "epoch": 50.98, | |
| "grad_norm": 1.9291437864303589, | |
| "learning_rate": 2.617142857142857e-05, | |
| "loss": 0.6199, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 50.98, | |
| "eval_accuracy": 0.6818181818181818, | |
| "eval_loss": 0.8825680017471313, | |
| "eval_runtime": 4.9861, | |
| "eval_samples_per_second": 57.359, | |
| "eval_steps_per_second": 3.61, | |
| "step": 752 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "grad_norm": 1.2054837942123413, | |
| "learning_rate": 2.6028571428571428e-05, | |
| "loss": 0.6379, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 52.0, | |
| "eval_accuracy": 0.6923076923076923, | |
| "eval_loss": 0.8582118153572083, | |
| "eval_runtime": 5.287, | |
| "eval_samples_per_second": 54.095, | |
| "eval_steps_per_second": 3.405, | |
| "step": 767 | |
| }, | |
| { | |
| "epoch": 52.95, | |
| "grad_norm": 1.3063265085220337, | |
| "learning_rate": 2.5895238095238094e-05, | |
| "loss": 0.6588, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 52.95, | |
| "eval_accuracy": 0.6818181818181818, | |
| "eval_loss": 0.8825483918190002, | |
| "eval_runtime": 4.6024, | |
| "eval_samples_per_second": 62.141, | |
| "eval_steps_per_second": 3.911, | |
| "step": 781 | |
| }, | |
| { | |
| "epoch": 53.97, | |
| "grad_norm": 1.2957918643951416, | |
| "learning_rate": 2.5752380952380952e-05, | |
| "loss": 0.5857, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 53.97, | |
| "eval_accuracy": 0.6748251748251748, | |
| "eval_loss": 0.8808070421218872, | |
| "eval_runtime": 5.2003, | |
| "eval_samples_per_second": 54.997, | |
| "eval_steps_per_second": 3.461, | |
| "step": 796 | |
| }, | |
| { | |
| "epoch": 54.98, | |
| "grad_norm": 0.9337932467460632, | |
| "learning_rate": 2.560952380952381e-05, | |
| "loss": 0.6076, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 54.98, | |
| "eval_accuracy": 0.6958041958041958, | |
| "eval_loss": 0.8554547429084778, | |
| "eval_runtime": 5.108, | |
| "eval_samples_per_second": 55.991, | |
| "eval_steps_per_second": 3.524, | |
| "step": 811 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "grad_norm": 1.4797887802124023, | |
| "learning_rate": 2.5466666666666668e-05, | |
| "loss": 0.5934, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 56.0, | |
| "eval_accuracy": 0.6888111888111889, | |
| "eval_loss": 0.8653700351715088, | |
| "eval_runtime": 4.9201, | |
| "eval_samples_per_second": 58.128, | |
| "eval_steps_per_second": 3.658, | |
| "step": 826 | |
| }, | |
| { | |
| "epoch": 56.95, | |
| "grad_norm": 0.9749770164489746, | |
| "learning_rate": 2.5333333333333334e-05, | |
| "loss": 0.6427, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 56.95, | |
| "eval_accuracy": 0.6853146853146853, | |
| "eval_loss": 0.861566424369812, | |
| "eval_runtime": 5.5361, | |
| "eval_samples_per_second": 51.661, | |
| "eval_steps_per_second": 3.251, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 57.97, | |
| "grad_norm": 1.1588115692138672, | |
| "learning_rate": 2.519047619047619e-05, | |
| "loss": 0.5782, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 57.97, | |
| "eval_accuracy": 0.6678321678321678, | |
| "eval_loss": 0.8711130619049072, | |
| "eval_runtime": 5.1513, | |
| "eval_samples_per_second": 55.52, | |
| "eval_steps_per_second": 3.494, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 58.98, | |
| "grad_norm": 2.290128707885742, | |
| "learning_rate": 2.504761904761905e-05, | |
| "loss": 0.5819, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 58.98, | |
| "eval_accuracy": 0.6748251748251748, | |
| "eval_loss": 0.8689377903938293, | |
| "eval_runtime": 5.3476, | |
| "eval_samples_per_second": 53.482, | |
| "eval_steps_per_second": 3.366, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "grad_norm": 1.8434489965438843, | |
| "learning_rate": 2.4904761904761908e-05, | |
| "loss": 0.5918, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 60.0, | |
| "eval_accuracy": 0.6923076923076923, | |
| "eval_loss": 0.8602246642112732, | |
| "eval_runtime": 5.1182, | |
| "eval_samples_per_second": 55.879, | |
| "eval_steps_per_second": 3.517, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 60.95, | |
| "grad_norm": 1.1310744285583496, | |
| "learning_rate": 2.4771428571428573e-05, | |
| "loss": 0.5845, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 60.95, | |
| "eval_accuracy": 0.6993006993006993, | |
| "eval_loss": 0.8458691835403442, | |
| "eval_runtime": 4.8028, | |
| "eval_samples_per_second": 59.549, | |
| "eval_steps_per_second": 3.748, | |
| "step": 899 | |
| }, | |
| { | |
| "epoch": 61.97, | |
| "grad_norm": 1.319384217262268, | |
| "learning_rate": 2.462857142857143e-05, | |
| "loss": 0.5667, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 61.97, | |
| "eval_accuracy": 0.7027972027972028, | |
| "eval_loss": 0.8466892838478088, | |
| "eval_runtime": 4.9074, | |
| "eval_samples_per_second": 58.279, | |
| "eval_steps_per_second": 3.668, | |
| "step": 914 | |
| }, | |
| { | |
| "epoch": 62.98, | |
| "grad_norm": 1.2371602058410645, | |
| "learning_rate": 2.448571428571429e-05, | |
| "loss": 0.5327, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 62.98, | |
| "eval_accuracy": 0.6923076923076923, | |
| "eval_loss": 0.8539786338806152, | |
| "eval_runtime": 5.32, | |
| "eval_samples_per_second": 53.759, | |
| "eval_steps_per_second": 3.383, | |
| "step": 929 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "grad_norm": 2.1778481006622314, | |
| "learning_rate": 2.434285714285714e-05, | |
| "loss": 0.523, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 64.0, | |
| "eval_accuracy": 0.7062937062937062, | |
| "eval_loss": 0.8323072791099548, | |
| "eval_runtime": 5.1152, | |
| "eval_samples_per_second": 55.912, | |
| "eval_steps_per_second": 3.519, | |
| "step": 944 | |
| }, | |
| { | |
| "epoch": 64.95, | |
| "grad_norm": 0.8219490051269531, | |
| "learning_rate": 2.420952380952381e-05, | |
| "loss": 0.548, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 64.95, | |
| "eval_accuracy": 0.6993006993006993, | |
| "eval_loss": 0.8406782746315002, | |
| "eval_runtime": 5.325, | |
| "eval_samples_per_second": 53.709, | |
| "eval_steps_per_second": 3.38, | |
| "step": 958 | |
| }, | |
| { | |
| "epoch": 65.97, | |
| "grad_norm": 2.0205609798431396, | |
| "learning_rate": 2.4066666666666664e-05, | |
| "loss": 0.5399, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 65.97, | |
| "eval_accuracy": 0.6993006993006993, | |
| "eval_loss": 0.837882936000824, | |
| "eval_runtime": 5.6808, | |
| "eval_samples_per_second": 50.345, | |
| "eval_steps_per_second": 3.169, | |
| "step": 973 | |
| }, | |
| { | |
| "epoch": 66.98, | |
| "grad_norm": 1.527024507522583, | |
| "learning_rate": 2.3923809523809522e-05, | |
| "loss": 0.5324, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 66.98, | |
| "eval_accuracy": 0.7027972027972028, | |
| "eval_loss": 0.8118743896484375, | |
| "eval_runtime": 5.543, | |
| "eval_samples_per_second": 51.596, | |
| "eval_steps_per_second": 3.247, | |
| "step": 988 | |
| }, | |
| { | |
| "epoch": 68.0, | |
| "grad_norm": 1.3294800519943237, | |
| "learning_rate": 2.378095238095238e-05, | |
| "loss": 0.5171, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 68.0, | |
| "eval_accuracy": 0.6923076923076923, | |
| "eval_loss": 0.8444966673851013, | |
| "eval_runtime": 5.5571, | |
| "eval_samples_per_second": 51.465, | |
| "eval_steps_per_second": 3.239, | |
| "step": 1003 | |
| }, | |
| { | |
| "epoch": 68.95, | |
| "grad_norm": 1.6269547939300537, | |
| "learning_rate": 2.3647619047619046e-05, | |
| "loss": 0.538, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 68.95, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.8196296095848083, | |
| "eval_runtime": 4.8636, | |
| "eval_samples_per_second": 58.805, | |
| "eval_steps_per_second": 3.701, | |
| "step": 1017 | |
| }, | |
| { | |
| "epoch": 69.97, | |
| "grad_norm": 1.0726577043533325, | |
| "learning_rate": 2.3504761904761904e-05, | |
| "loss": 0.5312, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 69.97, | |
| "eval_accuracy": 0.6853146853146853, | |
| "eval_loss": 0.8414965271949768, | |
| "eval_runtime": 5.2322, | |
| "eval_samples_per_second": 54.661, | |
| "eval_steps_per_second": 3.44, | |
| "step": 1032 | |
| }, | |
| { | |
| "epoch": 70.98, | |
| "grad_norm": 1.0595276355743408, | |
| "learning_rate": 2.3361904761904762e-05, | |
| "loss": 0.4914, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 70.98, | |
| "eval_accuracy": 0.6958041958041958, | |
| "eval_loss": 0.8184240460395813, | |
| "eval_runtime": 4.6109, | |
| "eval_samples_per_second": 62.027, | |
| "eval_steps_per_second": 3.904, | |
| "step": 1047 | |
| }, | |
| { | |
| "epoch": 72.0, | |
| "grad_norm": 1.3313194513320923, | |
| "learning_rate": 2.321904761904762e-05, | |
| "loss": 0.5055, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 72.0, | |
| "eval_accuracy": 0.6923076923076923, | |
| "eval_loss": 0.8218348622322083, | |
| "eval_runtime": 5.5535, | |
| "eval_samples_per_second": 51.499, | |
| "eval_steps_per_second": 3.241, | |
| "step": 1062 | |
| }, | |
| { | |
| "epoch": 72.95, | |
| "grad_norm": 1.9882231950759888, | |
| "learning_rate": 2.3085714285714286e-05, | |
| "loss": 0.5401, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 72.95, | |
| "eval_accuracy": 0.7027972027972028, | |
| "eval_loss": 0.815979540348053, | |
| "eval_runtime": 6.0512, | |
| "eval_samples_per_second": 47.264, | |
| "eval_steps_per_second": 2.975, | |
| "step": 1076 | |
| }, | |
| { | |
| "epoch": 73.97, | |
| "grad_norm": 1.8936933279037476, | |
| "learning_rate": 2.2942857142857144e-05, | |
| "loss": 0.4966, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 73.97, | |
| "eval_accuracy": 0.6888111888111889, | |
| "eval_loss": 0.8237566351890564, | |
| "eval_runtime": 4.6622, | |
| "eval_samples_per_second": 61.344, | |
| "eval_steps_per_second": 3.861, | |
| "step": 1091 | |
| }, | |
| { | |
| "epoch": 74.98, | |
| "grad_norm": 2.0124337673187256, | |
| "learning_rate": 2.2800000000000002e-05, | |
| "loss": 0.4768, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 74.98, | |
| "eval_accuracy": 0.6993006993006993, | |
| "eval_loss": 0.8185241222381592, | |
| "eval_runtime": 4.2442, | |
| "eval_samples_per_second": 67.386, | |
| "eval_steps_per_second": 4.241, | |
| "step": 1106 | |
| }, | |
| { | |
| "epoch": 76.0, | |
| "grad_norm": 1.7522495985031128, | |
| "learning_rate": 2.265714285714286e-05, | |
| "loss": 0.4789, | |
| "step": 1121 | |
| }, | |
| { | |
| "epoch": 76.0, | |
| "eval_accuracy": 0.7027972027972028, | |
| "eval_loss": 0.8260769844055176, | |
| "eval_runtime": 6.128, | |
| "eval_samples_per_second": 46.671, | |
| "eval_steps_per_second": 2.937, | |
| "step": 1121 | |
| }, | |
| { | |
| "epoch": 76.95, | |
| "grad_norm": 1.8254278898239136, | |
| "learning_rate": 2.2523809523809526e-05, | |
| "loss": 0.5176, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 76.95, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.8109525442123413, | |
| "eval_runtime": 6.0147, | |
| "eval_samples_per_second": 47.55, | |
| "eval_steps_per_second": 2.993, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 77.97, | |
| "grad_norm": 1.4586073160171509, | |
| "learning_rate": 2.238095238095238e-05, | |
| "loss": 0.466, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 77.97, | |
| "eval_accuracy": 0.6993006993006993, | |
| "eval_loss": 0.814057469367981, | |
| "eval_runtime": 4.782, | |
| "eval_samples_per_second": 59.808, | |
| "eval_steps_per_second": 3.764, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 78.98, | |
| "grad_norm": 1.7543085813522339, | |
| "learning_rate": 2.223809523809524e-05, | |
| "loss": 0.4736, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 78.98, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7970029711723328, | |
| "eval_runtime": 5.8124, | |
| "eval_samples_per_second": 49.205, | |
| "eval_steps_per_second": 3.097, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 80.0, | |
| "grad_norm": 1.6676491498947144, | |
| "learning_rate": 2.2095238095238096e-05, | |
| "loss": 0.4785, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 80.0, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.8062326312065125, | |
| "eval_runtime": 4.5174, | |
| "eval_samples_per_second": 63.31, | |
| "eval_steps_per_second": 3.985, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 80.95, | |
| "grad_norm": 1.827528476715088, | |
| "learning_rate": 2.1961904761904762e-05, | |
| "loss": 0.5309, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 80.95, | |
| "eval_accuracy": 0.6958041958041958, | |
| "eval_loss": 0.8051398396492004, | |
| "eval_runtime": 4.9397, | |
| "eval_samples_per_second": 57.899, | |
| "eval_steps_per_second": 3.644, | |
| "step": 1194 | |
| }, | |
| { | |
| "epoch": 81.97, | |
| "grad_norm": 2.3700757026672363, | |
| "learning_rate": 2.181904761904762e-05, | |
| "loss": 0.4571, | |
| "step": 1209 | |
| }, | |
| { | |
| "epoch": 81.97, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.8024002909660339, | |
| "eval_runtime": 6.3277, | |
| "eval_samples_per_second": 45.198, | |
| "eval_steps_per_second": 2.845, | |
| "step": 1209 | |
| }, | |
| { | |
| "epoch": 82.98, | |
| "grad_norm": 1.5626410245895386, | |
| "learning_rate": 2.1676190476190478e-05, | |
| "loss": 0.47, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 82.98, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.8030509948730469, | |
| "eval_runtime": 5.4575, | |
| "eval_samples_per_second": 52.405, | |
| "eval_steps_per_second": 3.298, | |
| "step": 1224 | |
| }, | |
| { | |
| "epoch": 84.0, | |
| "grad_norm": 1.4918785095214844, | |
| "learning_rate": 2.1533333333333333e-05, | |
| "loss": 0.4525, | |
| "step": 1239 | |
| }, | |
| { | |
| "epoch": 84.0, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7910680174827576, | |
| "eval_runtime": 4.6183, | |
| "eval_samples_per_second": 61.928, | |
| "eval_steps_per_second": 3.898, | |
| "step": 1239 | |
| }, | |
| { | |
| "epoch": 84.95, | |
| "grad_norm": 1.4631191492080688, | |
| "learning_rate": 2.1400000000000002e-05, | |
| "loss": 0.5058, | |
| "step": 1253 | |
| }, | |
| { | |
| "epoch": 84.95, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7877373695373535, | |
| "eval_runtime": 4.6086, | |
| "eval_samples_per_second": 62.058, | |
| "eval_steps_per_second": 3.906, | |
| "step": 1253 | |
| }, | |
| { | |
| "epoch": 85.97, | |
| "grad_norm": 1.463442087173462, | |
| "learning_rate": 2.1257142857142856e-05, | |
| "loss": 0.4627, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 85.97, | |
| "eval_accuracy": 0.7062937062937062, | |
| "eval_loss": 0.7917687296867371, | |
| "eval_runtime": 6.4645, | |
| "eval_samples_per_second": 44.241, | |
| "eval_steps_per_second": 2.784, | |
| "step": 1268 | |
| }, | |
| { | |
| "epoch": 86.98, | |
| "grad_norm": 1.5172946453094482, | |
| "learning_rate": 2.1114285714285714e-05, | |
| "loss": 0.4343, | |
| "step": 1283 | |
| }, | |
| { | |
| "epoch": 86.98, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.78819739818573, | |
| "eval_runtime": 5.3121, | |
| "eval_samples_per_second": 53.839, | |
| "eval_steps_per_second": 3.388, | |
| "step": 1283 | |
| }, | |
| { | |
| "epoch": 88.0, | |
| "grad_norm": 1.8876270055770874, | |
| "learning_rate": 2.0971428571428572e-05, | |
| "loss": 0.4442, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 88.0, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.805809736251831, | |
| "eval_runtime": 5.4992, | |
| "eval_samples_per_second": 52.007, | |
| "eval_steps_per_second": 3.273, | |
| "step": 1298 | |
| }, | |
| { | |
| "epoch": 88.95, | |
| "grad_norm": 1.4128847122192383, | |
| "learning_rate": 2.0838095238095238e-05, | |
| "loss": 0.4745, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 88.95, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7810028791427612, | |
| "eval_runtime": 5.1967, | |
| "eval_samples_per_second": 55.035, | |
| "eval_steps_per_second": 3.464, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 89.97, | |
| "grad_norm": 1.3974703550338745, | |
| "learning_rate": 2.0695238095238096e-05, | |
| "loss": 0.4282, | |
| "step": 1327 | |
| }, | |
| { | |
| "epoch": 89.97, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.7951435446739197, | |
| "eval_runtime": 5.9497, | |
| "eval_samples_per_second": 48.07, | |
| "eval_steps_per_second": 3.025, | |
| "step": 1327 | |
| }, | |
| { | |
| "epoch": 90.98, | |
| "grad_norm": 1.8035740852355957, | |
| "learning_rate": 2.055238095238095e-05, | |
| "loss": 0.4307, | |
| "step": 1342 | |
| }, | |
| { | |
| "epoch": 90.98, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7739275097846985, | |
| "eval_runtime": 5.0539, | |
| "eval_samples_per_second": 56.59, | |
| "eval_steps_per_second": 3.562, | |
| "step": 1342 | |
| }, | |
| { | |
| "epoch": 92.0, | |
| "grad_norm": 2.104257822036743, | |
| "learning_rate": 2.040952380952381e-05, | |
| "loss": 0.4403, | |
| "step": 1357 | |
| }, | |
| { | |
| "epoch": 92.0, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7788061499595642, | |
| "eval_runtime": 4.9104, | |
| "eval_samples_per_second": 58.244, | |
| "eval_steps_per_second": 3.666, | |
| "step": 1357 | |
| }, | |
| { | |
| "epoch": 92.95, | |
| "grad_norm": 1.6670126914978027, | |
| "learning_rate": 2.0276190476190475e-05, | |
| "loss": 0.4567, | |
| "step": 1371 | |
| }, | |
| { | |
| "epoch": 92.95, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7926862835884094, | |
| "eval_runtime": 5.6039, | |
| "eval_samples_per_second": 51.036, | |
| "eval_steps_per_second": 3.212, | |
| "step": 1371 | |
| }, | |
| { | |
| "epoch": 93.97, | |
| "grad_norm": 0.9627218842506409, | |
| "learning_rate": 2.0133333333333333e-05, | |
| "loss": 0.4233, | |
| "step": 1386 | |
| }, | |
| { | |
| "epoch": 93.97, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7885140180587769, | |
| "eval_runtime": 4.3051, | |
| "eval_samples_per_second": 66.433, | |
| "eval_steps_per_second": 4.181, | |
| "step": 1386 | |
| }, | |
| { | |
| "epoch": 94.98, | |
| "grad_norm": 1.5669583082199097, | |
| "learning_rate": 1.999047619047619e-05, | |
| "loss": 0.4347, | |
| "step": 1401 | |
| }, | |
| { | |
| "epoch": 94.98, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7849389314651489, | |
| "eval_runtime": 4.7453, | |
| "eval_samples_per_second": 60.27, | |
| "eval_steps_per_second": 3.793, | |
| "step": 1401 | |
| }, | |
| { | |
| "epoch": 96.0, | |
| "grad_norm": 1.5593161582946777, | |
| "learning_rate": 1.984761904761905e-05, | |
| "loss": 0.4167, | |
| "step": 1416 | |
| }, | |
| { | |
| "epoch": 96.0, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7880135774612427, | |
| "eval_runtime": 5.155, | |
| "eval_samples_per_second": 55.48, | |
| "eval_steps_per_second": 3.492, | |
| "step": 1416 | |
| }, | |
| { | |
| "epoch": 96.95, | |
| "grad_norm": 1.337389349937439, | |
| "learning_rate": 1.9714285714285714e-05, | |
| "loss": 0.4394, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 96.95, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7888504862785339, | |
| "eval_runtime": 4.6698, | |
| "eval_samples_per_second": 61.244, | |
| "eval_steps_per_second": 3.855, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 97.97, | |
| "grad_norm": 1.8546737432479858, | |
| "learning_rate": 1.9571428571428572e-05, | |
| "loss": 0.4359, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 97.97, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.778459370136261, | |
| "eval_runtime": 4.7578, | |
| "eval_samples_per_second": 60.112, | |
| "eval_steps_per_second": 3.783, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 98.98, | |
| "grad_norm": 1.3580291271209717, | |
| "learning_rate": 1.942857142857143e-05, | |
| "loss": 0.4085, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 98.98, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7851797342300415, | |
| "eval_runtime": 5.8892, | |
| "eval_samples_per_second": 48.564, | |
| "eval_steps_per_second": 3.056, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 100.0, | |
| "grad_norm": 1.4461547136306763, | |
| "learning_rate": 1.928571428571429e-05, | |
| "loss": 0.3965, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 100.0, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7784654498100281, | |
| "eval_runtime": 5.3557, | |
| "eval_samples_per_second": 53.401, | |
| "eval_steps_per_second": 3.361, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 100.95, | |
| "grad_norm": 1.765655755996704, | |
| "learning_rate": 1.9152380952380954e-05, | |
| "loss": 0.445, | |
| "step": 1489 | |
| }, | |
| { | |
| "epoch": 100.95, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7826104760169983, | |
| "eval_runtime": 5.628, | |
| "eval_samples_per_second": 50.818, | |
| "eval_steps_per_second": 3.198, | |
| "step": 1489 | |
| }, | |
| { | |
| "epoch": 101.97, | |
| "grad_norm": 1.7069743871688843, | |
| "learning_rate": 1.9009523809523812e-05, | |
| "loss": 0.3988, | |
| "step": 1504 | |
| }, | |
| { | |
| "epoch": 101.97, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.8045200109481812, | |
| "eval_runtime": 5.2204, | |
| "eval_samples_per_second": 54.785, | |
| "eval_steps_per_second": 3.448, | |
| "step": 1504 | |
| }, | |
| { | |
| "epoch": 102.98, | |
| "grad_norm": 1.2846873998641968, | |
| "learning_rate": 1.886666666666667e-05, | |
| "loss": 0.4129, | |
| "step": 1519 | |
| }, | |
| { | |
| "epoch": 102.98, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7685949206352234, | |
| "eval_runtime": 5.3375, | |
| "eval_samples_per_second": 53.583, | |
| "eval_steps_per_second": 3.372, | |
| "step": 1519 | |
| }, | |
| { | |
| "epoch": 104.0, | |
| "grad_norm": 1.9040664434432983, | |
| "learning_rate": 1.872380952380952e-05, | |
| "loss": 0.3937, | |
| "step": 1534 | |
| }, | |
| { | |
| "epoch": 104.0, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7911521196365356, | |
| "eval_runtime": 5.5547, | |
| "eval_samples_per_second": 51.488, | |
| "eval_steps_per_second": 3.24, | |
| "step": 1534 | |
| }, | |
| { | |
| "epoch": 104.95, | |
| "grad_norm": 1.998403549194336, | |
| "learning_rate": 1.8590476190476194e-05, | |
| "loss": 0.4356, | |
| "step": 1548 | |
| }, | |
| { | |
| "epoch": 104.95, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7922284007072449, | |
| "eval_runtime": 5.247, | |
| "eval_samples_per_second": 54.507, | |
| "eval_steps_per_second": 3.431, | |
| "step": 1548 | |
| }, | |
| { | |
| "epoch": 105.97, | |
| "grad_norm": 1.3670754432678223, | |
| "learning_rate": 1.8447619047619045e-05, | |
| "loss": 0.3969, | |
| "step": 1563 | |
| }, | |
| { | |
| "epoch": 105.97, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7751501202583313, | |
| "eval_runtime": 4.5814, | |
| "eval_samples_per_second": 62.427, | |
| "eval_steps_per_second": 3.929, | |
| "step": 1563 | |
| }, | |
| { | |
| "epoch": 106.98, | |
| "grad_norm": 1.0057995319366455, | |
| "learning_rate": 1.8304761904761903e-05, | |
| "loss": 0.4051, | |
| "step": 1578 | |
| }, | |
| { | |
| "epoch": 106.98, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7917311787605286, | |
| "eval_runtime": 5.1261, | |
| "eval_samples_per_second": 55.793, | |
| "eval_steps_per_second": 3.511, | |
| "step": 1578 | |
| }, | |
| { | |
| "epoch": 108.0, | |
| "grad_norm": 1.7176306247711182, | |
| "learning_rate": 1.816190476190476e-05, | |
| "loss": 0.3982, | |
| "step": 1593 | |
| }, | |
| { | |
| "epoch": 108.0, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.7916986346244812, | |
| "eval_runtime": 6.2925, | |
| "eval_samples_per_second": 45.451, | |
| "eval_steps_per_second": 2.861, | |
| "step": 1593 | |
| }, | |
| { | |
| "epoch": 108.95, | |
| "grad_norm": 1.6695342063903809, | |
| "learning_rate": 1.8028571428571427e-05, | |
| "loss": 0.4117, | |
| "step": 1607 | |
| }, | |
| { | |
| "epoch": 108.95, | |
| "eval_accuracy": 0.7062937062937062, | |
| "eval_loss": 0.8070589303970337, | |
| "eval_runtime": 4.8482, | |
| "eval_samples_per_second": 58.99, | |
| "eval_steps_per_second": 3.713, | |
| "step": 1607 | |
| }, | |
| { | |
| "epoch": 109.97, | |
| "grad_norm": 1.1871509552001953, | |
| "learning_rate": 1.7885714285714285e-05, | |
| "loss": 0.3666, | |
| "step": 1622 | |
| }, | |
| { | |
| "epoch": 109.97, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7839590907096863, | |
| "eval_runtime": 5.4971, | |
| "eval_samples_per_second": 52.027, | |
| "eval_steps_per_second": 3.274, | |
| "step": 1622 | |
| }, | |
| { | |
| "epoch": 110.98, | |
| "grad_norm": 2.196869134902954, | |
| "learning_rate": 1.7742857142857143e-05, | |
| "loss": 0.3894, | |
| "step": 1637 | |
| }, | |
| { | |
| "epoch": 110.98, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7790002226829529, | |
| "eval_runtime": 6.5509, | |
| "eval_samples_per_second": 43.658, | |
| "eval_steps_per_second": 2.748, | |
| "step": 1637 | |
| }, | |
| { | |
| "epoch": 112.0, | |
| "grad_norm": 2.341435432434082, | |
| "learning_rate": 1.76e-05, | |
| "loss": 0.3858, | |
| "step": 1652 | |
| }, | |
| { | |
| "epoch": 112.0, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.7961041927337646, | |
| "eval_runtime": 4.8263, | |
| "eval_samples_per_second": 59.258, | |
| "eval_steps_per_second": 3.73, | |
| "step": 1652 | |
| }, | |
| { | |
| "epoch": 112.95, | |
| "grad_norm": 1.5795267820358276, | |
| "learning_rate": 1.7466666666666667e-05, | |
| "loss": 0.4037, | |
| "step": 1666 | |
| }, | |
| { | |
| "epoch": 112.95, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7822238802909851, | |
| "eval_runtime": 5.6008, | |
| "eval_samples_per_second": 51.064, | |
| "eval_steps_per_second": 3.214, | |
| "step": 1666 | |
| }, | |
| { | |
| "epoch": 113.97, | |
| "grad_norm": 1.5341583490371704, | |
| "learning_rate": 1.7333333333333332e-05, | |
| "loss": 0.3886, | |
| "step": 1681 | |
| }, | |
| { | |
| "epoch": 113.97, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7748440504074097, | |
| "eval_runtime": 4.7211, | |
| "eval_samples_per_second": 60.579, | |
| "eval_steps_per_second": 3.813, | |
| "step": 1681 | |
| }, | |
| { | |
| "epoch": 114.98, | |
| "grad_norm": 1.056776523590088, | |
| "learning_rate": 1.719047619047619e-05, | |
| "loss": 0.3762, | |
| "step": 1696 | |
| }, | |
| { | |
| "epoch": 114.98, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7781729102134705, | |
| "eval_runtime": 6.5039, | |
| "eval_samples_per_second": 43.974, | |
| "eval_steps_per_second": 2.768, | |
| "step": 1696 | |
| }, | |
| { | |
| "epoch": 116.0, | |
| "grad_norm": 2.7518413066864014, | |
| "learning_rate": 1.704761904761905e-05, | |
| "loss": 0.3444, | |
| "step": 1711 | |
| }, | |
| { | |
| "epoch": 116.0, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.7745847702026367, | |
| "eval_runtime": 6.1348, | |
| "eval_samples_per_second": 46.619, | |
| "eval_steps_per_second": 2.934, | |
| "step": 1711 | |
| }, | |
| { | |
| "epoch": 116.95, | |
| "grad_norm": 1.8518308401107788, | |
| "learning_rate": 1.6914285714285714e-05, | |
| "loss": 0.3961, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 116.95, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7842003703117371, | |
| "eval_runtime": 4.8863, | |
| "eval_samples_per_second": 58.531, | |
| "eval_steps_per_second": 3.684, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 117.97, | |
| "grad_norm": 1.487108588218689, | |
| "learning_rate": 1.6771428571428572e-05, | |
| "loss": 0.3578, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 117.97, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7819164991378784, | |
| "eval_runtime": 5.6099, | |
| "eval_samples_per_second": 50.981, | |
| "eval_steps_per_second": 3.209, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 118.98, | |
| "grad_norm": 1.7816615104675293, | |
| "learning_rate": 1.662857142857143e-05, | |
| "loss": 0.3578, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 118.98, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7805832624435425, | |
| "eval_runtime": 5.5595, | |
| "eval_samples_per_second": 51.443, | |
| "eval_steps_per_second": 3.238, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 120.0, | |
| "grad_norm": 2.4715042114257812, | |
| "learning_rate": 1.6485714285714285e-05, | |
| "loss": 0.3489, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 120.0, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7809211015701294, | |
| "eval_runtime": 6.1205, | |
| "eval_samples_per_second": 46.728, | |
| "eval_steps_per_second": 2.941, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 120.95, | |
| "grad_norm": 1.1066619157791138, | |
| "learning_rate": 1.635238095238095e-05, | |
| "loss": 0.3622, | |
| "step": 1784 | |
| }, | |
| { | |
| "epoch": 120.95, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.7947035431861877, | |
| "eval_runtime": 5.0722, | |
| "eval_samples_per_second": 56.386, | |
| "eval_steps_per_second": 3.549, | |
| "step": 1784 | |
| }, | |
| { | |
| "epoch": 121.97, | |
| "grad_norm": 1.3460161685943604, | |
| "learning_rate": 1.620952380952381e-05, | |
| "loss": 0.3545, | |
| "step": 1799 | |
| }, | |
| { | |
| "epoch": 121.97, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7877638339996338, | |
| "eval_runtime": 5.4395, | |
| "eval_samples_per_second": 52.578, | |
| "eval_steps_per_second": 3.309, | |
| "step": 1799 | |
| }, | |
| { | |
| "epoch": 122.98, | |
| "grad_norm": 1.7303364276885986, | |
| "learning_rate": 1.6066666666666666e-05, | |
| "loss": 0.3361, | |
| "step": 1814 | |
| }, | |
| { | |
| "epoch": 122.98, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7854802012443542, | |
| "eval_runtime": 4.9544, | |
| "eval_samples_per_second": 57.726, | |
| "eval_steps_per_second": 3.633, | |
| "step": 1814 | |
| }, | |
| { | |
| "epoch": 124.0, | |
| "grad_norm": 1.5440623760223389, | |
| "learning_rate": 1.5923809523809524e-05, | |
| "loss": 0.3618, | |
| "step": 1829 | |
| }, | |
| { | |
| "epoch": 124.0, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7890083193778992, | |
| "eval_runtime": 4.6202, | |
| "eval_samples_per_second": 61.902, | |
| "eval_steps_per_second": 3.896, | |
| "step": 1829 | |
| }, | |
| { | |
| "epoch": 124.95, | |
| "grad_norm": 1.2985795736312866, | |
| "learning_rate": 1.579047619047619e-05, | |
| "loss": 0.3472, | |
| "step": 1843 | |
| }, | |
| { | |
| "epoch": 124.95, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7809839844703674, | |
| "eval_runtime": 4.5623, | |
| "eval_samples_per_second": 62.687, | |
| "eval_steps_per_second": 3.945, | |
| "step": 1843 | |
| }, | |
| { | |
| "epoch": 125.97, | |
| "grad_norm": 1.3930552005767822, | |
| "learning_rate": 1.5647619047619048e-05, | |
| "loss": 0.3511, | |
| "step": 1858 | |
| }, | |
| { | |
| "epoch": 125.97, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7897189259529114, | |
| "eval_runtime": 5.1961, | |
| "eval_samples_per_second": 55.041, | |
| "eval_steps_per_second": 3.464, | |
| "step": 1858 | |
| }, | |
| { | |
| "epoch": 126.98, | |
| "grad_norm": 2.0258781909942627, | |
| "learning_rate": 1.5504761904761906e-05, | |
| "loss": 0.3389, | |
| "step": 1873 | |
| }, | |
| { | |
| "epoch": 126.98, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7922906875610352, | |
| "eval_runtime": 5.0606, | |
| "eval_samples_per_second": 56.515, | |
| "eval_steps_per_second": 3.557, | |
| "step": 1873 | |
| }, | |
| { | |
| "epoch": 128.0, | |
| "grad_norm": 1.0623548030853271, | |
| "learning_rate": 1.5361904761904764e-05, | |
| "loss": 0.3391, | |
| "step": 1888 | |
| }, | |
| { | |
| "epoch": 128.0, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7781790494918823, | |
| "eval_runtime": 4.5848, | |
| "eval_samples_per_second": 62.38, | |
| "eval_steps_per_second": 3.926, | |
| "step": 1888 | |
| }, | |
| { | |
| "epoch": 128.95, | |
| "grad_norm": 1.2839761972427368, | |
| "learning_rate": 1.522857142857143e-05, | |
| "loss": 0.3746, | |
| "step": 1902 | |
| }, | |
| { | |
| "epoch": 128.95, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7838397026062012, | |
| "eval_runtime": 6.371, | |
| "eval_samples_per_second": 44.891, | |
| "eval_steps_per_second": 2.825, | |
| "step": 1902 | |
| }, | |
| { | |
| "epoch": 129.97, | |
| "grad_norm": 1.7424538135528564, | |
| "learning_rate": 1.5085714285714288e-05, | |
| "loss": 0.3238, | |
| "step": 1917 | |
| }, | |
| { | |
| "epoch": 129.97, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7943305969238281, | |
| "eval_runtime": 5.0078, | |
| "eval_samples_per_second": 57.111, | |
| "eval_steps_per_second": 3.594, | |
| "step": 1917 | |
| }, | |
| { | |
| "epoch": 130.98, | |
| "grad_norm": 1.9641584157943726, | |
| "learning_rate": 1.4942857142857143e-05, | |
| "loss": 0.3601, | |
| "step": 1932 | |
| }, | |
| { | |
| "epoch": 130.98, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.786338746547699, | |
| "eval_runtime": 5.7221, | |
| "eval_samples_per_second": 49.982, | |
| "eval_steps_per_second": 3.146, | |
| "step": 1932 | |
| }, | |
| { | |
| "epoch": 132.0, | |
| "grad_norm": 1.6700012683868408, | |
| "learning_rate": 1.48e-05, | |
| "loss": 0.3339, | |
| "step": 1947 | |
| }, | |
| { | |
| "epoch": 132.0, | |
| "eval_accuracy": 0.7132867132867133, | |
| "eval_loss": 0.7948570847511292, | |
| "eval_runtime": 4.5956, | |
| "eval_samples_per_second": 62.234, | |
| "eval_steps_per_second": 3.917, | |
| "step": 1947 | |
| }, | |
| { | |
| "epoch": 132.95, | |
| "grad_norm": 1.4314067363739014, | |
| "learning_rate": 1.4666666666666666e-05, | |
| "loss": 0.3805, | |
| "step": 1961 | |
| }, | |
| { | |
| "epoch": 132.95, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7823219299316406, | |
| "eval_runtime": 5.3022, | |
| "eval_samples_per_second": 53.94, | |
| "eval_steps_per_second": 3.395, | |
| "step": 1961 | |
| }, | |
| { | |
| "epoch": 133.97, | |
| "grad_norm": 1.9961072206497192, | |
| "learning_rate": 1.4523809523809524e-05, | |
| "loss": 0.3524, | |
| "step": 1976 | |
| }, | |
| { | |
| "epoch": 133.97, | |
| "eval_accuracy": 0.7097902097902098, | |
| "eval_loss": 0.8052372932434082, | |
| "eval_runtime": 4.8993, | |
| "eval_samples_per_second": 58.375, | |
| "eval_steps_per_second": 3.674, | |
| "step": 1976 | |
| }, | |
| { | |
| "epoch": 134.98, | |
| "grad_norm": 1.1320672035217285, | |
| "learning_rate": 1.4380952380952382e-05, | |
| "loss": 0.3103, | |
| "step": 1991 | |
| }, | |
| { | |
| "epoch": 134.98, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7809023261070251, | |
| "eval_runtime": 5.8892, | |
| "eval_samples_per_second": 48.563, | |
| "eval_steps_per_second": 3.056, | |
| "step": 1991 | |
| }, | |
| { | |
| "epoch": 136.0, | |
| "grad_norm": 2.298990249633789, | |
| "learning_rate": 1.4238095238095239e-05, | |
| "loss": 0.3484, | |
| "step": 2006 | |
| }, | |
| { | |
| "epoch": 136.0, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.787878692150116, | |
| "eval_runtime": 5.2761, | |
| "eval_samples_per_second": 54.207, | |
| "eval_steps_per_second": 3.412, | |
| "step": 2006 | |
| }, | |
| { | |
| "epoch": 136.95, | |
| "grad_norm": 2.3017499446868896, | |
| "learning_rate": 1.4104761904761906e-05, | |
| "loss": 0.3424, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 136.95, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7875267267227173, | |
| "eval_runtime": 4.7804, | |
| "eval_samples_per_second": 59.828, | |
| "eval_steps_per_second": 3.765, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 137.97, | |
| "grad_norm": 1.4992213249206543, | |
| "learning_rate": 1.3961904761904762e-05, | |
| "loss": 0.316, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 137.97, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7829337120056152, | |
| "eval_runtime": 4.9522, | |
| "eval_samples_per_second": 57.752, | |
| "eval_steps_per_second": 3.635, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 138.98, | |
| "grad_norm": 1.634419560432434, | |
| "learning_rate": 1.3819047619047619e-05, | |
| "loss": 0.3171, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 138.98, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7882408499717712, | |
| "eval_runtime": 4.8523, | |
| "eval_samples_per_second": 58.941, | |
| "eval_steps_per_second": 3.71, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 140.0, | |
| "grad_norm": 1.2980549335479736, | |
| "learning_rate": 1.3676190476190477e-05, | |
| "loss": 0.3155, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 140.0, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7830457091331482, | |
| "eval_runtime": 4.9611, | |
| "eval_samples_per_second": 57.649, | |
| "eval_steps_per_second": 3.628, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 140.95, | |
| "grad_norm": 1.5925828218460083, | |
| "learning_rate": 1.3542857142857142e-05, | |
| "loss": 0.3382, | |
| "step": 2079 | |
| }, | |
| { | |
| "epoch": 140.95, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7825762033462524, | |
| "eval_runtime": 4.6558, | |
| "eval_samples_per_second": 61.428, | |
| "eval_steps_per_second": 3.866, | |
| "step": 2079 | |
| }, | |
| { | |
| "epoch": 141.97, | |
| "grad_norm": 1.9007666110992432, | |
| "learning_rate": 1.34e-05, | |
| "loss": 0.3175, | |
| "step": 2094 | |
| }, | |
| { | |
| "epoch": 141.97, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7963916659355164, | |
| "eval_runtime": 5.2803, | |
| "eval_samples_per_second": 54.164, | |
| "eval_steps_per_second": 3.409, | |
| "step": 2094 | |
| }, | |
| { | |
| "epoch": 142.98, | |
| "grad_norm": 1.8862413167953491, | |
| "learning_rate": 1.3257142857142858e-05, | |
| "loss": 0.3444, | |
| "step": 2109 | |
| }, | |
| { | |
| "epoch": 142.98, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7858501672744751, | |
| "eval_runtime": 5.4501, | |
| "eval_samples_per_second": 52.476, | |
| "eval_steps_per_second": 3.303, | |
| "step": 2109 | |
| }, | |
| { | |
| "epoch": 144.0, | |
| "grad_norm": 2.4451475143432617, | |
| "learning_rate": 1.3114285714285715e-05, | |
| "loss": 0.3208, | |
| "step": 2124 | |
| }, | |
| { | |
| "epoch": 144.0, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7859659194946289, | |
| "eval_runtime": 4.5157, | |
| "eval_samples_per_second": 63.334, | |
| "eval_steps_per_second": 3.986, | |
| "step": 2124 | |
| }, | |
| { | |
| "epoch": 144.95, | |
| "grad_norm": 1.6911462545394897, | |
| "learning_rate": 1.298095238095238e-05, | |
| "loss": 0.3286, | |
| "step": 2138 | |
| }, | |
| { | |
| "epoch": 144.95, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7869133353233337, | |
| "eval_runtime": 5.6207, | |
| "eval_samples_per_second": 50.883, | |
| "eval_steps_per_second": 3.202, | |
| "step": 2138 | |
| }, | |
| { | |
| "epoch": 145.97, | |
| "grad_norm": 1.9792907238006592, | |
| "learning_rate": 1.2838095238095239e-05, | |
| "loss": 0.3319, | |
| "step": 2153 | |
| }, | |
| { | |
| "epoch": 145.97, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7916193604469299, | |
| "eval_runtime": 5.434, | |
| "eval_samples_per_second": 52.631, | |
| "eval_steps_per_second": 3.312, | |
| "step": 2153 | |
| }, | |
| { | |
| "epoch": 146.98, | |
| "grad_norm": 1.3403879404067993, | |
| "learning_rate": 1.2695238095238095e-05, | |
| "loss": 0.2954, | |
| "step": 2168 | |
| }, | |
| { | |
| "epoch": 146.98, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7937904000282288, | |
| "eval_runtime": 4.5934, | |
| "eval_samples_per_second": 62.264, | |
| "eval_steps_per_second": 3.919, | |
| "step": 2168 | |
| }, | |
| { | |
| "epoch": 148.0, | |
| "grad_norm": 2.6414599418640137, | |
| "learning_rate": 1.2552380952380953e-05, | |
| "loss": 0.3283, | |
| "step": 2183 | |
| }, | |
| { | |
| "epoch": 148.0, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7974053025245667, | |
| "eval_runtime": 4.6622, | |
| "eval_samples_per_second": 61.344, | |
| "eval_steps_per_second": 3.861, | |
| "step": 2183 | |
| }, | |
| { | |
| "epoch": 148.95, | |
| "grad_norm": 1.7404941320419312, | |
| "learning_rate": 1.241904761904762e-05, | |
| "loss": 0.3306, | |
| "step": 2197 | |
| }, | |
| { | |
| "epoch": 148.95, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.7795438766479492, | |
| "eval_runtime": 4.8202, | |
| "eval_samples_per_second": 59.334, | |
| "eval_steps_per_second": 3.734, | |
| "step": 2197 | |
| }, | |
| { | |
| "epoch": 149.97, | |
| "grad_norm": 1.3558377027511597, | |
| "learning_rate": 1.2276190476190477e-05, | |
| "loss": 0.3073, | |
| "step": 2212 | |
| }, | |
| { | |
| "epoch": 149.97, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7910019755363464, | |
| "eval_runtime": 5.1476, | |
| "eval_samples_per_second": 55.56, | |
| "eval_steps_per_second": 3.497, | |
| "step": 2212 | |
| }, | |
| { | |
| "epoch": 150.98, | |
| "grad_norm": 1.2379992008209229, | |
| "learning_rate": 1.2133333333333333e-05, | |
| "loss": 0.3089, | |
| "step": 2227 | |
| }, | |
| { | |
| "epoch": 150.98, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7942177653312683, | |
| "eval_runtime": 5.4133, | |
| "eval_samples_per_second": 52.833, | |
| "eval_steps_per_second": 3.325, | |
| "step": 2227 | |
| }, | |
| { | |
| "epoch": 152.0, | |
| "grad_norm": 1.3452534675598145, | |
| "learning_rate": 1.1990476190476191e-05, | |
| "loss": 0.2915, | |
| "step": 2242 | |
| }, | |
| { | |
| "epoch": 152.0, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7933531403541565, | |
| "eval_runtime": 6.3553, | |
| "eval_samples_per_second": 45.002, | |
| "eval_steps_per_second": 2.832, | |
| "step": 2242 | |
| }, | |
| { | |
| "epoch": 152.95, | |
| "grad_norm": 0.9324106574058533, | |
| "learning_rate": 1.1857142857142857e-05, | |
| "loss": 0.3286, | |
| "step": 2256 | |
| }, | |
| { | |
| "epoch": 152.95, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.7807846069335938, | |
| "eval_runtime": 5.094, | |
| "eval_samples_per_second": 56.144, | |
| "eval_steps_per_second": 3.534, | |
| "step": 2256 | |
| }, | |
| { | |
| "epoch": 153.97, | |
| "grad_norm": 1.5133529901504517, | |
| "learning_rate": 1.1714285714285715e-05, | |
| "loss": 0.2817, | |
| "step": 2271 | |
| }, | |
| { | |
| "epoch": 153.97, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.7788205742835999, | |
| "eval_runtime": 5.0531, | |
| "eval_samples_per_second": 56.599, | |
| "eval_steps_per_second": 3.562, | |
| "step": 2271 | |
| }, | |
| { | |
| "epoch": 154.98, | |
| "grad_norm": 2.8459701538085938, | |
| "learning_rate": 1.1571428571428573e-05, | |
| "loss": 0.3118, | |
| "step": 2286 | |
| }, | |
| { | |
| "epoch": 154.98, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7898407578468323, | |
| "eval_runtime": 4.8393, | |
| "eval_samples_per_second": 59.1, | |
| "eval_steps_per_second": 3.72, | |
| "step": 2286 | |
| }, | |
| { | |
| "epoch": 156.0, | |
| "grad_norm": 1.8526194095611572, | |
| "learning_rate": 1.1428571428571429e-05, | |
| "loss": 0.3155, | |
| "step": 2301 | |
| }, | |
| { | |
| "epoch": 156.0, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.796588122844696, | |
| "eval_runtime": 5.8687, | |
| "eval_samples_per_second": 48.733, | |
| "eval_steps_per_second": 3.067, | |
| "step": 2301 | |
| }, | |
| { | |
| "epoch": 156.95, | |
| "grad_norm": 1.130321741104126, | |
| "learning_rate": 1.1295238095238096e-05, | |
| "loss": 0.3156, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 156.95, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7947112321853638, | |
| "eval_runtime": 4.5714, | |
| "eval_samples_per_second": 62.564, | |
| "eval_steps_per_second": 3.938, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 157.97, | |
| "grad_norm": 1.3919442892074585, | |
| "learning_rate": 1.1152380952380953e-05, | |
| "loss": 0.2936, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 157.97, | |
| "eval_accuracy": 0.7167832167832168, | |
| "eval_loss": 0.7916660904884338, | |
| "eval_runtime": 4.554, | |
| "eval_samples_per_second": 62.802, | |
| "eval_steps_per_second": 3.953, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 158.98, | |
| "grad_norm": 2.1628525257110596, | |
| "learning_rate": 1.1009523809523809e-05, | |
| "loss": 0.3049, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 158.98, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.7779849171638489, | |
| "eval_runtime": 5.5631, | |
| "eval_samples_per_second": 51.41, | |
| "eval_steps_per_second": 3.236, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 160.0, | |
| "grad_norm": 1.0748810768127441, | |
| "learning_rate": 1.0866666666666667e-05, | |
| "loss": 0.2896, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 160.0, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.792616605758667, | |
| "eval_runtime": 5.3985, | |
| "eval_samples_per_second": 52.978, | |
| "eval_steps_per_second": 3.334, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 160.95, | |
| "grad_norm": 1.9327268600463867, | |
| "learning_rate": 1.0733333333333333e-05, | |
| "loss": 0.3194, | |
| "step": 2374 | |
| }, | |
| { | |
| "epoch": 160.95, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.8022683262825012, | |
| "eval_runtime": 5.5802, | |
| "eval_samples_per_second": 51.253, | |
| "eval_steps_per_second": 3.226, | |
| "step": 2374 | |
| }, | |
| { | |
| "epoch": 161.97, | |
| "grad_norm": 1.610187292098999, | |
| "learning_rate": 1.059047619047619e-05, | |
| "loss": 0.2918, | |
| "step": 2389 | |
| }, | |
| { | |
| "epoch": 161.97, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.7933264374732971, | |
| "eval_runtime": 5.0333, | |
| "eval_samples_per_second": 56.821, | |
| "eval_steps_per_second": 3.576, | |
| "step": 2389 | |
| }, | |
| { | |
| "epoch": 162.98, | |
| "grad_norm": 1.961053729057312, | |
| "learning_rate": 1.0447619047619049e-05, | |
| "loss": 0.2992, | |
| "step": 2404 | |
| }, | |
| { | |
| "epoch": 162.98, | |
| "eval_accuracy": 0.7412587412587412, | |
| "eval_loss": 0.7828559875488281, | |
| "eval_runtime": 5.4812, | |
| "eval_samples_per_second": 52.178, | |
| "eval_steps_per_second": 3.284, | |
| "step": 2404 | |
| }, | |
| { | |
| "epoch": 164.0, | |
| "grad_norm": 1.2120810747146606, | |
| "learning_rate": 1.0304761904761905e-05, | |
| "loss": 0.3, | |
| "step": 2419 | |
| }, | |
| { | |
| "epoch": 164.0, | |
| "eval_accuracy": 0.7202797202797203, | |
| "eval_loss": 0.7946493029594421, | |
| "eval_runtime": 5.351, | |
| "eval_samples_per_second": 53.448, | |
| "eval_steps_per_second": 3.364, | |
| "step": 2419 | |
| }, | |
| { | |
| "epoch": 164.95, | |
| "grad_norm": 1.112051010131836, | |
| "learning_rate": 1.0171428571428571e-05, | |
| "loss": 0.322, | |
| "step": 2433 | |
| }, | |
| { | |
| "epoch": 164.95, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.79693204164505, | |
| "eval_runtime": 6.2682, | |
| "eval_samples_per_second": 45.627, | |
| "eval_steps_per_second": 2.872, | |
| "step": 2433 | |
| }, | |
| { | |
| "epoch": 165.97, | |
| "grad_norm": 1.2547581195831299, | |
| "learning_rate": 1.0028571428571429e-05, | |
| "loss": 0.2994, | |
| "step": 2448 | |
| }, | |
| { | |
| "epoch": 165.97, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.8075766563415527, | |
| "eval_runtime": 5.4676, | |
| "eval_samples_per_second": 52.308, | |
| "eval_steps_per_second": 3.292, | |
| "step": 2448 | |
| }, | |
| { | |
| "epoch": 166.98, | |
| "grad_norm": 3.3027942180633545, | |
| "learning_rate": 9.885714285714285e-06, | |
| "loss": 0.2849, | |
| "step": 2463 | |
| }, | |
| { | |
| "epoch": 166.98, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.7950677275657654, | |
| "eval_runtime": 5.6445, | |
| "eval_samples_per_second": 50.668, | |
| "eval_steps_per_second": 3.189, | |
| "step": 2463 | |
| }, | |
| { | |
| "epoch": 168.0, | |
| "grad_norm": 3.2351794242858887, | |
| "learning_rate": 9.742857142857143e-06, | |
| "loss": 0.2745, | |
| "step": 2478 | |
| }, | |
| { | |
| "epoch": 168.0, | |
| "eval_accuracy": 0.7342657342657343, | |
| "eval_loss": 0.7892395853996277, | |
| "eval_runtime": 6.6609, | |
| "eval_samples_per_second": 42.937, | |
| "eval_steps_per_second": 2.702, | |
| "step": 2478 | |
| }, | |
| { | |
| "epoch": 168.95, | |
| "grad_norm": 2.2650949954986572, | |
| "learning_rate": 9.60952380952381e-06, | |
| "loss": 0.2974, | |
| "step": 2492 | |
| }, | |
| { | |
| "epoch": 168.95, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.791623055934906, | |
| "eval_runtime": 5.8468, | |
| "eval_samples_per_second": 48.916, | |
| "eval_steps_per_second": 3.079, | |
| "step": 2492 | |
| }, | |
| { | |
| "epoch": 169.97, | |
| "grad_norm": 1.3676173686981201, | |
| "learning_rate": 9.466666666666667e-06, | |
| "loss": 0.2656, | |
| "step": 2507 | |
| }, | |
| { | |
| "epoch": 169.97, | |
| "eval_accuracy": 0.7342657342657343, | |
| "eval_loss": 0.7994617223739624, | |
| "eval_runtime": 4.7953, | |
| "eval_samples_per_second": 59.642, | |
| "eval_steps_per_second": 3.754, | |
| "step": 2507 | |
| }, | |
| { | |
| "epoch": 170.98, | |
| "grad_norm": 1.5123904943466187, | |
| "learning_rate": 9.323809523809523e-06, | |
| "loss": 0.295, | |
| "step": 2522 | |
| }, | |
| { | |
| "epoch": 170.98, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.802618682384491, | |
| "eval_runtime": 6.3216, | |
| "eval_samples_per_second": 45.241, | |
| "eval_steps_per_second": 2.847, | |
| "step": 2522 | |
| }, | |
| { | |
| "epoch": 172.0, | |
| "grad_norm": 1.5210902690887451, | |
| "learning_rate": 9.180952380952381e-06, | |
| "loss": 0.2791, | |
| "step": 2537 | |
| }, | |
| { | |
| "epoch": 172.0, | |
| "eval_accuracy": 0.7342657342657343, | |
| "eval_loss": 0.797250509262085, | |
| "eval_runtime": 5.4447, | |
| "eval_samples_per_second": 52.528, | |
| "eval_steps_per_second": 3.306, | |
| "step": 2537 | |
| }, | |
| { | |
| "epoch": 172.95, | |
| "grad_norm": 1.0823432207107544, | |
| "learning_rate": 9.047619047619047e-06, | |
| "loss": 0.2836, | |
| "step": 2551 | |
| }, | |
| { | |
| "epoch": 172.95, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.8022569417953491, | |
| "eval_runtime": 5.0556, | |
| "eval_samples_per_second": 56.571, | |
| "eval_steps_per_second": 3.56, | |
| "step": 2551 | |
| }, | |
| { | |
| "epoch": 173.97, | |
| "grad_norm": 1.4100669622421265, | |
| "learning_rate": 8.904761904761905e-06, | |
| "loss": 0.2806, | |
| "step": 2566 | |
| }, | |
| { | |
| "epoch": 173.97, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.8013490438461304, | |
| "eval_runtime": 5.503, | |
| "eval_samples_per_second": 51.972, | |
| "eval_steps_per_second": 3.271, | |
| "step": 2566 | |
| }, | |
| { | |
| "epoch": 174.98, | |
| "grad_norm": 1.389672875404358, | |
| "learning_rate": 8.761904761904763e-06, | |
| "loss": 0.2661, | |
| "step": 2581 | |
| }, | |
| { | |
| "epoch": 174.98, | |
| "eval_accuracy": 0.7307692307692307, | |
| "eval_loss": 0.796493649482727, | |
| "eval_runtime": 5.7105, | |
| "eval_samples_per_second": 50.083, | |
| "eval_steps_per_second": 3.152, | |
| "step": 2581 | |
| }, | |
| { | |
| "epoch": 176.0, | |
| "grad_norm": 1.8702772855758667, | |
| "learning_rate": 8.61904761904762e-06, | |
| "loss": 0.2695, | |
| "step": 2596 | |
| }, | |
| { | |
| "epoch": 176.0, | |
| "eval_accuracy": 0.7272727272727273, | |
| "eval_loss": 0.8062567114830017, | |
| "eval_runtime": 7.451, | |
| "eval_samples_per_second": 38.384, | |
| "eval_steps_per_second": 2.416, | |
| "step": 2596 | |
| }, | |
| { | |
| "epoch": 176.95, | |
| "grad_norm": 3.131314992904663, | |
| "learning_rate": 8.485714285714287e-06, | |
| "loss": 0.286, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 176.95, | |
| "eval_accuracy": 0.7237762237762237, | |
| "eval_loss": 0.7963144779205322, | |
| "eval_runtime": 4.431, | |
| "eval_samples_per_second": 64.546, | |
| "eval_steps_per_second": 4.062, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 177.97, | |
| "grad_norm": 1.7147862911224365, | |
| "learning_rate": 8.342857142857143e-06, | |
| "loss": 0.2743, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 177.97, | |
| "eval_accuracy": 0.7412587412587412, | |
| "eval_loss": 0.7928534150123596, | |
| "eval_runtime": 5.7357, | |
| "eval_samples_per_second": 49.863, | |
| "eval_steps_per_second": 3.138, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 178.98, | |
| "grad_norm": 1.669243574142456, | |
| "learning_rate": 8.2e-06, | |
| "loss": 0.2775, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 178.98, | |
| "eval_accuracy": 0.7447552447552448, | |
| "eval_loss": 0.7855107188224792, | |
| "eval_runtime": 5.7848, | |
| "eval_samples_per_second": 49.44, | |
| "eval_steps_per_second": 3.112, | |
| "step": 2640 | |
| } | |
| ], | |
| "logging_steps": 100.0, | |
| "max_steps": 3500, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 250, | |
| "save_steps": 500, | |
| "total_flos": 1.462758962872032e+18, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |