| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 2.9984810126582278, | |
| "eval_steps": 500, | |
| "global_step": 2961, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.010126582278481013, | |
| "grad_norm": 2.9319232619463147, | |
| "learning_rate": 5e-06, | |
| "loss": 0.754, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.020253164556962026, | |
| "grad_norm": 2.3527842138237443, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6565, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.030379746835443037, | |
| "grad_norm": 2.218097297827581, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6343, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.04050632911392405, | |
| "grad_norm": 1.728885416020883, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6266, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.05063291139240506, | |
| "grad_norm": 1.7600150721761239, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6225, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.060759493670886074, | |
| "grad_norm": 2.036486573113631, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6185, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.07088607594936709, | |
| "grad_norm": 1.8472990337897077, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6137, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.0810126582278481, | |
| "grad_norm": 2.0268825324116926, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6035, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.09113924050632911, | |
| "grad_norm": 1.7502041363801772, | |
| "learning_rate": 5e-06, | |
| "loss": 0.6063, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.10126582278481013, | |
| "grad_norm": 1.7706135514845762, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5954, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.11139240506329114, | |
| "grad_norm": 1.6647485123629138, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5933, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.12151898734177215, | |
| "grad_norm": 2.299944971561371, | |
| "learning_rate": 5e-06, | |
| "loss": 0.598, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.13164556962025317, | |
| "grad_norm": 1.4323042297049071, | |
| "learning_rate": 5e-06, | |
| "loss": 0.597, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.14177215189873418, | |
| "grad_norm": 1.5735470897186044, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5859, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.1518987341772152, | |
| "grad_norm": 1.4396445107803788, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5833, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.1620253164556962, | |
| "grad_norm": 1.6048875360767698, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5926, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.17215189873417722, | |
| "grad_norm": 1.7629611019629856, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5878, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.18227848101265823, | |
| "grad_norm": 1.692513854306383, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5875, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.19240506329113924, | |
| "grad_norm": 2.4645583630800614, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5908, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.20253164556962025, | |
| "grad_norm": 1.6304964629686653, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5887, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.21265822784810126, | |
| "grad_norm": 2.117270092101592, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5872, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.22278481012658227, | |
| "grad_norm": 1.5677168838868496, | |
| "learning_rate": 5e-06, | |
| "loss": 0.575, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.23291139240506328, | |
| "grad_norm": 1.6515647911636342, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5747, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.2430379746835443, | |
| "grad_norm": 1.4834458457301771, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5834, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.25316455696202533, | |
| "grad_norm": 2.090691823913493, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5793, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.26329113924050634, | |
| "grad_norm": 1.4520944372541345, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5779, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.27341772151898736, | |
| "grad_norm": 1.7905780272726646, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5637, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.28354430379746837, | |
| "grad_norm": 1.6753642665165924, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5745, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.2936708860759494, | |
| "grad_norm": 1.8117100564747939, | |
| "learning_rate": 5e-06, | |
| "loss": 0.577, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.3037974683544304, | |
| "grad_norm": 1.5160896257954382, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5821, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.3139240506329114, | |
| "grad_norm": 1.5338602444115943, | |
| "learning_rate": 5e-06, | |
| "loss": 0.574, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.3240506329113924, | |
| "grad_norm": 1.5768816348730115, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5831, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.3341772151898734, | |
| "grad_norm": 1.5251244093370149, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5793, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.34430379746835443, | |
| "grad_norm": 1.5422210280401378, | |
| "learning_rate": 5e-06, | |
| "loss": 0.579, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.35443037974683544, | |
| "grad_norm": 1.4422891794630028, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5794, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.36455696202531646, | |
| "grad_norm": 1.778098375508798, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5702, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.37468354430379747, | |
| "grad_norm": 1.4276847831429058, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5762, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.3848101265822785, | |
| "grad_norm": 1.8350692685441172, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5742, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.3949367088607595, | |
| "grad_norm": 1.6654466906631886, | |
| "learning_rate": 5e-06, | |
| "loss": 0.57, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.4050632911392405, | |
| "grad_norm": 1.5625914451532021, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5688, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.4151898734177215, | |
| "grad_norm": 1.5962831101550994, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5567, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.4253164556962025, | |
| "grad_norm": 1.4805817323596187, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5723, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.43544303797468353, | |
| "grad_norm": 1.4164796466458769, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5712, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.44556962025316454, | |
| "grad_norm": 1.5450127362729664, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5704, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.45569620253164556, | |
| "grad_norm": 1.6570421627781717, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5729, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.46582278481012657, | |
| "grad_norm": 1.623306340780734, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5673, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.4759493670886076, | |
| "grad_norm": 1.747763610168544, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5609, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.4860759493670886, | |
| "grad_norm": 1.295527401866004, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5699, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.4962025316455696, | |
| "grad_norm": 1.3529876566902441, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5783, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.5063291139240507, | |
| "grad_norm": 1.4541590535564763, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5606, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.5164556962025316, | |
| "grad_norm": 1.5391153152763637, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5598, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.5265822784810127, | |
| "grad_norm": 1.3816263219270335, | |
| "learning_rate": 5e-06, | |
| "loss": 0.552, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.5367088607594936, | |
| "grad_norm": 1.407148137331511, | |
| "learning_rate": 5e-06, | |
| "loss": 0.561, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.5468354430379747, | |
| "grad_norm": 1.4939662242734717, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5691, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.5569620253164557, | |
| "grad_norm": 1.4769024813528058, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5628, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.5670886075949367, | |
| "grad_norm": 1.5234796484078978, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5571, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.5772151898734177, | |
| "grad_norm": 1.38748512162588, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5588, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.5873417721518988, | |
| "grad_norm": 1.6616670500045834, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5629, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.5974683544303797, | |
| "grad_norm": 1.3613770956091802, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5597, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.6075949367088608, | |
| "grad_norm": 1.9645875965732293, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5614, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.6177215189873417, | |
| "grad_norm": 1.3122962556953037, | |
| "learning_rate": 5e-06, | |
| "loss": 0.57, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.6278481012658228, | |
| "grad_norm": 1.5024516266743335, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5629, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.6379746835443038, | |
| "grad_norm": 1.3599801460050238, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5679, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.6481012658227848, | |
| "grad_norm": 1.3697118174446543, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5567, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.6582278481012658, | |
| "grad_norm": 1.447467777446668, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5614, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.6683544303797468, | |
| "grad_norm": 1.4332946464750023, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5629, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.6784810126582278, | |
| "grad_norm": 1.4245518863907969, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5607, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.6886075949367089, | |
| "grad_norm": 1.392960913673911, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5631, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.6987341772151898, | |
| "grad_norm": 1.3851040024314174, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5547, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.7088607594936709, | |
| "grad_norm": 1.5665731498555628, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5729, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.7189873417721518, | |
| "grad_norm": 1.3541133459928674, | |
| "learning_rate": 5e-06, | |
| "loss": 0.567, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.7291139240506329, | |
| "grad_norm": 1.3050422174455243, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5649, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.739240506329114, | |
| "grad_norm": 1.4474203168962876, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5587, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.7493670886075949, | |
| "grad_norm": 1.2849691486389958, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5584, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.759493670886076, | |
| "grad_norm": 1.4374182399089128, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5485, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.769620253164557, | |
| "grad_norm": 1.2784166011759992, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5619, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.779746835443038, | |
| "grad_norm": 1.338741085455322, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5625, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.789873417721519, | |
| "grad_norm": 1.3830345955439087, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5532, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 1.2560229071250701, | |
| "learning_rate": 5e-06, | |
| "loss": 0.557, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.810126582278481, | |
| "grad_norm": 1.3094948207554686, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5606, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.8202531645569621, | |
| "grad_norm": 1.2980406537156193, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5435, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.830379746835443, | |
| "grad_norm": 1.3336678028790407, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5644, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.8405063291139241, | |
| "grad_norm": 1.4362113888678505, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5615, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.850632911392405, | |
| "grad_norm": 1.5496595209680522, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5619, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.8607594936708861, | |
| "grad_norm": 1.4648845925364151, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5556, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.8708860759493671, | |
| "grad_norm": 1.3830147030582034, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5534, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.8810126582278481, | |
| "grad_norm": 1.3802375047462647, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5654, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.8911392405063291, | |
| "grad_norm": 1.665356350061181, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5542, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.9012658227848102, | |
| "grad_norm": 1.5275969827966436, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5492, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.9113924050632911, | |
| "grad_norm": 1.5257937868842386, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5509, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.9215189873417722, | |
| "grad_norm": 1.3059469433496236, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5563, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.9316455696202531, | |
| "grad_norm": 1.3687394115237168, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5527, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.9417721518987342, | |
| "grad_norm": 1.3468586235595001, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5594, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.9518987341772152, | |
| "grad_norm": 1.4377274430310882, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5512, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.9620253164556962, | |
| "grad_norm": 1.2680685522134791, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5486, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.9721518987341772, | |
| "grad_norm": 1.2807954706357672, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5566, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.9822784810126582, | |
| "grad_norm": 1.3423829666439555, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5485, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.9924050632911392, | |
| "grad_norm": 1.4283880670670543, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5515, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.999493670886076, | |
| "eval_loss": 0.13834980130195618, | |
| "eval_runtime": 507.9163, | |
| "eval_samples_per_second": 26.193, | |
| "eval_steps_per_second": 0.41, | |
| "step": 987 | |
| }, | |
| { | |
| "epoch": 1.0025316455696203, | |
| "grad_norm": 1.6161728864680243, | |
| "learning_rate": 5e-06, | |
| "loss": 0.5308, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.0126582278481013, | |
| "grad_norm": 1.5737881952906483, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4633, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.0227848101265822, | |
| "grad_norm": 1.5616941170427905, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4634, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.0329113924050632, | |
| "grad_norm": 1.5346519713859017, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4435, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.0430379746835443, | |
| "grad_norm": 1.3492138788127321, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4503, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.0531645569620254, | |
| "grad_norm": 1.4763337790140822, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4506, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.0632911392405062, | |
| "grad_norm": 1.4225225244349122, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4591, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.0734177215189873, | |
| "grad_norm": 1.377686472346139, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4498, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.0835443037974684, | |
| "grad_norm": 1.524899956605274, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4521, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.0936708860759494, | |
| "grad_norm": 1.4555671547809872, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4537, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.1037974683544305, | |
| "grad_norm": 1.352652609202294, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4591, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.1139240506329113, | |
| "grad_norm": 1.389780851883711, | |
| "learning_rate": 5e-06, | |
| "loss": 0.451, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.1240506329113924, | |
| "grad_norm": 1.5493151434769157, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4545, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.1341772151898735, | |
| "grad_norm": 1.4921163467682323, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4511, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.1443037974683543, | |
| "grad_norm": 1.3424106662166821, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4597, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.1544303797468354, | |
| "grad_norm": 1.4205445966678738, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4587, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.1645569620253164, | |
| "grad_norm": 1.3812137606947525, | |
| "learning_rate": 5e-06, | |
| "loss": 0.457, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.1746835443037975, | |
| "grad_norm": 1.504094206656827, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4582, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.1848101265822786, | |
| "grad_norm": 1.4384269712621547, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4478, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.1949367088607594, | |
| "grad_norm": 1.4604436825013358, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4561, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.2050632911392405, | |
| "grad_norm": 1.484357252258588, | |
| "learning_rate": 5e-06, | |
| "loss": 0.463, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.2151898734177216, | |
| "grad_norm": 1.3507024353159274, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4557, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.2253164556962026, | |
| "grad_norm": 1.5356768758566504, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4616, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.2354430379746835, | |
| "grad_norm": 1.507777364560312, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4492, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.2455696202531645, | |
| "grad_norm": 1.402742621286163, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4623, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.2556962025316456, | |
| "grad_norm": 1.4914361590080172, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4576, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.2658227848101267, | |
| "grad_norm": 1.4449809880641054, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4581, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.2759493670886077, | |
| "grad_norm": 1.401311614162818, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4613, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.2860759493670886, | |
| "grad_norm": 1.4035116916924748, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4546, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.2962025316455696, | |
| "grad_norm": 1.5360414127344846, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4638, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.3063291139240507, | |
| "grad_norm": 1.4721064580495236, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4605, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.3164556962025316, | |
| "grad_norm": 1.5018662441760837, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4656, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.3265822784810126, | |
| "grad_norm": 1.4057750381381058, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4546, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.3367088607594937, | |
| "grad_norm": 1.472547127671551, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4509, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.3468354430379748, | |
| "grad_norm": 1.378222631334251, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4592, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.3569620253164558, | |
| "grad_norm": 1.477645235480499, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4731, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.3670886075949367, | |
| "grad_norm": 1.4190131074856087, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4614, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.3772151898734177, | |
| "grad_norm": 1.598773468818255, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4642, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.3873417721518988, | |
| "grad_norm": 1.401701470107724, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4599, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.3974683544303796, | |
| "grad_norm": 1.5834771319486265, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4635, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.4075949367088607, | |
| "grad_norm": 1.398082340799174, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4581, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.4177215189873418, | |
| "grad_norm": 1.4006340071029044, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4602, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.4278481012658228, | |
| "grad_norm": 1.3336328229398797, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4602, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.437974683544304, | |
| "grad_norm": 1.3492817883075563, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4687, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.4481012658227848, | |
| "grad_norm": 1.342687219512621, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4651, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.4582278481012658, | |
| "grad_norm": 1.4950318280118493, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4643, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.4683544303797469, | |
| "grad_norm": 1.5341419482481558, | |
| "learning_rate": 5e-06, | |
| "loss": 0.467, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.4784810126582277, | |
| "grad_norm": 1.3873503711966886, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4692, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.4886075949367088, | |
| "grad_norm": 1.3838670814377412, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4606, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.4987341772151899, | |
| "grad_norm": 1.435159074036311, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4607, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.508860759493671, | |
| "grad_norm": 1.416449038632833, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4603, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.518987341772152, | |
| "grad_norm": 1.349053423409347, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4671, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.529113924050633, | |
| "grad_norm": 1.5418750945784405, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4642, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.539240506329114, | |
| "grad_norm": 1.5350888472278188, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4654, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.549367088607595, | |
| "grad_norm": 1.5918459039624637, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4623, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.5594936708860758, | |
| "grad_norm": 1.4044783253886717, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4644, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.5696202531645569, | |
| "grad_norm": 1.4733622287724428, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4699, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.579746835443038, | |
| "grad_norm": 1.4190534275279456, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4585, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.589873417721519, | |
| "grad_norm": 1.588541120324257, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4674, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 1.4489487698353771, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4659, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.6101265822784812, | |
| "grad_norm": 1.3499923545579962, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4582, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.620253164556962, | |
| "grad_norm": 1.3640430679010667, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4676, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.630379746835443, | |
| "grad_norm": 1.3690059243296218, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4612, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.640506329113924, | |
| "grad_norm": 1.2447374210211029, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4506, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.650632911392405, | |
| "grad_norm": 1.351219366349358, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4611, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.660759493670886, | |
| "grad_norm": 1.335122455907824, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4699, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.6708860759493671, | |
| "grad_norm": 1.5625860035007904, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4612, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.6810126582278482, | |
| "grad_norm": 1.6296205380869357, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4623, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.6911392405063292, | |
| "grad_norm": 1.4330065379895607, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4625, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.70126582278481, | |
| "grad_norm": 1.3278761181861125, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4627, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.7113924050632912, | |
| "grad_norm": 1.4304366765861536, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4646, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.721518987341772, | |
| "grad_norm": 1.3460103891901807, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4648, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.731645569620253, | |
| "grad_norm": 1.4336438757742112, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4708, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.7417721518987341, | |
| "grad_norm": 1.333163056745564, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4591, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.7518987341772152, | |
| "grad_norm": 1.3819280069426765, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4651, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.7620253164556963, | |
| "grad_norm": 1.413028622360716, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4714, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.7721518987341773, | |
| "grad_norm": 1.3989242299354576, | |
| "learning_rate": 5e-06, | |
| "loss": 0.466, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.7822784810126582, | |
| "grad_norm": 1.3712554248183217, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4727, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.7924050632911392, | |
| "grad_norm": 1.4780769754046013, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4679, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.80253164556962, | |
| "grad_norm": 1.4002831190634755, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4647, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.8126582278481012, | |
| "grad_norm": 1.3944675393169206, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4644, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.8227848101265822, | |
| "grad_norm": 1.4867966880006709, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4675, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.8329113924050633, | |
| "grad_norm": 1.404496416486814, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4624, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.8430379746835444, | |
| "grad_norm": 1.5305845929183253, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4739, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.8531645569620254, | |
| "grad_norm": 1.3084464311737185, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4655, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.8632911392405065, | |
| "grad_norm": 1.3059438675535409, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4678, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.8734177215189873, | |
| "grad_norm": 1.3867484618996055, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4609, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.8835443037974684, | |
| "grad_norm": 1.3420942943215763, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4663, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.8936708860759492, | |
| "grad_norm": 1.421266278043212, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4665, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.9037974683544303, | |
| "grad_norm": 1.36634296470369, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4676, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.9139240506329114, | |
| "grad_norm": 1.2659671091408768, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4669, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.9240506329113924, | |
| "grad_norm": 1.373264332933901, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4701, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.9341772151898735, | |
| "grad_norm": 1.3568414058185374, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4693, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.9443037974683546, | |
| "grad_norm": 1.3940261082697252, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4763, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.9544303797468354, | |
| "grad_norm": 1.3879188735035681, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4756, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.9645569620253165, | |
| "grad_norm": 1.363003506285662, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4667, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.9746835443037973, | |
| "grad_norm": 1.4533412320271968, | |
| "learning_rate": 5e-06, | |
| "loss": 0.469, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.9848101265822784, | |
| "grad_norm": 1.4479310447515883, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4748, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.9949367088607595, | |
| "grad_norm": 1.2966648783230335, | |
| "learning_rate": 5e-06, | |
| "loss": 0.468, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 0.1398227959871292, | |
| "eval_runtime": 506.8788, | |
| "eval_samples_per_second": 26.247, | |
| "eval_steps_per_second": 0.41, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 2.0050632911392405, | |
| "grad_norm": 3.06600016953371, | |
| "learning_rate": 5e-06, | |
| "loss": 0.4085, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 2.0151898734177216, | |
| "grad_norm": 2.0901467389116335, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3545, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 2.0253164556962027, | |
| "grad_norm": 1.7198026240231687, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3486, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 2.0354430379746837, | |
| "grad_norm": 1.6596559930998485, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3485, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 2.0455696202531644, | |
| "grad_norm": 1.6010412767976676, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3469, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 2.0556962025316454, | |
| "grad_norm": 1.619456570416184, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3412, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 2.0658227848101265, | |
| "grad_norm": 1.5477510460994646, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3499, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 2.0759493670886076, | |
| "grad_norm": 1.6995317518377548, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3431, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 2.0860759493670886, | |
| "grad_norm": 1.6672377260042517, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3509, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 2.0962025316455697, | |
| "grad_norm": 1.5137523020681678, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3485, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 2.1063291139240508, | |
| "grad_norm": 1.5728338684227074, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3414, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 2.116455696202532, | |
| "grad_norm": 1.7215854826795536, | |
| "learning_rate": 5e-06, | |
| "loss": 0.356, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 2.1265822784810124, | |
| "grad_norm": 1.547269577634038, | |
| "learning_rate": 5e-06, | |
| "loss": 0.351, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 2.1367088607594935, | |
| "grad_norm": 1.7846477440752144, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3475, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 2.1468354430379746, | |
| "grad_norm": 1.6766174751768645, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3507, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 2.1569620253164556, | |
| "grad_norm": 1.6129490070505184, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3544, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 2.1670886075949367, | |
| "grad_norm": 1.6006266055236678, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3514, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 2.1772151898734178, | |
| "grad_norm": 1.5978325737155568, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3525, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 2.187341772151899, | |
| "grad_norm": 1.5951449058616392, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3507, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 2.19746835443038, | |
| "grad_norm": 1.5513789073675182, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3496, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 2.207594936708861, | |
| "grad_norm": 1.5229143655144641, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3494, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 2.2177215189873416, | |
| "grad_norm": 1.660242717310412, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3553, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 2.2278481012658227, | |
| "grad_norm": 1.6739776038113652, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3607, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 2.2379746835443037, | |
| "grad_norm": 1.7005249152138244, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3486, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 2.248101265822785, | |
| "grad_norm": 1.6489218772469403, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3592, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 2.258227848101266, | |
| "grad_norm": 1.6209760997530658, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3584, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 2.268354430379747, | |
| "grad_norm": 1.6153685359109242, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3554, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 2.278481012658228, | |
| "grad_norm": 1.6268217139529644, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3542, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 2.2886075949367086, | |
| "grad_norm": 1.7752820541706038, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3621, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 2.2987341772151897, | |
| "grad_norm": 1.5709969247849427, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3524, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 2.3088607594936708, | |
| "grad_norm": 1.5534897904208944, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3563, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 2.318987341772152, | |
| "grad_norm": 1.6608945730503153, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3565, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 2.329113924050633, | |
| "grad_norm": 1.5801622846665264, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3622, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 2.339240506329114, | |
| "grad_norm": 1.5320500459364543, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3589, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 2.349367088607595, | |
| "grad_norm": 1.581287461651533, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3613, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 2.359493670886076, | |
| "grad_norm": 1.6038211850984352, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3617, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 2.369620253164557, | |
| "grad_norm": 1.625966530104118, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3598, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 2.379746835443038, | |
| "grad_norm": 1.7645641137037487, | |
| "learning_rate": 5e-06, | |
| "loss": 0.362, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 2.389873417721519, | |
| "grad_norm": 1.6940524616595496, | |
| "learning_rate": 5e-06, | |
| "loss": 0.361, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 1.5384052283713547, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3639, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 2.410126582278481, | |
| "grad_norm": 1.6483060393926068, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3586, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 2.420253164556962, | |
| "grad_norm": 1.6074760148392138, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3648, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 2.430379746835443, | |
| "grad_norm": 1.7080201956192906, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3554, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 2.440506329113924, | |
| "grad_norm": 1.625757072296565, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3531, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 2.4506329113924052, | |
| "grad_norm": 1.6544192748138193, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3659, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 2.460759493670886, | |
| "grad_norm": 1.6016056620993897, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3618, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 2.470886075949367, | |
| "grad_norm": 1.6466444516635712, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3651, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 2.481012658227848, | |
| "grad_norm": 1.6213376274576505, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3611, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 2.491139240506329, | |
| "grad_norm": 1.61939669911579, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3655, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 2.50126582278481, | |
| "grad_norm": 1.7185252999562315, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3717, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 2.511392405063291, | |
| "grad_norm": 1.5701430968600476, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3616, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 2.5215189873417723, | |
| "grad_norm": 1.6525891145092144, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3618, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 2.5316455696202533, | |
| "grad_norm": 1.6331648861432553, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3608, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.5417721518987344, | |
| "grad_norm": 1.5954836812756767, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3564, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 2.5518987341772155, | |
| "grad_norm": 1.6221389574344411, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3603, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 2.562025316455696, | |
| "grad_norm": 1.645557544125129, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3633, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 2.572151898734177, | |
| "grad_norm": 1.7704063332654312, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3658, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 2.5822784810126582, | |
| "grad_norm": 1.6779345444764124, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3706, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.5924050632911393, | |
| "grad_norm": 1.6024752118154975, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3618, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 2.6025316455696204, | |
| "grad_norm": 1.5688766989229237, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3719, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 2.6126582278481014, | |
| "grad_norm": 1.5972321010318007, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3695, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 2.622784810126582, | |
| "grad_norm": 1.7926779657765413, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3671, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 2.632911392405063, | |
| "grad_norm": 1.6497318559260037, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3648, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 2.643037974683544, | |
| "grad_norm": 1.825449432151336, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3675, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 2.6531645569620252, | |
| "grad_norm": 1.6048797855449353, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3638, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 2.6632911392405063, | |
| "grad_norm": 1.6207591062208428, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3683, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 2.6734177215189874, | |
| "grad_norm": 1.7272301549499494, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3685, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 2.6835443037974684, | |
| "grad_norm": 1.6364841258418936, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3684, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 2.6936708860759495, | |
| "grad_norm": 1.6937643730038003, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3723, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 2.7037974683544306, | |
| "grad_norm": 1.5774730979594618, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3671, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 2.7139240506329116, | |
| "grad_norm": 1.6768910733845062, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3715, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 2.7240506329113923, | |
| "grad_norm": 1.6389744367173145, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3674, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 2.7341772151898733, | |
| "grad_norm": 1.7387883655123013, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3701, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 2.7443037974683544, | |
| "grad_norm": 1.622855925296899, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3692, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 2.7544303797468355, | |
| "grad_norm": 1.6049744247340423, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3605, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 2.7645569620253165, | |
| "grad_norm": 1.674299981616671, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3681, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 2.7746835443037976, | |
| "grad_norm": 1.6563350827590924, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3732, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 2.7848101265822782, | |
| "grad_norm": 1.6562436415268564, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3666, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 2.7949367088607593, | |
| "grad_norm": 1.5565903819361853, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3706, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 2.8050632911392404, | |
| "grad_norm": 1.7142515863322454, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3679, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 2.8151898734177214, | |
| "grad_norm": 1.5712425643932983, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3761, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 2.8253164556962025, | |
| "grad_norm": 1.597329501161643, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3697, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 2.8354430379746836, | |
| "grad_norm": 1.6642023307988005, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3613, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 2.8455696202531646, | |
| "grad_norm": 1.6302362931142675, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3773, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 2.8556962025316457, | |
| "grad_norm": 1.594346958058013, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3699, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 2.8658227848101268, | |
| "grad_norm": 1.5803480355180004, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3711, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 2.875949367088608, | |
| "grad_norm": 1.4800494871531567, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3691, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 2.8860759493670884, | |
| "grad_norm": 1.6631915526393122, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3747, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 2.8962025316455695, | |
| "grad_norm": 1.630862108030936, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3674, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 2.9063291139240506, | |
| "grad_norm": 1.6893754841333202, | |
| "learning_rate": 5e-06, | |
| "loss": 0.373, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 2.9164556962025316, | |
| "grad_norm": 1.705060151220597, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3723, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 2.9265822784810127, | |
| "grad_norm": 1.6188859279801948, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3796, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 2.9367088607594938, | |
| "grad_norm": 1.629353248790061, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3828, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 2.946835443037975, | |
| "grad_norm": 1.7509445495897462, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3713, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 2.9569620253164555, | |
| "grad_norm": 1.6650857655447995, | |
| "learning_rate": 5e-06, | |
| "loss": 0.369, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 2.9670886075949365, | |
| "grad_norm": 1.8285248313648312, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3721, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 2.9772151898734176, | |
| "grad_norm": 1.6243458834916522, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3759, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 2.9873417721518987, | |
| "grad_norm": 1.6911778320262214, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3746, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 2.9974683544303797, | |
| "grad_norm": 1.7509767039817499, | |
| "learning_rate": 5e-06, | |
| "loss": 0.3711, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 2.9984810126582278, | |
| "eval_loss": 0.15120381116867065, | |
| "eval_runtime": 508.8119, | |
| "eval_samples_per_second": 26.147, | |
| "eval_steps_per_second": 0.409, | |
| "step": 2961 | |
| }, | |
| { | |
| "epoch": 2.9984810126582278, | |
| "step": 2961, | |
| "total_flos": 2479683262218240.0, | |
| "train_loss": 0.46617310421968466, | |
| "train_runtime": 84499.4991, | |
| "train_samples_per_second": 8.974, | |
| "train_steps_per_second": 0.035 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 2961, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2479683262218240.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |