| { | |
| "best_metric": 1.0779144763946533, | |
| "best_model_checkpoint": "output/bob-dylan/checkpoint-3542", | |
| "epoch": 11.0, | |
| "global_step": 3542, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.02, | |
| "learning_rate": 0.00013711788223044424, | |
| "loss": 3.6967, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "learning_rate": 0.0001368717255202631, | |
| "loss": 3.7692, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "learning_rate": 0.00013646211919423798, | |
| "loss": 3.4916, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "learning_rate": 0.00013589004389254062, | |
| "loss": 3.6186, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "learning_rate": 0.00013515686922297834, | |
| "loss": 3.5756, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "learning_rate": 0.00013426435048201062, | |
| "loss": 3.236, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "learning_rate": 0.0001332146244523866, | |
| "loss": 3.3956, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "learning_rate": 0.00013201020428746477, | |
| "loss": 3.3277, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "learning_rate": 0.0001306539734944624, | |
| "loss": 3.2617, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "learning_rate": 0.00012914917903103908, | |
| "loss": 3.554, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "learning_rate": 0.00012749942353174222, | |
| "loss": 3.2625, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "learning_rate": 0.00012570865668292503, | |
| "loss": 3.2986, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "learning_rate": 0.0001237811657667863, | |
| "loss": 3.264, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "learning_rate": 0.00012172156539717071, | |
| "loss": 3.0437, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "learning_rate": 0.00011953478647170303, | |
| "loss": 3.0527, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "learning_rate": 0.000117226064366706, | |
| "loss": 3.3623, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "learning_rate": 0.0001148009264031647, | |
| "loss": 3.0801, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "learning_rate": 0.0001122651786137447, | |
| "loss": 3.0471, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "learning_rate": 0.00010962489184254581, | |
| "loss": 3.051, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "learning_rate": 0.00010688638721086951, | |
| "loss": 3.2191, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "learning_rate": 0.0001040562209837965, | |
| "loss": 2.9343, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "learning_rate": 0.00010114116887380613, | |
| "loss": 2.9893, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "learning_rate": 9.814820981901533e-05, | |
| "loss": 3.1068, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "learning_rate": 9.508450927487455e-05, | |
| "loss": 3.1636, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "learning_rate": 9.195740205932179e-05, | |
| "loss": 3.1864, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "learning_rate": 8.877437479246497e-05, | |
| "loss": 2.9116, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "learning_rate": 8.554304797283438e-05, | |
| "loss": 3.1795, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "learning_rate": 8.227115773311617e-05, | |
| "loss": 2.9074, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "learning_rate": 7.896653731904552e-05, | |
| "loss": 3.1763, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "learning_rate": 7.56370983358012e-05, | |
| "loss": 3.0588, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "learning_rate": 7.229081180679942e-05, | |
| "loss": 2.9373, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "learning_rate": 6.893568909023427e-05, | |
| "loss": 3.092, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "learning_rate": 6.557976269905237e-05, | |
| "loss": 3.0733, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "learning_rate": 6.223106707028106e-05, | |
| "loss": 2.9187, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "learning_rate": 5.889761932974993e-05, | |
| "loss": 2.9659, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "learning_rate": 5.5587400098257335e-05, | |
| "loss": 3.1799, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "learning_rate": 5.230833438513365e-05, | |
| "loss": 2.9959, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "learning_rate": 4.9068272614944106e-05, | |
| "loss": 2.8529, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "learning_rate": 4.58749718327555e-05, | |
| "loss": 3.0337, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "learning_rate": 4.2736077132963006e-05, | |
| "loss": 3.0827, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "learning_rate": 3.9659103356138536e-05, | |
| "loss": 2.8453, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "learning_rate": 3.6651417097720435e-05, | |
| "loss": 2.9619, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "learning_rate": 3.372021907161731e-05, | |
| "loss": 3.0061, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "learning_rate": 3.0872526870949537e-05, | |
| "loss": 2.8687, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "learning_rate": 2.8115158167201102e-05, | |
| "loss": 2.9913, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "learning_rate": 2.5454714388004492e-05, | |
| "loss": 2.9778, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "learning_rate": 2.289756491263597e-05, | |
| "loss": 3.0748, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "learning_rate": 2.0449831823058788e-05, | |
| "loss": 2.7658, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "learning_rate": 1.8117375247021725e-05, | |
| "loss": 2.9792, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "learning_rate": 1.5905779328303487e-05, | |
| "loss": 2.7718, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "learning_rate": 1.3820338857691364e-05, | |
| "loss": 2.9954, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "learning_rate": 1.1866046596701035e-05, | |
| "loss": 2.9246, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "learning_rate": 1.0047581324385938e-05, | |
| "loss": 2.9875, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "learning_rate": 8.36929663585326e-06, | |
| "loss": 2.673, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "learning_rate": 6.835210519304257e-06, | |
| "loss": 2.8515, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "learning_rate": 5.448995736552248e-06, | |
| "loss": 2.9416, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "learning_rate": 4.213971030048682e-06, | |
| "loss": 2.8573, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "learning_rate": 3.133093177468323e-06, | |
| "loss": 2.6045, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "learning_rate": 2.208949912875789e-06, | |
| "loss": 3.0414, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "learning_rate": 1.4437537314208725e-06, | |
| "loss": 2.7666, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "learning_rate": 8.39336592394954e-07, | |
| "loss": 2.7331, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "learning_rate": 3.971455333297437e-07, | |
| "loss": 2.847, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "learning_rate": 1.1823920563887646e-07, | |
| "loss": 2.7128, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "learning_rate": 3.2853400962779e-09, | |
| "loss": 2.6074, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "eval_loss": 2.670438766479492, | |
| "eval_runtime": 18.6137, | |
| "eval_samples_per_second": 22.833, | |
| "eval_steps_per_second": 2.901, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "learning_rate": 1.6396932901668288e-07, | |
| "loss": 2.7644, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "learning_rate": 4.814968134727319e-07, | |
| "loss": 2.6609, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "learning_rate": 9.651980029285235e-07, | |
| "loss": 2.681, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "learning_rate": 1.6138929181565879e-06, | |
| "loss": 2.6512, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "learning_rate": 2.4259990810865978e-06, | |
| "loss": 2.6379, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "learning_rate": 3.3995353752283744e-06, | |
| "loss": 2.9539, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "learning_rate": 4.532126878565439e-06, | |
| "loss": 2.7261, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "learning_rate": 5.821010657128926e-06, | |
| "loss": 2.8164, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "learning_rate": 7.26304250511898e-06, | |
| "loss": 2.8454, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "learning_rate": 8.854704615130857e-06, | |
| "loss": 2.7863, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "learning_rate": 1.0592114159774732e-05, | |
| "loss": 2.8572, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "learning_rate": 1.2471032763754147e-05, | |
| "loss": 2.8415, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "learning_rate": 1.4486876843296586e-05, | |
| "loss": 2.6313, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "learning_rate": 1.663472878771285e-05, | |
| "loss": 2.7228, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "learning_rate": 1.8909348955808624e-05, | |
| "loss": 2.5958, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "learning_rate": 2.1305188457882567e-05, | |
| "loss": 2.7318, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "learning_rate": 2.3816402692130242e-05, | |
| "loss": 2.7112, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "learning_rate": 2.6436865602431172e-05, | |
| "loss": 2.7692, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "learning_rate": 2.9160184622738906e-05, | |
| "loss": 2.7993, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "learning_rate": 3.197971627161534e-05, | |
| "loss": 2.8784, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "learning_rate": 3.4888582358869375e-05, | |
| "loss": 2.7869, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "learning_rate": 3.7879686764761624e-05, | |
| "loss": 2.8056, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "learning_rate": 4.09457327508451e-05, | |
| "loss": 2.553, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "learning_rate": 4.407924076020965e-05, | |
| "loss": 3.1824, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "learning_rate": 4.7272566663709714e-05, | |
| "loss": 2.8178, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "learning_rate": 5.0517920407661915e-05, | |
| "loss": 2.6659, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "learning_rate": 5.3807385017522816e-05, | |
| "loss": 2.5079, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "learning_rate": 5.7132935911187364e-05, | |
| "loss": 3.0069, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "learning_rate": 6.048646047479427e-05, | |
| "loss": 2.7084, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "learning_rate": 6.385977785328252e-05, | |
| "loss": 2.738, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "learning_rate": 6.724465890742127e-05, | |
| "loss": 2.7037, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "learning_rate": 7.063284628862763e-05, | |
| "loss": 2.7949, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "learning_rate": 7.401607458260099e-05, | |
| "loss": 2.8525, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "learning_rate": 7.738609047263364e-05, | |
| "loss": 2.7132, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "learning_rate": 8.07346728734085e-05, | |
| "loss": 2.7785, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 1.57, | |
| "learning_rate": 8.405365298617048e-05, | |
| "loss": 2.7346, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 1.59, | |
| "learning_rate": 8.733493422634478e-05, | |
| "loss": 2.8556, | |
| "step": 505 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "learning_rate": 9.057051197499066e-05, | |
| "loss": 2.7322, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "learning_rate": 9.375249310590655e-05, | |
| "loss": 2.6462, | |
| "step": 515 | |
| }, | |
| { | |
| "epoch": 1.64, | |
| "learning_rate": 9.687311524075135e-05, | |
| "loss": 2.7953, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "learning_rate": 9.992476568520869e-05, | |
| "loss": 2.7156, | |
| "step": 525 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "learning_rate": 0.00010290000000000001, | |
| "loss": 2.8443, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 1.68, | |
| "learning_rate": 0.00010579156016144291, | |
| "loss": 2.73, | |
| "step": 535 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "learning_rate": 0.00010859239226725277, | |
| "loss": 2.6283, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "learning_rate": 0.00011129566374439388, | |
| "loss": 2.7128, | |
| "step": 545 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "learning_rate": 0.00011389478001700271, | |
| "loss": 2.848, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "learning_rate": 0.00011638340059372158, | |
| "loss": 2.6623, | |
| "step": 555 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "learning_rate": 0.00011875545453519892, | |
| "loss": 2.5335, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "learning_rate": 0.00012100515526402216, | |
| "loss": 2.7063, | |
| "step": 565 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "learning_rate": 0.00012312701468095605, | |
| "loss": 2.5592, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "learning_rate": 0.00012511585655304897, | |
| "loss": 2.658, | |
| "step": 575 | |
| }, | |
| { | |
| "epoch": 1.82, | |
| "learning_rate": 0.00012696682914094848, | |
| "loss": 2.7849, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 1.84, | |
| "learning_rate": 0.00012867541703462067, | |
| "loss": 2.4521, | |
| "step": 585 | |
| }, | |
| { | |
| "epoch": 1.86, | |
| "learning_rate": 0.0001302374521686013, | |
| "loss": 2.7116, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "learning_rate": 0.00013164912398990668, | |
| "loss": 2.5754, | |
| "step": 595 | |
| }, | |
| { | |
| "epoch": 1.89, | |
| "learning_rate": 0.0001329069887538007, | |
| "loss": 2.6935, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "learning_rate": 0.00013400797792473965, | |
| "loss": 2.4856, | |
| "step": 605 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "learning_rate": 0.0001349494056620028, | |
| "loss": 2.7162, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 1.93, | |
| "learning_rate": 0.000135728975371746, | |
| "loss": 2.8196, | |
| "step": 615 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "learning_rate": 0.0001363447853094957, | |
| "loss": 2.6532, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "learning_rate": 0.00013679533321941633, | |
| "loss": 2.5941, | |
| "step": 625 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "learning_rate": 0.00013707951999903246, | |
| "loss": 2.4365, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "learning_rate": 0.00013719665238046719, | |
| "loss": 2.6559, | |
| "step": 635 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "eval_loss": 2.319049835205078, | |
| "eval_runtime": 20.2442, | |
| "eval_samples_per_second": 22.179, | |
| "eval_steps_per_second": 2.816, | |
| "step": 636 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "learning_rate": 0.00013714644462165502, | |
| "loss": 2.3683, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "learning_rate": 0.00013692901920340388, | |
| "loss": 2.3765, | |
| "step": 645 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "learning_rate": 0.00013654490653060555, | |
| "loss": 2.7581, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "learning_rate": 0.00013599504363832372, | |
| "loss": 2.5057, | |
| "step": 655 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "learning_rate": 0.00013528077190591619, | |
| "loss": 2.5706, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "learning_rate": 0.00013440383378476688, | |
| "loss": 2.5847, | |
| "step": 665 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "learning_rate": 0.00013336636854761118, | |
| "loss": 2.3126, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "learning_rate": 0.00013217090706982377, | |
| "loss": 2.309, | |
| "step": 675 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "learning_rate": 0.00013082036565539919, | |
| "loss": 2.5847, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "learning_rate": 0.00012931803892268768, | |
| "loss": 2.4275, | |
| "step": 685 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "learning_rate": 0.00012766759176724058, | |
| "loss": 2.3388, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "learning_rate": 0.00012587305042137224, | |
| "loss": 2.4289, | |
| "step": 695 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "learning_rate": 0.00012393879263224768, | |
| "loss": 2.3432, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "learning_rate": 0.00012186953698245773, | |
| "loss": 2.1438, | |
| "step": 705 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "learning_rate": 0.00011967033137913228, | |
| "loss": 2.4482, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "learning_rate": 0.00011734654073967362, | |
| "loss": 2.596, | |
| "step": 715 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "learning_rate": 0.00011490383390414927, | |
| "loss": 2.3528, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 2.28, | |
| "learning_rate": 0.00011234816980627178, | |
| "loss": 2.2922, | |
| "step": 725 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "learning_rate": 0.0001096857829367009, | |
| "loss": 2.2967, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "learning_rate": 0.00010692316813413058, | |
| "loss": 2.3893, | |
| "step": 735 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "learning_rate": 0.0001040670647412614, | |
| "loss": 2.2958, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "learning_rate": 0.00010112444016431127, | |
| "loss": 2.3878, | |
| "step": 745 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "learning_rate": 9.810247287616931e-05, | |
| "loss": 2.0135, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "learning_rate": 9.500853490465723e-05, | |
| "loss": 2.5128, | |
| "step": 755 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "learning_rate": 9.185017384861694e-05, | |
| "loss": 2.2281, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "learning_rate": 8.86350944656966e-05, | |
| "loss": 2.2168, | |
| "step": 765 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "learning_rate": 8.537113987675077e-05, | |
| "loss": 2.2228, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "learning_rate": 8.206627243270665e-05, | |
| "loss": 2.5163, | |
| "step": 775 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "learning_rate": 7.872855429057025e-05, | |
| "loss": 2.5593, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 2.47, | |
| "learning_rate": 7.536612774595843e-05, | |
| "loss": 1.8992, | |
| "step": 785 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "learning_rate": 7.198719537013403e-05, | |
| "loss": 2.2331, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "learning_rate": 6.860000000000001e-05, | |
| "loss": 2.2157, | |
| "step": 795 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "learning_rate": 6.521280462986602e-05, | |
| "loss": 2.2611, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 2.53, | |
| "learning_rate": 6.183387225404161e-05, | |
| "loss": 2.2053, | |
| "step": 805 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "learning_rate": 5.8471445709429735e-05, | |
| "loss": 2.1826, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "learning_rate": 5.513372756729345e-05, | |
| "loss": 1.995, | |
| "step": 815 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "learning_rate": 5.182886012324932e-05, | |
| "loss": 2.3863, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "learning_rate": 4.856490553430346e-05, | |
| "loss": 2.3232, | |
| "step": 825 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "learning_rate": 4.534982615138311e-05, | |
| "loss": 2.3315, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "learning_rate": 4.2191465095342816e-05, | |
| "loss": 2.3571, | |
| "step": 835 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "learning_rate": 3.909752712383074e-05, | |
| "loss": 2.1077, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "learning_rate": 3.607555983568871e-05, | |
| "loss": 2.1479, | |
| "step": 845 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "learning_rate": 3.313293525873858e-05, | |
| "loss": 2.2508, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "learning_rate": 3.027683186586951e-05, | |
| "loss": 2.5729, | |
| "step": 855 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "learning_rate": 2.7514217063299187e-05, | |
| "loss": 2.4242, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 2.72, | |
| "learning_rate": 2.485183019372827e-05, | |
| "loss": 2.2222, | |
| "step": 865 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "learning_rate": 2.2296166095850762e-05, | |
| "loss": 2.3744, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "learning_rate": 1.9853459260326405e-05, | |
| "loss": 2.2229, | |
| "step": 875 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "learning_rate": 1.752966862086776e-05, | |
| "loss": 2.4542, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 2.78, | |
| "learning_rate": 1.5330463017542246e-05, | |
| "loss": 2.253, | |
| "step": 885 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "learning_rate": 1.3261207367752365e-05, | |
| "loss": 2.2606, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "learning_rate": 1.1326949578627828e-05, | |
| "loss": 2.535, | |
| "step": 895 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "learning_rate": 9.532408232759462e-06, | |
| "loss": 2.3032, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "learning_rate": 7.881961077312348e-06, | |
| "loss": 2.2779, | |
| "step": 905 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "learning_rate": 6.379634344600831e-06, | |
| "loss": 2.2839, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "learning_rate": 5.029092930176238e-06, | |
| "loss": 2.0958, | |
| "step": 915 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "learning_rate": 3.833631452388814e-06, | |
| "loss": 2.1064, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "learning_rate": 2.7961662152331326e-06, | |
| "loss": 2.3978, | |
| "step": 925 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "learning_rate": 1.919228094083838e-06, | |
| "loss": 2.2206, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "learning_rate": 1.204956361676291e-06, | |
| "loss": 2.2185, | |
| "step": 935 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "learning_rate": 6.550934693944858e-07, | |
| "loss": 2.4076, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 2.97, | |
| "learning_rate": 2.709807965961209e-07, | |
| "loss": 2.2124, | |
| "step": 945 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "learning_rate": 5.355537834497188e-08, | |
| "loss": 2.6099, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "eval_loss": 2.0997118949890137, | |
| "eval_runtime": 20.2461, | |
| "eval_samples_per_second": 22.177, | |
| "eval_steps_per_second": 2.815, | |
| "step": 954 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "learning_rate": 3.347619532822632e-09, | |
| "loss": 2.1903, | |
| "step": 955 | |
| }, | |
| { | |
| "epoch": 3.02, | |
| "learning_rate": 1.2048000096755528e-07, | |
| "loss": 2.2037, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 3.03, | |
| "learning_rate": 4.0466678058365933e-07, | |
| "loss": 1.8544, | |
| "step": 965 | |
| }, | |
| { | |
| "epoch": 3.05, | |
| "learning_rate": 8.552146905042755e-07, | |
| "loss": 2.1987, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 3.07, | |
| "learning_rate": 1.4710246282540082e-06, | |
| "loss": 2.1019, | |
| "step": 975 | |
| }, | |
| { | |
| "epoch": 3.08, | |
| "learning_rate": 2.250594337997185e-06, | |
| "loss": 2.1517, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 3.1, | |
| "learning_rate": 3.192022075260327e-06, | |
| "loss": 2.1346, | |
| "step": 985 | |
| }, | |
| { | |
| "epoch": 3.11, | |
| "learning_rate": 4.293011246199299e-06, | |
| "loss": 2.0847, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 3.13, | |
| "learning_rate": 5.550876010093297e-06, | |
| "loss": 1.9358, | |
| "step": 995 | |
| }, | |
| { | |
| "epoch": 3.14, | |
| "learning_rate": 6.962547831398709e-06, | |
| "loss": 2.2579, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 3.16, | |
| "learning_rate": 8.524582965379288e-06, | |
| "loss": 2.0876, | |
| "step": 1005 | |
| }, | |
| { | |
| "epoch": 3.18, | |
| "learning_rate": 1.0233170859051466e-05, | |
| "loss": 1.8737, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 3.19, | |
| "learning_rate": 1.2084143446950978e-05, | |
| "loss": 2.1, | |
| "step": 1015 | |
| }, | |
| { | |
| "epoch": 3.21, | |
| "learning_rate": 1.4072985319043973e-05, | |
| "loss": 2.0016, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 3.22, | |
| "learning_rate": 1.619484473597781e-05, | |
| "loss": 2.1936, | |
| "step": 1025 | |
| }, | |
| { | |
| "epoch": 3.24, | |
| "learning_rate": 1.844454546480105e-05, | |
| "loss": 2.2236, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 3.25, | |
| "learning_rate": 2.081659940627838e-05, | |
| "loss": 1.94, | |
| "step": 1035 | |
| }, | |
| { | |
| "epoch": 3.27, | |
| "learning_rate": 2.330521998299727e-05, | |
| "loss": 2.103, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 3.29, | |
| "learning_rate": 2.5904336255606053e-05, | |
| "loss": 2.0909, | |
| "step": 1045 | |
| }, | |
| { | |
| "epoch": 3.3, | |
| "learning_rate": 2.860760773274715e-05, | |
| "loss": 2.1943, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 3.32, | |
| "learning_rate": 3.1408439838557e-05, | |
| "loss": 2.3713, | |
| "step": 1055 | |
| }, | |
| { | |
| "epoch": 3.33, | |
| "learning_rate": 3.4300000000000014e-05, | |
| "loss": 2.1534, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 3.35, | |
| "learning_rate": 3.727523431479128e-05, | |
| "loss": 2.1973, | |
| "step": 1065 | |
| }, | |
| { | |
| "epoch": 3.36, | |
| "learning_rate": 4.0326884759248605e-05, | |
| "loss": 2.1412, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 3.38, | |
| "learning_rate": 4.3447506894093424e-05, | |
| "loss": 2.1419, | |
| "step": 1075 | |
| }, | |
| { | |
| "epoch": 3.4, | |
| "learning_rate": 4.662948802500929e-05, | |
| "loss": 2.179, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 3.41, | |
| "learning_rate": 4.986506577365512e-05, | |
| "loss": 2.0405, | |
| "step": 1085 | |
| }, | |
| { | |
| "epoch": 3.43, | |
| "learning_rate": 5.314634701382942e-05, | |
| "loss": 2.1397, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 3.44, | |
| "learning_rate": 5.646532712659141e-05, | |
| "loss": 2.0391, | |
| "step": 1095 | |
| }, | |
| { | |
| "epoch": 3.46, | |
| "learning_rate": 5.9813909527366384e-05, | |
| "loss": 2.2312, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 3.47, | |
| "learning_rate": 6.318392541739896e-05, | |
| "loss": 1.8413, | |
| "step": 1105 | |
| }, | |
| { | |
| "epoch": 3.49, | |
| "learning_rate": 6.656715371137238e-05, | |
| "loss": 2.2519, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 3.51, | |
| "learning_rate": 6.995534109257869e-05, | |
| "loss": 1.871, | |
| "step": 1115 | |
| }, | |
| { | |
| "epoch": 3.52, | |
| "learning_rate": 7.334022214671738e-05, | |
| "loss": 1.9633, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 3.54, | |
| "learning_rate": 7.671353952520564e-05, | |
| "loss": 2.3068, | |
| "step": 1125 | |
| }, | |
| { | |
| "epoch": 3.55, | |
| "learning_rate": 8.006706408881254e-05, | |
| "loss": 2.083, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 3.57, | |
| "learning_rate": 8.33926149824772e-05, | |
| "loss": 2.1478, | |
| "step": 1135 | |
| }, | |
| { | |
| "epoch": 3.58, | |
| "learning_rate": 8.668207959233806e-05, | |
| "loss": 2.1284, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 3.6, | |
| "learning_rate": 8.992743333629024e-05, | |
| "loss": 2.1958, | |
| "step": 1145 | |
| }, | |
| { | |
| "epoch": 3.62, | |
| "learning_rate": 9.312075923979032e-05, | |
| "loss": 1.9476, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 3.63, | |
| "learning_rate": 9.625426724915486e-05, | |
| "loss": 2.1252, | |
| "step": 1155 | |
| }, | |
| { | |
| "epoch": 3.65, | |
| "learning_rate": 9.932031323523828e-05, | |
| "loss": 1.921, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 3.66, | |
| "learning_rate": 0.00010231141764113059, | |
| "loss": 2.1346, | |
| "step": 1165 | |
| }, | |
| { | |
| "epoch": 3.68, | |
| "learning_rate": 0.00010522028372838456, | |
| "loss": 2.1729, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 3.69, | |
| "learning_rate": 0.00010803981537726111, | |
| "loss": 1.7231, | |
| "step": 1175 | |
| }, | |
| { | |
| "epoch": 3.71, | |
| "learning_rate": 0.0001107631343975688, | |
| "loss": 1.9233, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 3.73, | |
| "learning_rate": 0.00011338359730786977, | |
| "loss": 2.2671, | |
| "step": 1185 | |
| }, | |
| { | |
| "epoch": 3.74, | |
| "learning_rate": 0.0001158948115421174, | |
| "loss": 1.9335, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 3.76, | |
| "learning_rate": 0.00011829065104419134, | |
| "loss": 2.0918, | |
| "step": 1195 | |
| }, | |
| { | |
| "epoch": 3.77, | |
| "learning_rate": 0.00012056527121228711, | |
| "loss": 2.2418, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 3.79, | |
| "learning_rate": 0.00012271312315670338, | |
| "loss": 2.1403, | |
| "step": 1205 | |
| }, | |
| { | |
| "epoch": 3.81, | |
| "learning_rate": 0.0001247289672362458, | |
| "loss": 2.1248, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 3.82, | |
| "learning_rate": 0.00012660788584022527, | |
| "loss": 2.0255, | |
| "step": 1215 | |
| }, | |
| { | |
| "epoch": 3.84, | |
| "learning_rate": 0.00012834529538486916, | |
| "loss": 2.1805, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 3.85, | |
| "learning_rate": 0.000129936957494881, | |
| "loss": 1.9559, | |
| "step": 1225 | |
| }, | |
| { | |
| "epoch": 3.87, | |
| "learning_rate": 0.00013137898934287106, | |
| "loss": 2.1316, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 3.88, | |
| "learning_rate": 0.00013266787312143455, | |
| "loss": 2.4223, | |
| "step": 1235 | |
| }, | |
| { | |
| "epoch": 3.9, | |
| "learning_rate": 0.0001338004646247716, | |
| "loss": 2.0369, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 3.92, | |
| "learning_rate": 0.00013477400091891338, | |
| "loss": 2.0681, | |
| "step": 1245 | |
| }, | |
| { | |
| "epoch": 3.93, | |
| "learning_rate": 0.00013558610708184338, | |
| "loss": 2.1821, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 3.95, | |
| "learning_rate": 0.00013623480199707148, | |
| "loss": 2.2183, | |
| "step": 1255 | |
| }, | |
| { | |
| "epoch": 3.96, | |
| "learning_rate": 0.00013671850318652728, | |
| "loss": 2.0026, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 3.98, | |
| "learning_rate": 0.00013703603067098332, | |
| "loss": 1.9368, | |
| "step": 1265 | |
| }, | |
| { | |
| "epoch": 3.99, | |
| "learning_rate": 0.0001371866098485905, | |
| "loss": 2.1825, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "eval_loss": 2.0486814975738525, | |
| "eval_runtime": 20.2715, | |
| "eval_samples_per_second": 22.149, | |
| "eval_steps_per_second": 2.812, | |
| "step": 1272 | |
| }, | |
| { | |
| "epoch": 4.01, | |
| "learning_rate": 0.0001371698733845033, | |
| "loss": 2.0223, | |
| "step": 1275 | |
| }, | |
| { | |
| "epoch": 4.03, | |
| "learning_rate": 0.000136985862106986, | |
| "loss": 2.0536, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 4.04, | |
| "learning_rate": 0.0001366350249078127, | |
| "loss": 1.962, | |
| "step": 1285 | |
| }, | |
| { | |
| "epoch": 4.06, | |
| "learning_rate": 0.00013611821764720515, | |
| "loss": 2.0683, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 4.07, | |
| "learning_rate": 0.00013543670106597888, | |
| "loss": 1.9271, | |
| "step": 1295 | |
| }, | |
| { | |
| "epoch": 4.09, | |
| "learning_rate": 0.00013459213770999188, | |
| "loss": 2.0152, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 4.1, | |
| "learning_rate": 0.00013358658787439754, | |
| "loss": 1.7685, | |
| "step": 1305 | |
| }, | |
| { | |
| "epoch": 4.12, | |
| "learning_rate": 0.0001324225045775965, | |
| "loss": 2.2589, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 4.14, | |
| "learning_rate": 0.00013110272757714818, | |
| "loss": 1.9458, | |
| "step": 1315 | |
| }, | |
| { | |
| "epoch": 4.15, | |
| "learning_rate": 0.00012963047644223968, | |
| "loss": 2.0727, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 4.17, | |
| "learning_rate": 0.00012800934269961248, | |
| "loss": 2.0622, | |
| "step": 1325 | |
| }, | |
| { | |
| "epoch": 4.18, | |
| "learning_rate": 0.0001262432810721057, | |
| "loss": 1.7668, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 4.2, | |
| "learning_rate": 0.00012433659983118975, | |
| "loss": 1.6953, | |
| "step": 1335 | |
| }, | |
| { | |
| "epoch": 4.21, | |
| "learning_rate": 0.00012229395028702625, | |
| "loss": 2.1107, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 4.23, | |
| "learning_rate": 0.00012012031544169091, | |
| "loss": 1.7084, | |
| "step": 1345 | |
| }, | |
| { | |
| "epoch": 4.25, | |
| "learning_rate": 0.00011782099783324114, | |
| "loss": 2.281, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 4.26, | |
| "learning_rate": 0.00011540160660028247, | |
| "loss": 2.0195, | |
| "step": 1355 | |
| }, | |
| { | |
| "epoch": 4.28, | |
| "learning_rate": 0.0001128680437985883, | |
| "loss": 1.9339, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 4.29, | |
| "learning_rate": 0.00011022649000315548, | |
| "loss": 1.8057, | |
| "step": 1365 | |
| }, | |
| { | |
| "epoch": 4.31, | |
| "learning_rate": 0.00010748338923081677, | |
| "loss": 1.9412, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 4.32, | |
| "learning_rate": 0.00010464543322019205, | |
| "loss": 1.6871, | |
| "step": 1375 | |
| }, | |
| { | |
| "epoch": 4.34, | |
| "learning_rate": 0.0001017195451073291, | |
| "loss": 1.8352, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 4.36, | |
| "learning_rate": 9.871286253685277e-05, | |
| "loss": 1.8437, | |
| "step": 1385 | |
| }, | |
| { | |
| "epoch": 4.37, | |
| "learning_rate": 9.56327202498264e-05, | |
| "loss": 2.0612, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 4.39, | |
| "learning_rate": 9.248663219079991e-05, | |
| "loss": 1.9911, | |
| "step": 1395 | |
| }, | |
| { | |
| "epoch": 4.4, | |
| "learning_rate": 8.928227317769423e-05, | |
| "loss": 1.8443, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 4.42, | |
| "learning_rate": 8.602746017923975e-05, | |
| "loss": 1.9599, | |
| "step": 1405 | |
| }, | |
| { | |
| "epoch": 4.43, | |
| "learning_rate": 8.273013324563943e-05, | |
| "loss": 1.8763, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 4.45, | |
| "learning_rate": 7.939833613897684e-05, | |
| "loss": 1.7859, | |
| "step": 1415 | |
| }, | |
| { | |
| "epoch": 4.47, | |
| "learning_rate": 7.60401967106234e-05, | |
| "loss": 2.1938, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 4.48, | |
| "learning_rate": 7.266390707350893e-05, | |
| "loss": 1.554, | |
| "step": 1425 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "learning_rate": 6.927770361762858e-05, | |
| "loss": 1.9093, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 4.51, | |
| "learning_rate": 6.588984691753754e-05, | |
| "loss": 1.6916, | |
| "step": 1435 | |
| }, | |
| { | |
| "epoch": 4.53, | |
| "learning_rate": 6.250860158084673e-05, | |
| "loss": 1.9688, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 4.54, | |
| "learning_rate": 5.91422160868796e-05, | |
| "loss": 1.9263, | |
| "step": 1445 | |
| }, | |
| { | |
| "epoch": 4.56, | |
| "learning_rate": 5.579890266467603e-05, | |
| "loss": 1.6181, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 4.58, | |
| "learning_rate": 5.248681725942451e-05, | |
| "loss": 2.0187, | |
| "step": 1455 | |
| }, | |
| { | |
| "epoch": 4.59, | |
| "learning_rate": 4.921403963620125e-05, | |
| "loss": 1.6975, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 4.61, | |
| "learning_rate": 4.598855366954619e-05, | |
| "loss": 1.995, | |
| "step": 1465 | |
| }, | |
| { | |
| "epoch": 4.62, | |
| "learning_rate": 4.28182278669633e-05, | |
| "loss": 1.9041, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 4.64, | |
| "learning_rate": 3.9710796173857146e-05, | |
| "loss": 1.8769, | |
| "step": 1475 | |
| }, | |
| { | |
| "epoch": 4.65, | |
| "learning_rate": 3.667383910672967e-05, | |
| "loss": 1.6473, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 4.67, | |
| "learning_rate": 3.371476526066344e-05, | |
| "loss": 1.5014, | |
| "step": 1485 | |
| }, | |
| { | |
| "epoch": 4.69, | |
| "learning_rate": 3.084079323620596e-05, | |
| "loss": 1.8409, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 4.7, | |
| "learning_rate": 2.805893402973855e-05, | |
| "loss": 1.8548, | |
| "step": 1495 | |
| }, | |
| { | |
| "epoch": 4.72, | |
| "learning_rate": 2.5375973930294628e-05, | |
| "loss": 1.8673, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 4.73, | |
| "learning_rate": 2.2798457964544384e-05, | |
| "loss": 1.8588, | |
| "step": 1505 | |
| }, | |
| { | |
| "epoch": 4.75, | |
| "learning_rate": 2.0332673930335255e-05, | |
| "loss": 2.0708, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 4.76, | |
| "learning_rate": 1.7984637057737467e-05, | |
| "loss": 1.782, | |
| "step": 1515 | |
| }, | |
| { | |
| "epoch": 4.78, | |
| "learning_rate": 1.5760075335011966e-05, | |
| "loss": 1.9664, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 4.8, | |
| "learning_rate": 1.3664415535298983e-05, | |
| "loss": 1.9151, | |
| "step": 1525 | |
| }, | |
| { | |
| "epoch": 4.81, | |
| "learning_rate": 1.1702769978116211e-05, | |
| "loss": 1.9315, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 4.83, | |
| "learning_rate": 9.879924057958108e-06, | |
| "loss": 1.8344, | |
| "step": 1535 | |
| }, | |
| { | |
| "epoch": 4.84, | |
| "learning_rate": 8.20032457042391e-06, | |
| "loss": 1.8142, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 4.86, | |
| "learning_rate": 6.6680688643500544e-06, | |
| "loss": 1.7706, | |
| "step": 1545 | |
| }, | |
| { | |
| "epoch": 4.87, | |
| "learning_rate": 5.2868948464103564e-06, | |
| "loss": 2.0751, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 4.89, | |
| "learning_rate": 4.060171862569283e-06, | |
| "loss": 2.0826, | |
| "step": 1555 | |
| }, | |
| { | |
| "epoch": 4.91, | |
| "learning_rate": 2.9908924786305843e-06, | |
| "loss": 1.9677, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 4.92, | |
| "learning_rate": 2.0816651799335755e-06, | |
| "loss": 1.7251, | |
| "step": 1565 | |
| }, | |
| { | |
| "epoch": 4.94, | |
| "learning_rate": 1.3347080080066644e-06, | |
| "loss": 1.9202, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 4.95, | |
| "learning_rate": 7.518431496995702e-07, | |
| "loss": 1.833, | |
| "step": 1575 | |
| }, | |
| { | |
| "epoch": 4.97, | |
| "learning_rate": 3.34492491995841e-07, | |
| "loss": 1.8957, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 4.98, | |
| "learning_rate": 8.367415334837176e-08, | |
| "loss": 1.7695, | |
| "step": 1585 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "learning_rate": 0.0, | |
| "loss": 1.9115, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "eval_loss": 1.906249761581421, | |
| "eval_runtime": 20.2347, | |
| "eval_samples_per_second": 22.19, | |
| "eval_steps_per_second": 2.817, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 5.02, | |
| "learning_rate": 8.367415334837176e-08, | |
| "loss": 1.6316, | |
| "step": 1595 | |
| }, | |
| { | |
| "epoch": 5.03, | |
| "learning_rate": 3.3449249199583335e-07, | |
| "loss": 1.7125, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 5.05, | |
| "learning_rate": 7.51843149699555e-07, | |
| "loss": 1.631, | |
| "step": 1605 | |
| }, | |
| { | |
| "epoch": 5.06, | |
| "learning_rate": 1.3347080080066491e-06, | |
| "loss": 1.5559, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 5.08, | |
| "learning_rate": 2.0816651799335526e-06, | |
| "loss": 1.7693, | |
| "step": 1615 | |
| }, | |
| { | |
| "epoch": 5.09, | |
| "learning_rate": 2.990892478630592e-06, | |
| "loss": 1.5954, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 5.11, | |
| "learning_rate": 4.060171862569298e-06, | |
| "loss": 1.9519, | |
| "step": 1625 | |
| }, | |
| { | |
| "epoch": 5.13, | |
| "learning_rate": 5.2868948464102726e-06, | |
| "loss": 1.5427, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 5.14, | |
| "learning_rate": 6.668068864349971e-06, | |
| "loss": 1.5142, | |
| "step": 1635 | |
| }, | |
| { | |
| "epoch": 5.16, | |
| "learning_rate": 8.200324570423812e-06, | |
| "loss": 1.4704, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 5.17, | |
| "learning_rate": 9.879924057958001e-06, | |
| "loss": 1.724, | |
| "step": 1645 | |
| }, | |
| { | |
| "epoch": 5.19, | |
| "learning_rate": 1.1702769978116166e-05, | |
| "loss": 1.7845, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 5.2, | |
| "learning_rate": 1.366441553529893e-05, | |
| "loss": 1.5651, | |
| "step": 1655 | |
| }, | |
| { | |
| "epoch": 5.22, | |
| "learning_rate": 1.5760075335011912e-05, | |
| "loss": 1.5684, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 5.24, | |
| "learning_rate": 1.7984637057737406e-05, | |
| "loss": 1.8357, | |
| "step": 1665 | |
| }, | |
| { | |
| "epoch": 5.25, | |
| "learning_rate": 2.0332673930335194e-05, | |
| "loss": 1.6561, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 5.27, | |
| "learning_rate": 2.2798457964544323e-05, | |
| "loss": 1.7467, | |
| "step": 1675 | |
| }, | |
| { | |
| "epoch": 5.28, | |
| "learning_rate": 2.5375973930294556e-05, | |
| "loss": 1.9728, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 5.3, | |
| "learning_rate": 2.805893402973858e-05, | |
| "loss": 1.813, | |
| "step": 1685 | |
| }, | |
| { | |
| "epoch": 5.31, | |
| "learning_rate": 3.084079323620599e-05, | |
| "loss": 1.7629, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 5.33, | |
| "learning_rate": 3.3714765260663475e-05, | |
| "loss": 1.7309, | |
| "step": 1695 | |
| }, | |
| { | |
| "epoch": 5.35, | |
| "learning_rate": 3.66738391067297e-05, | |
| "loss": 1.5675, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 5.36, | |
| "learning_rate": 3.971079617385719e-05, | |
| "loss": 1.4041, | |
| "step": 1705 | |
| }, | |
| { | |
| "epoch": 5.38, | |
| "learning_rate": 4.28182278669631e-05, | |
| "loss": 1.5323, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 5.39, | |
| "learning_rate": 4.5988553669545996e-05, | |
| "loss": 1.7705, | |
| "step": 1715 | |
| }, | |
| { | |
| "epoch": 5.41, | |
| "learning_rate": 4.921403963620105e-05, | |
| "loss": 1.7652, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 5.42, | |
| "learning_rate": 5.2486817259424427e-05, | |
| "loss": 1.8281, | |
| "step": 1725 | |
| }, | |
| { | |
| "epoch": 5.44, | |
| "learning_rate": 5.579890266467595e-05, | |
| "loss": 1.6899, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 5.46, | |
| "learning_rate": 5.914221608687952e-05, | |
| "loss": 1.6888, | |
| "step": 1735 | |
| }, | |
| { | |
| "epoch": 5.47, | |
| "learning_rate": 6.250860158084665e-05, | |
| "loss": 1.7589, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 5.49, | |
| "learning_rate": 6.588984691753746e-05, | |
| "loss": 1.6297, | |
| "step": 1745 | |
| }, | |
| { | |
| "epoch": 5.5, | |
| "learning_rate": 6.92777036176285e-05, | |
| "loss": 1.8036, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 5.52, | |
| "learning_rate": 7.266390707350887e-05, | |
| "loss": 1.8385, | |
| "step": 1755 | |
| }, | |
| { | |
| "epoch": 5.53, | |
| "learning_rate": 7.604019671062332e-05, | |
| "loss": 1.7527, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 5.55, | |
| "learning_rate": 7.939833613897687e-05, | |
| "loss": 1.5839, | |
| "step": 1765 | |
| }, | |
| { | |
| "epoch": 5.57, | |
| "learning_rate": 8.273013324563949e-05, | |
| "loss": 1.9817, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 5.58, | |
| "learning_rate": 8.602746017923979e-05, | |
| "loss": 2.1168, | |
| "step": 1775 | |
| }, | |
| { | |
| "epoch": 5.6, | |
| "learning_rate": 8.928227317769404e-05, | |
| "loss": 1.8069, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 5.61, | |
| "learning_rate": 9.248663219079973e-05, | |
| "loss": 1.886, | |
| "step": 1785 | |
| }, | |
| { | |
| "epoch": 5.63, | |
| "learning_rate": 9.56327202498262e-05, | |
| "loss": 1.8172, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 5.64, | |
| "learning_rate": 9.871286253685269e-05, | |
| "loss": 1.6818, | |
| "step": 1795 | |
| }, | |
| { | |
| "epoch": 5.66, | |
| "learning_rate": 0.00010171954510732892, | |
| "loss": 1.4689, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 5.68, | |
| "learning_rate": 0.00010464543322019198, | |
| "loss": 1.3972, | |
| "step": 1805 | |
| }, | |
| { | |
| "epoch": 5.69, | |
| "learning_rate": 0.00010748338923081671, | |
| "loss": 1.8636, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 5.71, | |
| "learning_rate": 0.0001102264900031554, | |
| "loss": 1.5656, | |
| "step": 1815 | |
| }, | |
| { | |
| "epoch": 5.72, | |
| "learning_rate": 0.00011286804379858823, | |
| "loss": 1.834, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 5.74, | |
| "learning_rate": 0.00011540160660028242, | |
| "loss": 1.6874, | |
| "step": 1825 | |
| }, | |
| { | |
| "epoch": 5.75, | |
| "learning_rate": 0.00011782099783324108, | |
| "loss": 1.7932, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 5.77, | |
| "learning_rate": 0.00012012031544169092, | |
| "loss": 1.8093, | |
| "step": 1835 | |
| }, | |
| { | |
| "epoch": 5.79, | |
| "learning_rate": 0.00012229395028702622, | |
| "loss": 1.4865, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 5.8, | |
| "learning_rate": 0.00012433659983118975, | |
| "loss": 1.7283, | |
| "step": 1845 | |
| }, | |
| { | |
| "epoch": 5.82, | |
| "learning_rate": 0.0001262432810721057, | |
| "loss": 1.5611, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 5.83, | |
| "learning_rate": 0.00012800934269961248, | |
| "loss": 1.7974, | |
| "step": 1855 | |
| }, | |
| { | |
| "epoch": 5.85, | |
| "learning_rate": 0.0001296304764422396, | |
| "loss": 1.6754, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 5.86, | |
| "learning_rate": 0.0001311027275771481, | |
| "loss": 1.5481, | |
| "step": 1865 | |
| }, | |
| { | |
| "epoch": 5.88, | |
| "learning_rate": 0.0001324225045775964, | |
| "loss": 1.7931, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 5.9, | |
| "learning_rate": 0.0001335865878743975, | |
| "loss": 1.9985, | |
| "step": 1875 | |
| }, | |
| { | |
| "epoch": 5.91, | |
| "learning_rate": 0.00013459213770999182, | |
| "loss": 1.9837, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 5.93, | |
| "learning_rate": 0.00013543670106597888, | |
| "loss": 1.7165, | |
| "step": 1885 | |
| }, | |
| { | |
| "epoch": 5.94, | |
| "learning_rate": 0.00013611821764720515, | |
| "loss": 1.9364, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 5.96, | |
| "learning_rate": 0.0001366350249078127, | |
| "loss": 1.6662, | |
| "step": 1895 | |
| }, | |
| { | |
| "epoch": 5.97, | |
| "learning_rate": 0.000136985862106986, | |
| "loss": 1.8671, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 5.99, | |
| "learning_rate": 0.0001371698733845033, | |
| "loss": 1.7105, | |
| "step": 1905 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "eval_loss": 1.9032058715820312, | |
| "eval_runtime": 20.2363, | |
| "eval_samples_per_second": 22.188, | |
| "eval_steps_per_second": 2.817, | |
| "step": 1908 | |
| }, | |
| { | |
| "epoch": 6.01, | |
| "learning_rate": 0.0001371866098485905, | |
| "loss": 1.4646, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 6.02, | |
| "learning_rate": 0.00013703603067098332, | |
| "loss": 1.572, | |
| "step": 1915 | |
| }, | |
| { | |
| "epoch": 6.04, | |
| "learning_rate": 0.00013671850318652725, | |
| "loss": 1.6881, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 6.05, | |
| "learning_rate": 0.00013623480199707148, | |
| "loss": 1.8478, | |
| "step": 1925 | |
| }, | |
| { | |
| "epoch": 6.07, | |
| "learning_rate": 0.00013558610708184343, | |
| "loss": 1.6836, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 6.08, | |
| "learning_rate": 0.0001347740009189134, | |
| "loss": 1.7053, | |
| "step": 1935 | |
| }, | |
| { | |
| "epoch": 6.1, | |
| "learning_rate": 0.00013380046462477165, | |
| "loss": 1.5923, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 6.12, | |
| "learning_rate": 0.00013266787312143458, | |
| "loss": 1.5511, | |
| "step": 1945 | |
| }, | |
| { | |
| "epoch": 6.13, | |
| "learning_rate": 0.00013137898934287114, | |
| "loss": 1.6041, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 6.15, | |
| "learning_rate": 0.00012993695749488105, | |
| "loss": 1.5811, | |
| "step": 1955 | |
| }, | |
| { | |
| "epoch": 6.16, | |
| "learning_rate": 0.0001283452953848691, | |
| "loss": 1.6477, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 6.18, | |
| "learning_rate": 0.00012660788584022533, | |
| "loss": 1.6821, | |
| "step": 1965 | |
| }, | |
| { | |
| "epoch": 6.19, | |
| "learning_rate": 0.00012472896723624585, | |
| "loss": 1.9101, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 6.21, | |
| "learning_rate": 0.00012271312315670352, | |
| "loss": 1.7961, | |
| "step": 1975 | |
| }, | |
| { | |
| "epoch": 6.23, | |
| "learning_rate": 0.00012056527121228716, | |
| "loss": 1.7949, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 6.24, | |
| "learning_rate": 0.00011829065104419132, | |
| "loss": 1.4401, | |
| "step": 1985 | |
| }, | |
| { | |
| "epoch": 6.26, | |
| "learning_rate": 0.00011589481154211747, | |
| "loss": 1.6164, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 6.27, | |
| "learning_rate": 0.00011338359730786976, | |
| "loss": 1.8185, | |
| "step": 1995 | |
| }, | |
| { | |
| "epoch": 6.29, | |
| "learning_rate": 0.00011076313439756885, | |
| "loss": 1.9472, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 6.31, | |
| "learning_rate": 0.00010803981537726108, | |
| "loss": 1.5825, | |
| "step": 2005 | |
| }, | |
| { | |
| "epoch": 6.32, | |
| "learning_rate": 0.00010522028372838475, | |
| "loss": 1.6182, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 6.34, | |
| "learning_rate": 0.00010231141764113088, | |
| "loss": 1.6272, | |
| "step": 2015 | |
| }, | |
| { | |
| "epoch": 6.35, | |
| "learning_rate": 9.932031323523847e-05, | |
| "loss": 1.4251, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 6.37, | |
| "learning_rate": 9.625426724915494e-05, | |
| "loss": 1.5946, | |
| "step": 2025 | |
| }, | |
| { | |
| "epoch": 6.38, | |
| "learning_rate": 9.31207592397905e-05, | |
| "loss": 1.5311, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 6.4, | |
| "learning_rate": 8.992743333629033e-05, | |
| "loss": 1.7549, | |
| "step": 2035 | |
| }, | |
| { | |
| "epoch": 6.42, | |
| "learning_rate": 8.6682079592338e-05, | |
| "loss": 1.7651, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 6.43, | |
| "learning_rate": 8.339261498247729e-05, | |
| "loss": 1.7894, | |
| "step": 2045 | |
| }, | |
| { | |
| "epoch": 6.45, | |
| "learning_rate": 8.006706408881262e-05, | |
| "loss": 1.2016, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 6.46, | |
| "learning_rate": 7.671353952520582e-05, | |
| "loss": 1.5423, | |
| "step": 2055 | |
| }, | |
| { | |
| "epoch": 6.48, | |
| "learning_rate": 7.334022214671746e-05, | |
| "loss": 1.2711, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 6.49, | |
| "learning_rate": 6.995534109257865e-05, | |
| "loss": 1.582, | |
| "step": 2065 | |
| }, | |
| { | |
| "epoch": 6.51, | |
| "learning_rate": 6.656715371137246e-05, | |
| "loss": 1.55, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 6.53, | |
| "learning_rate": 6.318392541739893e-05, | |
| "loss": 1.5744, | |
| "step": 2075 | |
| }, | |
| { | |
| "epoch": 6.54, | |
| "learning_rate": 5.9813909527366465e-05, | |
| "loss": 1.7316, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 6.56, | |
| "learning_rate": 5.646532712659148e-05, | |
| "loss": 1.5926, | |
| "step": 2085 | |
| }, | |
| { | |
| "epoch": 6.57, | |
| "learning_rate": 5.314634701382963e-05, | |
| "loss": 1.3707, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 6.59, | |
| "learning_rate": 4.98650657736552e-05, | |
| "loss": 1.3874, | |
| "step": 2095 | |
| }, | |
| { | |
| "epoch": 6.6, | |
| "learning_rate": 4.6629488025009487e-05, | |
| "loss": 1.6198, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 6.62, | |
| "learning_rate": 4.3447506894093505e-05, | |
| "loss": 1.6056, | |
| "step": 2105 | |
| }, | |
| { | |
| "epoch": 6.64, | |
| "learning_rate": 4.0326884759248795e-05, | |
| "loss": 1.3487, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 6.65, | |
| "learning_rate": 3.7275234314791357e-05, | |
| "loss": 1.5276, | |
| "step": 2115 | |
| }, | |
| { | |
| "epoch": 6.67, | |
| "learning_rate": 3.429999999999998e-05, | |
| "loss": 1.6535, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 6.68, | |
| "learning_rate": 3.140843983855718e-05, | |
| "loss": 1.8044, | |
| "step": 2125 | |
| }, | |
| { | |
| "epoch": 6.7, | |
| "learning_rate": 2.860760773274722e-05, | |
| "loss": 1.705, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 6.71, | |
| "learning_rate": 2.5904336255606023e-05, | |
| "loss": 1.4356, | |
| "step": 2135 | |
| }, | |
| { | |
| "epoch": 6.73, | |
| "learning_rate": 2.3305219982997338e-05, | |
| "loss": 1.4672, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 6.75, | |
| "learning_rate": 2.0816599406278358e-05, | |
| "loss": 1.5759, | |
| "step": 2145 | |
| }, | |
| { | |
| "epoch": 6.76, | |
| "learning_rate": 1.8444545464801106e-05, | |
| "loss": 1.655, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 6.78, | |
| "learning_rate": 1.6194844735977787e-05, | |
| "loss": 1.5595, | |
| "step": 2155 | |
| }, | |
| { | |
| "epoch": 6.79, | |
| "learning_rate": 1.4072985319044027e-05, | |
| "loss": 1.5155, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 6.81, | |
| "learning_rate": 1.208414344695116e-05, | |
| "loss": 1.6561, | |
| "step": 2165 | |
| }, | |
| { | |
| "epoch": 6.82, | |
| "learning_rate": 1.0233170859051572e-05, | |
| "loss": 1.4243, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 6.84, | |
| "learning_rate": 8.524582965379327e-06, | |
| "loss": 1.3178, | |
| "step": 2175 | |
| }, | |
| { | |
| "epoch": 6.86, | |
| "learning_rate": 6.9625478313988e-06, | |
| "loss": 1.2895, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 6.87, | |
| "learning_rate": 5.5508760100933275e-06, | |
| "loss": 1.8021, | |
| "step": 2185 | |
| }, | |
| { | |
| "epoch": 6.89, | |
| "learning_rate": 4.293011246199375e-06, | |
| "loss": 1.2922, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 6.9, | |
| "learning_rate": 3.19202207526035e-06, | |
| "loss": 1.514, | |
| "step": 2195 | |
| }, | |
| { | |
| "epoch": 6.92, | |
| "learning_rate": 2.2505943379971774e-06, | |
| "loss": 1.7064, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 6.93, | |
| "learning_rate": 1.4710246282540463e-06, | |
| "loss": 1.7752, | |
| "step": 2205 | |
| }, | |
| { | |
| "epoch": 6.95, | |
| "learning_rate": 8.552146905042831e-07, | |
| "loss": 1.4982, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 6.97, | |
| "learning_rate": 4.0466678058365933e-07, | |
| "loss": 1.4699, | |
| "step": 2215 | |
| }, | |
| { | |
| "epoch": 6.98, | |
| "learning_rate": 1.2048000096755528e-07, | |
| "loss": 1.7008, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "learning_rate": 3.347619532822632e-09, | |
| "loss": 1.5587, | |
| "step": 2225 | |
| }, | |
| { | |
| "epoch": 7.0, | |
| "eval_loss": 1.8221737146377563, | |
| "eval_runtime": 20.3992, | |
| "eval_samples_per_second": 22.011, | |
| "eval_steps_per_second": 2.794, | |
| "step": 2226 | |
| }, | |
| { | |
| "epoch": 7.01, | |
| "learning_rate": 5.355537834497188e-08, | |
| "loss": 1.248, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 7.03, | |
| "learning_rate": 2.709807965961209e-07, | |
| "loss": 1.3298, | |
| "step": 2235 | |
| }, | |
| { | |
| "epoch": 7.04, | |
| "learning_rate": 6.550934693944553e-07, | |
| "loss": 1.4704, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 7.06, | |
| "learning_rate": 1.2049563616762301e-06, | |
| "loss": 1.4347, | |
| "step": 2245 | |
| }, | |
| { | |
| "epoch": 7.08, | |
| "learning_rate": 1.919228094083792e-06, | |
| "loss": 1.3554, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 7.09, | |
| "learning_rate": 2.7961662152331403e-06, | |
| "loss": 1.428, | |
| "step": 2255 | |
| }, | |
| { | |
| "epoch": 7.11, | |
| "learning_rate": 3.833631452388745e-06, | |
| "loss": 1.4177, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 7.12, | |
| "learning_rate": 5.029092930176208e-06, | |
| "loss": 1.3861, | |
| "step": 2265 | |
| }, | |
| { | |
| "epoch": 7.14, | |
| "learning_rate": 6.379634344600846e-06, | |
| "loss": 1.3699, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 7.15, | |
| "learning_rate": 7.88196107731231e-06, | |
| "loss": 1.4377, | |
| "step": 2275 | |
| }, | |
| { | |
| "epoch": 7.17, | |
| "learning_rate": 9.532408232759425e-06, | |
| "loss": 1.4078, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 7.19, | |
| "learning_rate": 1.1326949578627714e-05, | |
| "loss": 1.4196, | |
| "step": 2285 | |
| }, | |
| { | |
| "epoch": 7.2, | |
| "learning_rate": 1.3261207367752312e-05, | |
| "loss": 1.4826, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 7.22, | |
| "learning_rate": 1.533046301754235e-05, | |
| "loss": 1.4701, | |
| "step": 2295 | |
| }, | |
| { | |
| "epoch": 7.23, | |
| "learning_rate": 1.7529668620867698e-05, | |
| "loss": 1.4641, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 7.25, | |
| "learning_rate": 1.985345926032643e-05, | |
| "loss": 1.4294, | |
| "step": 2305 | |
| }, | |
| { | |
| "epoch": 7.26, | |
| "learning_rate": 2.22961660958507e-05, | |
| "loss": 1.4558, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 7.28, | |
| "learning_rate": 2.4851830193728118e-05, | |
| "loss": 1.5906, | |
| "step": 2315 | |
| }, | |
| { | |
| "epoch": 7.3, | |
| "learning_rate": 2.751421706329902e-05, | |
| "loss": 1.6085, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 7.31, | |
| "learning_rate": 3.027683186586924e-05, | |
| "loss": 1.2953, | |
| "step": 2325 | |
| }, | |
| { | |
| "epoch": 7.33, | |
| "learning_rate": 3.313293525873851e-05, | |
| "loss": 1.4112, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 7.34, | |
| "learning_rate": 3.607555983568874e-05, | |
| "loss": 1.3934, | |
| "step": 2335 | |
| }, | |
| { | |
| "epoch": 7.36, | |
| "learning_rate": 3.909752712383054e-05, | |
| "loss": 1.3835, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 7.37, | |
| "learning_rate": 4.219146509534274e-05, | |
| "loss": 1.4961, | |
| "step": 2345 | |
| }, | |
| { | |
| "epoch": 7.39, | |
| "learning_rate": 4.5349826151383146e-05, | |
| "loss": 1.2224, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 7.41, | |
| "learning_rate": 4.856490553430337e-05, | |
| "loss": 1.4432, | |
| "step": 2355 | |
| }, | |
| { | |
| "epoch": 7.42, | |
| "learning_rate": 5.182886012324924e-05, | |
| "loss": 1.5696, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 7.44, | |
| "learning_rate": 5.5133727567293246e-05, | |
| "loss": 1.6135, | |
| "step": 2365 | |
| }, | |
| { | |
| "epoch": 7.45, | |
| "learning_rate": 5.8471445709429775e-05, | |
| "loss": 1.3836, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 7.47, | |
| "learning_rate": 6.18338722540417e-05, | |
| "loss": 1.0351, | |
| "step": 2375 | |
| }, | |
| { | |
| "epoch": 7.48, | |
| "learning_rate": 6.521280462986592e-05, | |
| "loss": 1.5105, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 7.5, | |
| "learning_rate": 6.859999999999982e-05, | |
| "loss": 1.6121, | |
| "step": 2385 | |
| }, | |
| { | |
| "epoch": 7.52, | |
| "learning_rate": 7.198719537013396e-05, | |
| "loss": 1.6241, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 7.53, | |
| "learning_rate": 7.536612774595818e-05, | |
| "loss": 1.522, | |
| "step": 2395 | |
| }, | |
| { | |
| "epoch": 7.55, | |
| "learning_rate": 7.872855429057012e-05, | |
| "loss": 1.4888, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 7.56, | |
| "learning_rate": 8.206627243270664e-05, | |
| "loss": 1.4955, | |
| "step": 2405 | |
| }, | |
| { | |
| "epoch": 7.58, | |
| "learning_rate": 8.537113987675064e-05, | |
| "loss": 1.4161, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 7.59, | |
| "learning_rate": 8.86350944656965e-05, | |
| "loss": 1.8936, | |
| "step": 2415 | |
| }, | |
| { | |
| "epoch": 7.61, | |
| "learning_rate": 9.185017384861673e-05, | |
| "loss": 1.6053, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 7.63, | |
| "learning_rate": 9.500853490465716e-05, | |
| "loss": 1.5626, | |
| "step": 2425 | |
| }, | |
| { | |
| "epoch": 7.64, | |
| "learning_rate": 9.810247287616934e-05, | |
| "loss": 1.1654, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 7.66, | |
| "learning_rate": 0.00010112444016431114, | |
| "loss": 1.353, | |
| "step": 2435 | |
| }, | |
| { | |
| "epoch": 7.67, | |
| "learning_rate": 0.00010406706474126137, | |
| "loss": 1.5396, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 7.69, | |
| "learning_rate": 0.00010692316813413065, | |
| "loss": 1.3699, | |
| "step": 2445 | |
| }, | |
| { | |
| "epoch": 7.7, | |
| "learning_rate": 0.0001096857829367009, | |
| "loss": 1.3597, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 7.72, | |
| "learning_rate": 0.0001123481698062718, | |
| "loss": 1.5768, | |
| "step": 2455 | |
| }, | |
| { | |
| "epoch": 7.74, | |
| "learning_rate": 0.00011490383390414922, | |
| "loss": 1.3661, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 7.75, | |
| "learning_rate": 0.00011734654073967348, | |
| "loss": 1.4304, | |
| "step": 2465 | |
| }, | |
| { | |
| "epoch": 7.77, | |
| "learning_rate": 0.00011967033137913221, | |
| "loss": 1.4475, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 7.78, | |
| "learning_rate": 0.00012186953698245757, | |
| "loss": 1.5644, | |
| "step": 2475 | |
| }, | |
| { | |
| "epoch": 7.8, | |
| "learning_rate": 0.00012393879263224763, | |
| "loss": 1.2184, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 7.81, | |
| "learning_rate": 0.00012587305042137222, | |
| "loss": 1.55, | |
| "step": 2485 | |
| }, | |
| { | |
| "epoch": 7.83, | |
| "learning_rate": 0.00012766759176724053, | |
| "loss": 1.5614, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 7.85, | |
| "learning_rate": 0.00012931803892268765, | |
| "loss": 1.4788, | |
| "step": 2495 | |
| }, | |
| { | |
| "epoch": 7.86, | |
| "learning_rate": 0.0001308203656553991, | |
| "loss": 1.3325, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 7.88, | |
| "learning_rate": 0.00013217090706982374, | |
| "loss": 1.6197, | |
| "step": 2505 | |
| }, | |
| { | |
| "epoch": 7.89, | |
| "learning_rate": 0.0001333663685476112, | |
| "loss": 1.5014, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 7.91, | |
| "learning_rate": 0.00013440383378476682, | |
| "loss": 1.4764, | |
| "step": 2515 | |
| }, | |
| { | |
| "epoch": 7.92, | |
| "learning_rate": 0.00013528077190591619, | |
| "loss": 1.8615, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 7.94, | |
| "learning_rate": 0.00013599504363832375, | |
| "loss": 1.2448, | |
| "step": 2525 | |
| }, | |
| { | |
| "epoch": 7.96, | |
| "learning_rate": 0.00013654490653060552, | |
| "loss": 1.3242, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 7.97, | |
| "learning_rate": 0.00013692901920340386, | |
| "loss": 1.3224, | |
| "step": 2535 | |
| }, | |
| { | |
| "epoch": 7.99, | |
| "learning_rate": 0.00013714644462165502, | |
| "loss": 1.4513, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "eval_loss": 1.8965427875518799, | |
| "eval_runtime": 20.4074, | |
| "eval_samples_per_second": 22.002, | |
| "eval_steps_per_second": 2.793, | |
| "step": 2544 | |
| }, | |
| { | |
| "epoch": 8.0, | |
| "learning_rate": 0.00013719665238046719, | |
| "loss": 1.3932, | |
| "step": 2545 | |
| }, | |
| { | |
| "epoch": 8.02, | |
| "learning_rate": 0.00013707951999903246, | |
| "loss": 1.2432, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 8.03, | |
| "learning_rate": 0.00013679533321941633, | |
| "loss": 1.2907, | |
| "step": 2555 | |
| }, | |
| { | |
| "epoch": 8.05, | |
| "learning_rate": 0.00013634478530949573, | |
| "loss": 1.2857, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 8.07, | |
| "learning_rate": 0.00013572897537174604, | |
| "loss": 1.3178, | |
| "step": 2565 | |
| }, | |
| { | |
| "epoch": 8.08, | |
| "learning_rate": 0.00013494940566200278, | |
| "loss": 1.3444, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 8.1, | |
| "learning_rate": 0.0001340079779247397, | |
| "loss": 1.3778, | |
| "step": 2575 | |
| }, | |
| { | |
| "epoch": 8.11, | |
| "learning_rate": 0.00013290698875380075, | |
| "loss": 1.5204, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 8.13, | |
| "learning_rate": 0.00013164912398990663, | |
| "loss": 1.196, | |
| "step": 2585 | |
| }, | |
| { | |
| "epoch": 8.14, | |
| "learning_rate": 0.00013023745216860123, | |
| "loss": 1.5108, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 8.16, | |
| "learning_rate": 0.00012867541703462073, | |
| "loss": 1.414, | |
| "step": 2595 | |
| }, | |
| { | |
| "epoch": 8.18, | |
| "learning_rate": 0.00012696682914094848, | |
| "loss": 1.299, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 8.19, | |
| "learning_rate": 0.00012511585655304892, | |
| "loss": 1.5168, | |
| "step": 2605 | |
| }, | |
| { | |
| "epoch": 8.21, | |
| "learning_rate": 0.00012312701468095605, | |
| "loss": 1.4117, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 8.22, | |
| "learning_rate": 0.0001210051552640223, | |
| "loss": 1.4762, | |
| "step": 2615 | |
| }, | |
| { | |
| "epoch": 8.24, | |
| "learning_rate": 0.00011875545453519897, | |
| "loss": 1.3127, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 8.25, | |
| "learning_rate": 0.00011638340059372173, | |
| "loss": 1.5447, | |
| "step": 2625 | |
| }, | |
| { | |
| "epoch": 8.27, | |
| "learning_rate": 0.00011389478001700295, | |
| "loss": 1.3891, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 8.29, | |
| "learning_rate": 0.00011129566374439389, | |
| "loss": 1.7022, | |
| "step": 2635 | |
| }, | |
| { | |
| "epoch": 8.3, | |
| "learning_rate": 0.00010859239226725287, | |
| "loss": 1.2989, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 8.32, | |
| "learning_rate": 0.00010579156016144313, | |
| "loss": 1.4672, | |
| "step": 2645 | |
| }, | |
| { | |
| "epoch": 8.33, | |
| "learning_rate": 0.00010289999999999993, | |
| "loss": 1.4019, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 8.35, | |
| "learning_rate": 9.992476568520875e-05, | |
| "loss": 1.3174, | |
| "step": 2655 | |
| }, | |
| { | |
| "epoch": 8.36, | |
| "learning_rate": 9.687311524075153e-05, | |
| "loss": 1.4993, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 8.38, | |
| "learning_rate": 9.375249310590639e-05, | |
| "loss": 1.4268, | |
| "step": 2665 | |
| }, | |
| { | |
| "epoch": 8.4, | |
| "learning_rate": 9.057051197499064e-05, | |
| "loss": 1.3762, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 8.41, | |
| "learning_rate": 8.733493422634493e-05, | |
| "loss": 1.6376, | |
| "step": 2675 | |
| }, | |
| { | |
| "epoch": 8.43, | |
| "learning_rate": 8.405365298617051e-05, | |
| "loss": 1.3535, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 8.44, | |
| "learning_rate": 8.073467287340865e-05, | |
| "loss": 1.4565, | |
| "step": 2685 | |
| }, | |
| { | |
| "epoch": 8.46, | |
| "learning_rate": 7.738609047263366e-05, | |
| "loss": 1.4825, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 8.47, | |
| "learning_rate": 7.401607458260121e-05, | |
| "loss": 1.2098, | |
| "step": 2695 | |
| }, | |
| { | |
| "epoch": 8.49, | |
| "learning_rate": 7.063284628862766e-05, | |
| "loss": 1.114, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 8.51, | |
| "learning_rate": 6.724465890742147e-05, | |
| "loss": 1.1928, | |
| "step": 2705 | |
| }, | |
| { | |
| "epoch": 8.52, | |
| "learning_rate": 6.385977785328291e-05, | |
| "loss": 1.2321, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 8.54, | |
| "learning_rate": 6.048646047479429e-05, | |
| "loss": 1.4018, | |
| "step": 2715 | |
| }, | |
| { | |
| "epoch": 8.55, | |
| "learning_rate": 5.71329359111875e-05, | |
| "loss": 1.4829, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 8.57, | |
| "learning_rate": 5.3807385017523074e-05, | |
| "loss": 1.3501, | |
| "step": 2725 | |
| }, | |
| { | |
| "epoch": 8.58, | |
| "learning_rate": 5.051792040766187e-05, | |
| "loss": 1.344, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 8.6, | |
| "learning_rate": 4.7272566663709795e-05, | |
| "loss": 1.0449, | |
| "step": 2735 | |
| }, | |
| { | |
| "epoch": 8.62, | |
| "learning_rate": 4.407924076020983e-05, | |
| "loss": 1.3851, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 8.63, | |
| "learning_rate": 4.0945732750844954e-05, | |
| "loss": 1.3359, | |
| "step": 2745 | |
| }, | |
| { | |
| "epoch": 8.65, | |
| "learning_rate": 3.787968676476165e-05, | |
| "loss": 1.2086, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 8.66, | |
| "learning_rate": 3.488858235886944e-05, | |
| "loss": 1.2969, | |
| "step": 2755 | |
| }, | |
| { | |
| "epoch": 8.68, | |
| "learning_rate": 3.1979716271615364e-05, | |
| "loss": 1.1869, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 8.69, | |
| "learning_rate": 2.916018462273902e-05, | |
| "loss": 1.4603, | |
| "step": 2765 | |
| }, | |
| { | |
| "epoch": 8.71, | |
| "learning_rate": 2.643686560243124e-05, | |
| "loss": 1.2891, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 8.73, | |
| "learning_rate": 2.3816402692130164e-05, | |
| "loss": 1.4753, | |
| "step": 2775 | |
| }, | |
| { | |
| "epoch": 8.74, | |
| "learning_rate": 2.1305188457882628e-05, | |
| "loss": 1.4468, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 8.76, | |
| "learning_rate": 1.890934895580877e-05, | |
| "loss": 1.2143, | |
| "step": 2785 | |
| }, | |
| { | |
| "epoch": 8.77, | |
| "learning_rate": 1.6634728787713087e-05, | |
| "loss": 1.4829, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 8.79, | |
| "learning_rate": 1.4486876843296578e-05, | |
| "loss": 1.5124, | |
| "step": 2795 | |
| }, | |
| { | |
| "epoch": 8.81, | |
| "learning_rate": 1.247103276375423e-05, | |
| "loss": 1.1507, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 8.82, | |
| "learning_rate": 1.0592114159774876e-05, | |
| "loss": 1.2259, | |
| "step": 2805 | |
| }, | |
| { | |
| "epoch": 8.84, | |
| "learning_rate": 8.854704615130826e-06, | |
| "loss": 1.3434, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 8.85, | |
| "learning_rate": 7.263042505119003e-06, | |
| "loss": 1.4906, | |
| "step": 2815 | |
| }, | |
| { | |
| "epoch": 8.87, | |
| "learning_rate": 5.82101065712901e-06, | |
| "loss": 1.2034, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 8.88, | |
| "learning_rate": 4.532126878565386e-06, | |
| "loss": 1.3118, | |
| "step": 2825 | |
| }, | |
| { | |
| "epoch": 8.9, | |
| "learning_rate": 3.3995353752283744e-06, | |
| "loss": 1.4476, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 8.92, | |
| "learning_rate": 2.4259990810866283e-06, | |
| "loss": 1.5325, | |
| "step": 2835 | |
| }, | |
| { | |
| "epoch": 8.93, | |
| "learning_rate": 1.6138929181565955e-06, | |
| "loss": 1.3443, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 8.95, | |
| "learning_rate": 9.651980029285464e-07, | |
| "loss": 1.665, | |
| "step": 2845 | |
| }, | |
| { | |
| "epoch": 8.96, | |
| "learning_rate": 4.814968134727699e-07, | |
| "loss": 1.3313, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 8.98, | |
| "learning_rate": 1.6396932901667525e-07, | |
| "loss": 1.504, | |
| "step": 2855 | |
| }, | |
| { | |
| "epoch": 8.99, | |
| "learning_rate": 1.339015140952895e-08, | |
| "loss": 1.2679, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 9.0, | |
| "eval_loss": 1.7906934022903442, | |
| "eval_runtime": 20.4159, | |
| "eval_samples_per_second": 21.993, | |
| "eval_steps_per_second": 2.792, | |
| "step": 2862 | |
| }, | |
| { | |
| "epoch": 9.01, | |
| "learning_rate": 3.01266154967001e-08, | |
| "loss": 1.4496, | |
| "step": 2865 | |
| }, | |
| { | |
| "epoch": 9.03, | |
| "learning_rate": 2.1413789301401304e-07, | |
| "loss": 1.2338, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 9.04, | |
| "learning_rate": 5.649750921872831e-07, | |
| "loss": 1.1185, | |
| "step": 2875 | |
| }, | |
| { | |
| "epoch": 9.06, | |
| "learning_rate": 1.0817823527948346e-06, | |
| "loss": 1.0349, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 9.07, | |
| "learning_rate": 1.7632989340210505e-06, | |
| "loss": 1.1285, | |
| "step": 2885 | |
| }, | |
| { | |
| "epoch": 9.09, | |
| "learning_rate": 2.6078622900081364e-06, | |
| "loss": 1.1348, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 9.1, | |
| "learning_rate": 3.613412125602455e-06, | |
| "loss": 1.0489, | |
| "step": 2895 | |
| }, | |
| { | |
| "epoch": 9.12, | |
| "learning_rate": 4.777495422403447e-06, | |
| "loss": 1.1893, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 9.14, | |
| "learning_rate": 6.0972724228519425e-06, | |
| "loss": 1.3905, | |
| "step": 2905 | |
| }, | |
| { | |
| "epoch": 9.15, | |
| "learning_rate": 7.569523557760349e-06, | |
| "loss": 1.2451, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 9.17, | |
| "learning_rate": 9.190657300387452e-06, | |
| "loss": 1.2443, | |
| "step": 2915 | |
| }, | |
| { | |
| "epoch": 9.18, | |
| "learning_rate": 1.0956718927894355e-05, | |
| "loss": 1.2626, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 9.2, | |
| "learning_rate": 1.2863400168810178e-05, | |
| "loss": 1.0339, | |
| "step": 2925 | |
| }, | |
| { | |
| "epoch": 9.21, | |
| "learning_rate": 1.4906049712973553e-05, | |
| "loss": 1.1284, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 9.23, | |
| "learning_rate": 1.7079684558309144e-05, | |
| "loss": 1.2964, | |
| "step": 2935 | |
| }, | |
| { | |
| "epoch": 9.25, | |
| "learning_rate": 1.9379002166758836e-05, | |
| "loss": 0.9701, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 9.26, | |
| "learning_rate": 2.1798393399717496e-05, | |
| "loss": 1.2379, | |
| "step": 2945 | |
| }, | |
| { | |
| "epoch": 9.28, | |
| "learning_rate": 2.4331956201411865e-05, | |
| "loss": 1.2294, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 9.29, | |
| "learning_rate": 2.69735099968445e-05, | |
| "loss": 1.2279, | |
| "step": 2955 | |
| }, | |
| { | |
| "epoch": 9.31, | |
| "learning_rate": 2.9716610769183196e-05, | |
| "loss": 1.1694, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 9.32, | |
| "learning_rate": 3.255456677980771e-05, | |
| "loss": 1.3376, | |
| "step": 2965 | |
| }, | |
| { | |
| "epoch": 9.34, | |
| "learning_rate": 3.548045489267097e-05, | |
| "loss": 1.1979, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 9.36, | |
| "learning_rate": 3.848713746314718e-05, | |
| "loss": 1.0518, | |
| "step": 2975 | |
| }, | |
| { | |
| "epoch": 9.37, | |
| "learning_rate": 4.1567279750173456e-05, | |
| "loss": 1.1192, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 9.39, | |
| "learning_rate": 4.471336780920016e-05, | |
| "loss": 1.258, | |
| "step": 2985 | |
| }, | |
| { | |
| "epoch": 9.4, | |
| "learning_rate": 4.791772682230585e-05, | |
| "loss": 1.1698, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 9.42, | |
| "learning_rate": 5.1172539820760084e-05, | |
| "loss": 1.2648, | |
| "step": 2995 | |
| }, | |
| { | |
| "epoch": 9.43, | |
| "learning_rate": 5.4469866754360636e-05, | |
| "loss": 1.1929, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 9.45, | |
| "learning_rate": 5.7801663861022995e-05, | |
| "loss": 1.1723, | |
| "step": 3005 | |
| }, | |
| { | |
| "epoch": 9.47, | |
| "learning_rate": 6.115980328937633e-05, | |
| "loss": 1.225, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 9.48, | |
| "learning_rate": 6.453609292649126e-05, | |
| "loss": 1.281, | |
| "step": 3015 | |
| }, | |
| { | |
| "epoch": 9.5, | |
| "learning_rate": 6.792229638237138e-05, | |
| "loss": 1.1719, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 9.51, | |
| "learning_rate": 7.131015308246217e-05, | |
| "loss": 1.3758, | |
| "step": 3025 | |
| }, | |
| { | |
| "epoch": 9.53, | |
| "learning_rate": 7.469139841915347e-05, | |
| "loss": 1.3241, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 9.54, | |
| "learning_rate": 7.805778391312036e-05, | |
| "loss": 1.1407, | |
| "step": 3035 | |
| }, | |
| { | |
| "epoch": 9.56, | |
| "learning_rate": 8.140109733532393e-05, | |
| "loss": 1.2411, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 9.58, | |
| "learning_rate": 8.471318274057568e-05, | |
| "loss": 1.413, | |
| "step": 3045 | |
| }, | |
| { | |
| "epoch": 9.59, | |
| "learning_rate": 8.798596036379883e-05, | |
| "loss": 1.222, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 9.61, | |
| "learning_rate": 9.121144633045388e-05, | |
| "loss": 1.3584, | |
| "step": 3055 | |
| }, | |
| { | |
| "epoch": 9.62, | |
| "learning_rate": 9.438177213303655e-05, | |
| "loss": 1.1437, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 9.64, | |
| "learning_rate": 9.74892038261427e-05, | |
| "loss": 1.2668, | |
| "step": 3065 | |
| }, | |
| { | |
| "epoch": 9.65, | |
| "learning_rate": 0.00010052616089327018, | |
| "loss": 1.3637, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 9.67, | |
| "learning_rate": 0.00010348523473933642, | |
| "loss": 1.1909, | |
| "step": 3075 | |
| }, | |
| { | |
| "epoch": 9.69, | |
| "learning_rate": 0.0001063592067637941, | |
| "loss": 1.3763, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 9.7, | |
| "learning_rate": 0.00010914106597026132, | |
| "loss": 1.4548, | |
| "step": 3085 | |
| }, | |
| { | |
| "epoch": 9.72, | |
| "learning_rate": 0.00011182402606970514, | |
| "loss": 1.3378, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 9.73, | |
| "learning_rate": 0.00011440154203545576, | |
| "loss": 1.3677, | |
| "step": 3095 | |
| }, | |
| { | |
| "epoch": 9.75, | |
| "learning_rate": 0.00011686732606966472, | |
| "loss": 1.2485, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 9.76, | |
| "learning_rate": 0.00011921536294226233, | |
| "loss": 1.3363, | |
| "step": 3105 | |
| }, | |
| { | |
| "epoch": 9.78, | |
| "learning_rate": 0.00012143992466498816, | |
| "loss": 1.1575, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 9.8, | |
| "learning_rate": 0.00012353558446470098, | |
| "loss": 1.1208, | |
| "step": 3115 | |
| }, | |
| { | |
| "epoch": 9.81, | |
| "learning_rate": 0.00012549723002188375, | |
| "loss": 1.4087, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 9.83, | |
| "learning_rate": 0.00012732007594204206, | |
| "loss": 1.3399, | |
| "step": 3125 | |
| }, | |
| { | |
| "epoch": 9.84, | |
| "learning_rate": 0.00012899967542957612, | |
| "loss": 1.4988, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 9.86, | |
| "learning_rate": 0.00013053193113564998, | |
| "loss": 0.989, | |
| "step": 3135 | |
| }, | |
| { | |
| "epoch": 9.87, | |
| "learning_rate": 0.00013191310515358958, | |
| "loss": 1.349, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 9.89, | |
| "learning_rate": 0.00013313982813743067, | |
| "loss": 1.1874, | |
| "step": 3145 | |
| }, | |
| { | |
| "epoch": 9.91, | |
| "learning_rate": 0.00013420910752136937, | |
| "loss": 1.2978, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 9.92, | |
| "learning_rate": 0.00013511833482006638, | |
| "loss": 1.5572, | |
| "step": 3155 | |
| }, | |
| { | |
| "epoch": 9.94, | |
| "learning_rate": 0.00013586529199199334, | |
| "loss": 1.3689, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 9.95, | |
| "learning_rate": 0.00013644815685030044, | |
| "loss": 1.1297, | |
| "step": 3165 | |
| }, | |
| { | |
| "epoch": 9.97, | |
| "learning_rate": 0.00013686550750800414, | |
| "loss": 1.3856, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 9.98, | |
| "learning_rate": 0.00013711632584665164, | |
| "loss": 1.2732, | |
| "step": 3175 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "learning_rate": 0.0001372, | |
| "loss": 1.5975, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_loss": 1.7831153869628906, | |
| "eval_runtime": 20.3518, | |
| "eval_samples_per_second": 22.062, | |
| "eval_steps_per_second": 2.801, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 9.95, | |
| "learning_rate": 0.00013645750858358395, | |
| "loss": 1.2433, | |
| "step": 3185 | |
| }, | |
| { | |
| "epoch": 9.97, | |
| "learning_rate": 0.0001368696722497127, | |
| "loss": 1.547, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 9.98, | |
| "learning_rate": 0.00013711736829567482, | |
| "loss": 1.4594, | |
| "step": 3195 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "learning_rate": 0.0001372, | |
| "loss": 1.3407, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "eval_loss": 1.139600157737732, | |
| "eval_runtime": 5.2723, | |
| "eval_samples_per_second": 82.317, | |
| "eval_steps_per_second": 10.432, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 10.02, | |
| "learning_rate": 0.00013711736829567482, | |
| "loss": 1.4415, | |
| "step": 3205 | |
| }, | |
| { | |
| "epoch": 10.03, | |
| "learning_rate": 0.00013686967224971273, | |
| "loss": 1.2348, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 10.05, | |
| "learning_rate": 0.00013645750858358398, | |
| "loss": 1.4623, | |
| "step": 3215 | |
| }, | |
| { | |
| "epoch": 10.06, | |
| "learning_rate": 0.00013588187023566163, | |
| "loss": 1.437, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 10.08, | |
| "learning_rate": 0.00013514414396914573, | |
| "loss": 1.6916, | |
| "step": 3225 | |
| }, | |
| { | |
| "epoch": 10.09, | |
| "learning_rate": 0.00013424610703122958, | |
| "loss": 1.7023, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 10.11, | |
| "learning_rate": 0.00013318992287155525, | |
| "loss": 1.3172, | |
| "step": 3235 | |
| }, | |
| { | |
| "epoch": 10.12, | |
| "learning_rate": 0.00013197813593027435, | |
| "loss": 1.2053, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 10.14, | |
| "learning_rate": 0.00013061366550826825, | |
| "loss": 1.1869, | |
| "step": 3245 | |
| }, | |
| { | |
| "epoch": 10.16, | |
| "learning_rate": 0.00012909979873429724, | |
| "loss": 1.2981, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 10.17, | |
| "learning_rate": 0.0001274401826460187, | |
| "loss": 1.6608, | |
| "step": 3255 | |
| }, | |
| { | |
| "epoch": 10.19, | |
| "learning_rate": 0.00012563881540395474, | |
| "loss": 1.3115, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 10.2, | |
| "learning_rate": 0.00012370003665957216, | |
| "loss": 1.2824, | |
| "step": 3265 | |
| }, | |
| { | |
| "epoch": 10.22, | |
| "learning_rate": 0.00012162851710068375, | |
| "loss": 1.4082, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 10.23, | |
| "learning_rate": 0.00011942924719935029, | |
| "loss": 1.3048, | |
| "step": 3275 | |
| }, | |
| { | |
| "epoch": 10.25, | |
| "learning_rate": 0.00011710752518939736, | |
| "loss": 1.3276, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 10.27, | |
| "learning_rate": 0.0001146689443025054, | |
| "loss": 1.4064, | |
| "step": 3285 | |
| }, | |
| { | |
| "epoch": 10.28, | |
| "learning_rate": 0.00011211937929362613, | |
| "loss": 1.2408, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 10.3, | |
| "learning_rate": 0.00010946497228818107, | |
| "loss": 1.3932, | |
| "step": 3295 | |
| }, | |
| { | |
| "epoch": 10.31, | |
| "learning_rate": 0.00010671211798514499, | |
| "loss": 1.4576, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 10.33, | |
| "learning_rate": 0.00010386744825165496, | |
| "loss": 1.455, | |
| "step": 3305 | |
| }, | |
| { | |
| "epoch": 10.34, | |
| "learning_rate": 0.00010093781614626351, | |
| "loss": 1.3289, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 10.36, | |
| "learning_rate": 9.793027940931756e-05, | |
| "loss": 1.2645, | |
| "step": 3315 | |
| }, | |
| { | |
| "epoch": 10.38, | |
| "learning_rate": 9.485208346024504e-05, | |
| "loss": 1.39, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 10.39, | |
| "learning_rate": 9.17106439427063e-05, | |
| "loss": 1.3945, | |
| "step": 3325 | |
| }, | |
| { | |
| "epoch": 10.41, | |
| "learning_rate": 8.851352885965625e-05, | |
| "loss": 1.5375, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 10.42, | |
| "learning_rate": 8.526844034136417e-05, | |
| "loss": 1.4077, | |
| "step": 3335 | |
| }, | |
| { | |
| "epoch": 10.44, | |
| "learning_rate": 8.198319609030632e-05, | |
| "loss": 1.4331, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 10.45, | |
| "learning_rate": 7.866571054763788e-05, | |
| "loss": 1.8602, | |
| "step": 3345 | |
| }, | |
| { | |
| "epoch": 10.47, | |
| "learning_rate": 7.532397582660805e-05, | |
| "loss": 1.4865, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 10.48, | |
| "learning_rate": 7.19660424588612e-05, | |
| "loss": 1.2815, | |
| "step": 3355 | |
| }, | |
| { | |
| "epoch": 10.5, | |
| "learning_rate": 6.859999999999997e-05, | |
| "loss": 1.4705, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 10.52, | |
| "learning_rate": 6.523395754113922e-05, | |
| "loss": 1.1969, | |
| "step": 3365 | |
| }, | |
| { | |
| "epoch": 10.53, | |
| "learning_rate": 6.187602417339237e-05, | |
| "loss": 1.4564, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 10.55, | |
| "learning_rate": 5.853428945236207e-05, | |
| "loss": 1.4113, | |
| "step": 3375 | |
| }, | |
| { | |
| "epoch": 10.56, | |
| "learning_rate": 5.521680390969362e-05, | |
| "loss": 1.4642, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 10.58, | |
| "learning_rate": 5.193155965863624e-05, | |
| "loss": 1.4196, | |
| "step": 3385 | |
| }, | |
| { | |
| "epoch": 10.59, | |
| "learning_rate": 4.8686471140344147e-05, | |
| "loss": 1.3666, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 10.61, | |
| "learning_rate": 4.548935605729363e-05, | |
| "loss": 1.3908, | |
| "step": 3395 | |
| }, | |
| { | |
| "epoch": 10.62, | |
| "learning_rate": 4.23479165397549e-05, | |
| "loss": 1.4785, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 10.64, | |
| "learning_rate": 3.926972059068282e-05, | |
| "loss": 1.4775, | |
| "step": 3405 | |
| }, | |
| { | |
| "epoch": 10.66, | |
| "learning_rate": 3.626218385373685e-05, | |
| "loss": 1.4841, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 10.67, | |
| "learning_rate": 3.333255174834496e-05, | |
| "loss": 1.4263, | |
| "step": 3415 | |
| }, | |
| { | |
| "epoch": 10.69, | |
| "learning_rate": 3.0487882014855373e-05, | |
| "loss": 1.4815, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 10.7, | |
| "learning_rate": 2.7735027711819264e-05, | |
| "loss": 1.3612, | |
| "step": 3425 | |
| }, | |
| { | |
| "epoch": 10.72, | |
| "learning_rate": 2.508062070637383e-05, | |
| "loss": 1.3586, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 10.73, | |
| "learning_rate": 2.253105569749455e-05, | |
| "loss": 1.4036, | |
| "step": 3435 | |
| }, | |
| { | |
| "epoch": 10.75, | |
| "learning_rate": 2.0092474810602945e-05, | |
| "loss": 1.2455, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 10.77, | |
| "learning_rate": 1.7770752800649997e-05, | |
| "loss": 1.3747, | |
| "step": 3445 | |
| }, | |
| { | |
| "epoch": 10.78, | |
| "learning_rate": 1.5571482899316204e-05, | |
| "loss": 1.2848, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 10.8, | |
| "learning_rate": 1.3499963340427795e-05, | |
| "loss": 1.5623, | |
| "step": 3455 | |
| }, | |
| { | |
| "epoch": 10.81, | |
| "learning_rate": 1.1561184596045504e-05, | |
| "loss": 1.4704, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 10.83, | |
| "learning_rate": 9.759817353981509e-06, | |
| "loss": 1.3271, | |
| "step": 3465 | |
| }, | |
| { | |
| "epoch": 10.84, | |
| "learning_rate": 8.100201265702836e-06, | |
| "loss": 1.2696, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 10.86, | |
| "learning_rate": 6.586334491731833e-06, | |
| "loss": 1.5138, | |
| "step": 3475 | |
| }, | |
| { | |
| "epoch": 10.88, | |
| "learning_rate": 5.221864069725821e-06, | |
| "loss": 1.344, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 10.89, | |
| "learning_rate": 4.010077128444735e-06, | |
| "loss": 1.3544, | |
| "step": 3485 | |
| }, | |
| { | |
| "epoch": 10.91, | |
| "learning_rate": 2.9538929687704825e-06, | |
| "loss": 1.6602, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 10.92, | |
| "learning_rate": 2.0558560308543213e-06, | |
| "loss": 1.3761, | |
| "step": 3495 | |
| }, | |
| { | |
| "epoch": 10.94, | |
| "learning_rate": 1.3181297643384459e-06, | |
| "loss": 1.3709, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 10.95, | |
| "learning_rate": 7.424914164160148e-07, | |
| "loss": 1.3595, | |
| "step": 3505 | |
| }, | |
| { | |
| "epoch": 10.97, | |
| "learning_rate": 3.303277502872983e-07, | |
| "loss": 1.4077, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 10.98, | |
| "learning_rate": 8.263170432518063e-08, | |
| "loss": 1.4356, | |
| "step": 3515 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "learning_rate": 0.0, | |
| "loss": 1.7243, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_loss": 1.1156859397888184, | |
| "eval_runtime": 5.2715, | |
| "eval_samples_per_second": 82.33, | |
| "eval_steps_per_second": 10.433, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 10.95, | |
| "learning_rate": 9.414215321223168e-07, | |
| "loss": 1.3031, | |
| "step": 3525 | |
| }, | |
| { | |
| "epoch": 10.96, | |
| "learning_rate": 4.696220449804098e-07, | |
| "loss": 1.0973, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 10.98, | |
| "learning_rate": 1.5992243352901425e-07, | |
| "loss": 1.1205, | |
| "step": 3535 | |
| }, | |
| { | |
| "epoch": 10.99, | |
| "learning_rate": 1.3059553632214649e-08, | |
| "loss": 1.5828, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 11.0, | |
| "eval_loss": 1.0779144763946533, | |
| "eval_runtime": 8.8469, | |
| "eval_samples_per_second": 46.57, | |
| "eval_steps_per_second": 5.878, | |
| "step": 3542 | |
| } | |
| ], | |
| "max_steps": 3864, | |
| "num_train_epochs": 12, | |
| "total_flos": 3691011244032000.0, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |