| { | |
| "best_metric": 0.9837209302325581, | |
| "best_model_checkpoint": "./vit_finetuned_synset/checkpoint-5000", | |
| "epoch": 6.0, | |
| "eval_steps": 1000, | |
| "global_step": 5160, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.011627906976744186, | |
| "grad_norm": 9.393110275268555, | |
| "learning_rate": 3.488372093023256e-07, | |
| "loss": 3.7721, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.023255813953488372, | |
| "grad_norm": 9.336662292480469, | |
| "learning_rate": 7.364341085271319e-07, | |
| "loss": 3.7166, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.03488372093023256, | |
| "grad_norm": 9.135019302368164, | |
| "learning_rate": 1.1240310077519381e-06, | |
| "loss": 3.6974, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.046511627906976744, | |
| "grad_norm": 9.181822776794434, | |
| "learning_rate": 1.5116279069767443e-06, | |
| "loss": 3.7336, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.05813953488372093, | |
| "grad_norm": 9.79133415222168, | |
| "learning_rate": 1.8992248062015506e-06, | |
| "loss": 3.7501, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.06976744186046512, | |
| "grad_norm": 9.377440452575684, | |
| "learning_rate": 2.2868217054263567e-06, | |
| "loss": 3.7906, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.08139534883720931, | |
| "grad_norm": 9.488425254821777, | |
| "learning_rate": 2.635658914728683e-06, | |
| "loss": 3.708, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.09302325581395349, | |
| "grad_norm": 9.35010814666748, | |
| "learning_rate": 3.0232558139534885e-06, | |
| "loss": 3.6929, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.10465116279069768, | |
| "grad_norm": 10.471288681030273, | |
| "learning_rate": 3.372093023255814e-06, | |
| "loss": 3.6442, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.11627906976744186, | |
| "grad_norm": 8.739205360412598, | |
| "learning_rate": 3.7596899224806203e-06, | |
| "loss": 3.664, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.12790697674418605, | |
| "grad_norm": 9.124735832214355, | |
| "learning_rate": 4.1472868217054264e-06, | |
| "loss": 3.6163, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.13953488372093023, | |
| "grad_norm": 8.19368839263916, | |
| "learning_rate": 4.5348837209302326e-06, | |
| "loss": 3.5854, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.1511627906976744, | |
| "grad_norm": 9.776670455932617, | |
| "learning_rate": 4.922480620155039e-06, | |
| "loss": 3.5337, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.16279069767441862, | |
| "grad_norm": 8.68013858795166, | |
| "learning_rate": 5.310077519379846e-06, | |
| "loss": 3.5436, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.1744186046511628, | |
| "grad_norm": 8.815572738647461, | |
| "learning_rate": 5.697674418604652e-06, | |
| "loss": 3.4187, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.18604651162790697, | |
| "grad_norm": 8.750410079956055, | |
| "learning_rate": 6.085271317829458e-06, | |
| "loss": 3.4981, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.19767441860465115, | |
| "grad_norm": 8.929957389831543, | |
| "learning_rate": 6.472868217054265e-06, | |
| "loss": 3.481, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.20930232558139536, | |
| "grad_norm": 10.56398868560791, | |
| "learning_rate": 6.86046511627907e-06, | |
| "loss": 3.4054, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.22093023255813954, | |
| "grad_norm": 9.072072982788086, | |
| "learning_rate": 7.248062015503876e-06, | |
| "loss": 3.317, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.23255813953488372, | |
| "grad_norm": 8.503347396850586, | |
| "learning_rate": 7.635658914728683e-06, | |
| "loss": 3.3538, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.2441860465116279, | |
| "grad_norm": 9.191971778869629, | |
| "learning_rate": 8.023255813953488e-06, | |
| "loss": 3.2774, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.2558139534883721, | |
| "grad_norm": 8.65476131439209, | |
| "learning_rate": 8.410852713178295e-06, | |
| "loss": 3.2667, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.26744186046511625, | |
| "grad_norm": 9.394301414489746, | |
| "learning_rate": 8.7984496124031e-06, | |
| "loss": 3.1822, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.27906976744186046, | |
| "grad_norm": 9.087491035461426, | |
| "learning_rate": 9.186046511627908e-06, | |
| "loss": 3.0977, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.29069767441860467, | |
| "grad_norm": 9.139519691467285, | |
| "learning_rate": 9.573643410852715e-06, | |
| "loss": 3.0849, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.3023255813953488, | |
| "grad_norm": 8.674009323120117, | |
| "learning_rate": 9.96124031007752e-06, | |
| "loss": 2.983, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.313953488372093, | |
| "grad_norm": 8.377427101135254, | |
| "learning_rate": 1.0348837209302327e-05, | |
| "loss": 2.885, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.32558139534883723, | |
| "grad_norm": 8.761062622070312, | |
| "learning_rate": 1.0736434108527132e-05, | |
| "loss": 2.8172, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.3372093023255814, | |
| "grad_norm": 9.388766288757324, | |
| "learning_rate": 1.112403100775194e-05, | |
| "loss": 2.777, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.3488372093023256, | |
| "grad_norm": 8.205259323120117, | |
| "learning_rate": 1.1511627906976746e-05, | |
| "loss": 2.7128, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.36046511627906974, | |
| "grad_norm": 9.509353637695312, | |
| "learning_rate": 1.189922480620155e-05, | |
| "loss": 2.5911, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.37209302325581395, | |
| "grad_norm": 8.547121047973633, | |
| "learning_rate": 1.2286821705426357e-05, | |
| "loss": 2.5844, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.38372093023255816, | |
| "grad_norm": 9.776483535766602, | |
| "learning_rate": 1.2674418604651164e-05, | |
| "loss": 2.4906, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.3953488372093023, | |
| "grad_norm": 8.681229591369629, | |
| "learning_rate": 1.306201550387597e-05, | |
| "loss": 2.4379, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.4069767441860465, | |
| "grad_norm": 8.074151992797852, | |
| "learning_rate": 1.3449612403100776e-05, | |
| "loss": 2.3478, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.4186046511627907, | |
| "grad_norm": 9.828341484069824, | |
| "learning_rate": 1.3837209302325583e-05, | |
| "loss": 2.2452, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.43023255813953487, | |
| "grad_norm": 8.04952335357666, | |
| "learning_rate": 1.422480620155039e-05, | |
| "loss": 2.1285, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.4418604651162791, | |
| "grad_norm": 8.767853736877441, | |
| "learning_rate": 1.4612403100775195e-05, | |
| "loss": 1.9757, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.45348837209302323, | |
| "grad_norm": 8.500136375427246, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 1.8691, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.46511627906976744, | |
| "grad_norm": 8.077757835388184, | |
| "learning_rate": 1.5387596899224807e-05, | |
| "loss": 1.9578, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.47674418604651164, | |
| "grad_norm": 7.547516345977783, | |
| "learning_rate": 1.5775193798449613e-05, | |
| "loss": 1.869, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.4883720930232558, | |
| "grad_norm": 6.425692558288574, | |
| "learning_rate": 1.616279069767442e-05, | |
| "loss": 1.5731, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 7.883898735046387, | |
| "learning_rate": 1.6550387596899227e-05, | |
| "loss": 1.7579, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.5116279069767442, | |
| "grad_norm": 9.056062698364258, | |
| "learning_rate": 1.6937984496124032e-05, | |
| "loss": 1.5161, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.5232558139534884, | |
| "grad_norm": 7.17751407623291, | |
| "learning_rate": 1.7325581395348837e-05, | |
| "loss": 1.5369, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.5348837209302325, | |
| "grad_norm": 6.8209333419799805, | |
| "learning_rate": 1.7713178294573646e-05, | |
| "loss": 1.2922, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.5465116279069767, | |
| "grad_norm": 8.709400177001953, | |
| "learning_rate": 1.810077519379845e-05, | |
| "loss": 1.1708, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.5581395348837209, | |
| "grad_norm": 9.301350593566895, | |
| "learning_rate": 1.8488372093023256e-05, | |
| "loss": 1.2179, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.5697674418604651, | |
| "grad_norm": 6.705427646636963, | |
| "learning_rate": 1.8875968992248065e-05, | |
| "loss": 1.1499, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.5813953488372093, | |
| "grad_norm": 7.037912368774414, | |
| "learning_rate": 1.926356589147287e-05, | |
| "loss": 1.2365, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.5930232558139535, | |
| "grad_norm": 7.428440093994141, | |
| "learning_rate": 1.9651162790697676e-05, | |
| "loss": 1.0535, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.6046511627906976, | |
| "grad_norm": 8.269710540771484, | |
| "learning_rate": 1.999569336778639e-05, | |
| "loss": 0.9278, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.6162790697674418, | |
| "grad_norm": 5.8355255126953125, | |
| "learning_rate": 1.9952627045650302e-05, | |
| "loss": 1.0103, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.627906976744186, | |
| "grad_norm": 7.757998943328857, | |
| "learning_rate": 1.9909560723514216e-05, | |
| "loss": 0.9842, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.6395348837209303, | |
| "grad_norm": 6.82904577255249, | |
| "learning_rate": 1.9866494401378126e-05, | |
| "loss": 0.7977, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.6511627906976745, | |
| "grad_norm": 6.6139936447143555, | |
| "learning_rate": 1.9823428079242037e-05, | |
| "loss": 1.0064, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.6627906976744186, | |
| "grad_norm": 4.732098579406738, | |
| "learning_rate": 1.9780361757105947e-05, | |
| "loss": 0.6664, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.6744186046511628, | |
| "grad_norm": 6.268205165863037, | |
| "learning_rate": 1.9737295434969854e-05, | |
| "loss": 0.8618, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.686046511627907, | |
| "grad_norm": 6.771559238433838, | |
| "learning_rate": 1.9694229112833765e-05, | |
| "loss": 0.765, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.6976744186046512, | |
| "grad_norm": 5.093957901000977, | |
| "learning_rate": 1.9651162790697676e-05, | |
| "loss": 0.7082, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.7093023255813954, | |
| "grad_norm": 7.572436332702637, | |
| "learning_rate": 1.9608096468561586e-05, | |
| "loss": 0.7704, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.7209302325581395, | |
| "grad_norm": 7.583284378051758, | |
| "learning_rate": 1.9565030146425497e-05, | |
| "loss": 0.6531, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.7325581395348837, | |
| "grad_norm": 4.94130277633667, | |
| "learning_rate": 1.9521963824289407e-05, | |
| "loss": 0.6521, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.7441860465116279, | |
| "grad_norm": 7.676270961761475, | |
| "learning_rate": 1.9478897502153318e-05, | |
| "loss": 0.5925, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.7558139534883721, | |
| "grad_norm": 4.58935022354126, | |
| "learning_rate": 1.9435831180017228e-05, | |
| "loss": 0.649, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.7674418604651163, | |
| "grad_norm": 5.692972183227539, | |
| "learning_rate": 1.939276485788114e-05, | |
| "loss": 0.6009, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.7790697674418605, | |
| "grad_norm": 7.2515997886657715, | |
| "learning_rate": 1.934969853574505e-05, | |
| "loss": 0.6493, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.7906976744186046, | |
| "grad_norm": 8.415182113647461, | |
| "learning_rate": 1.930663221360896e-05, | |
| "loss": 0.6981, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.8023255813953488, | |
| "grad_norm": 6.320438385009766, | |
| "learning_rate": 1.926356589147287e-05, | |
| "loss": 0.6512, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.813953488372093, | |
| "grad_norm": 5.826692581176758, | |
| "learning_rate": 1.922049956933678e-05, | |
| "loss": 0.5459, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.8255813953488372, | |
| "grad_norm": 7.534470081329346, | |
| "learning_rate": 1.917743324720069e-05, | |
| "loss": 0.6567, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.8372093023255814, | |
| "grad_norm": 5.1702165603637695, | |
| "learning_rate": 1.9134366925064602e-05, | |
| "loss": 0.4341, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.8488372093023255, | |
| "grad_norm": 4.03882360458374, | |
| "learning_rate": 1.9091300602928512e-05, | |
| "loss": 0.5627, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.8604651162790697, | |
| "grad_norm": 4.468400478363037, | |
| "learning_rate": 1.9048234280792423e-05, | |
| "loss": 0.5918, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.872093023255814, | |
| "grad_norm": 3.214733362197876, | |
| "learning_rate": 1.9005167958656333e-05, | |
| "loss": 0.4649, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.8837209302325582, | |
| "grad_norm": 5.7742767333984375, | |
| "learning_rate": 1.8962101636520244e-05, | |
| "loss": 0.6414, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.8953488372093024, | |
| "grad_norm": 6.380870342254639, | |
| "learning_rate": 1.8919035314384154e-05, | |
| "loss": 0.4913, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.9069767441860465, | |
| "grad_norm": 6.435304641723633, | |
| "learning_rate": 1.8875968992248065e-05, | |
| "loss": 0.5424, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.9186046511627907, | |
| "grad_norm": 2.9689719676971436, | |
| "learning_rate": 1.8832902670111976e-05, | |
| "loss": 0.5135, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.9302325581395349, | |
| "grad_norm": 6.150023460388184, | |
| "learning_rate": 1.8789836347975886e-05, | |
| "loss": 0.4817, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.9418604651162791, | |
| "grad_norm": 5.568307876586914, | |
| "learning_rate": 1.8746770025839797e-05, | |
| "loss": 0.5294, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.9534883720930233, | |
| "grad_norm": 5.479521751403809, | |
| "learning_rate": 1.8703703703703707e-05, | |
| "loss": 0.5302, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.9651162790697675, | |
| "grad_norm": 5.573367595672607, | |
| "learning_rate": 1.8660637381567618e-05, | |
| "loss": 0.4978, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.9767441860465116, | |
| "grad_norm": 6.795954704284668, | |
| "learning_rate": 1.8617571059431525e-05, | |
| "loss": 0.5098, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.9883720930232558, | |
| "grad_norm": 6.2184295654296875, | |
| "learning_rate": 1.8574504737295435e-05, | |
| "loss": 0.374, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 5.73797607421875, | |
| "learning_rate": 1.8531438415159346e-05, | |
| "loss": 0.4492, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 1.0116279069767442, | |
| "grad_norm": 6.402333736419678, | |
| "learning_rate": 1.8488372093023256e-05, | |
| "loss": 0.3968, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 1.0232558139534884, | |
| "grad_norm": 2.261317014694214, | |
| "learning_rate": 1.8445305770887167e-05, | |
| "loss": 0.4758, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 1.0348837209302326, | |
| "grad_norm": 5.35994815826416, | |
| "learning_rate": 1.8402239448751077e-05, | |
| "loss": 0.4221, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 1.0465116279069768, | |
| "grad_norm": 7.571375846862793, | |
| "learning_rate": 1.8359173126614988e-05, | |
| "loss": 0.4654, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 1.058139534883721, | |
| "grad_norm": 5.946591854095459, | |
| "learning_rate": 1.83161068044789e-05, | |
| "loss": 0.4499, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 1.069767441860465, | |
| "grad_norm": 9.613181114196777, | |
| "learning_rate": 1.827304048234281e-05, | |
| "loss": 0.3823, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 1.0813953488372092, | |
| "grad_norm": 3.0806286334991455, | |
| "learning_rate": 1.822997416020672e-05, | |
| "loss": 0.3525, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 1.0930232558139534, | |
| "grad_norm": 7.319512367248535, | |
| "learning_rate": 1.818690783807063e-05, | |
| "loss": 0.389, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 1.1046511627906976, | |
| "grad_norm": 7.906378746032715, | |
| "learning_rate": 1.814384151593454e-05, | |
| "loss": 0.3431, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 1.1162790697674418, | |
| "grad_norm": 8.479047775268555, | |
| "learning_rate": 1.810077519379845e-05, | |
| "loss": 0.2573, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 1.127906976744186, | |
| "grad_norm": 5.967707633972168, | |
| "learning_rate": 1.805770887166236e-05, | |
| "loss": 0.4355, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 1.1395348837209303, | |
| "grad_norm": 5.773985385894775, | |
| "learning_rate": 1.8014642549526272e-05, | |
| "loss": 0.3693, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 1.1511627906976745, | |
| "grad_norm": 4.215901851654053, | |
| "learning_rate": 1.7971576227390183e-05, | |
| "loss": 0.4744, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.1627906976744187, | |
| "grad_norm": 4.95302677154541, | |
| "learning_rate": 1.7928509905254093e-05, | |
| "loss": 0.4826, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.1627906976744187, | |
| "eval_accuracy": 0.9575581395348837, | |
| "eval_loss": 0.19445417821407318, | |
| "eval_precision": 0.9557833419213569, | |
| "eval_recall": 0.954954954954955, | |
| "eval_runtime": 82.8887, | |
| "eval_samples_per_second": 20.751, | |
| "eval_steps_per_second": 2.594, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.1744186046511629, | |
| "grad_norm": 4.786319732666016, | |
| "learning_rate": 1.7885443583118004e-05, | |
| "loss": 0.3214, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.1860465116279069, | |
| "grad_norm": 2.6006505489349365, | |
| "learning_rate": 1.7842377260981914e-05, | |
| "loss": 0.3656, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.197674418604651, | |
| "grad_norm": 1.0553005933761597, | |
| "learning_rate": 1.7799310938845825e-05, | |
| "loss": 0.2502, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.2093023255813953, | |
| "grad_norm": 4.529180526733398, | |
| "learning_rate": 1.7756244616709735e-05, | |
| "loss": 0.4485, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.2209302325581395, | |
| "grad_norm": 5.82290506362915, | |
| "learning_rate": 1.7713178294573646e-05, | |
| "loss": 0.4492, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.2325581395348837, | |
| "grad_norm": 5.273308277130127, | |
| "learning_rate": 1.7670111972437556e-05, | |
| "loss": 0.453, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.244186046511628, | |
| "grad_norm": 8.547164916992188, | |
| "learning_rate": 1.7627045650301467e-05, | |
| "loss": 0.3011, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.255813953488372, | |
| "grad_norm": 5.207535266876221, | |
| "learning_rate": 1.7583979328165377e-05, | |
| "loss": 0.4841, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.2674418604651163, | |
| "grad_norm": 7.522132396697998, | |
| "learning_rate": 1.7540913006029288e-05, | |
| "loss": 0.3502, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.2790697674418605, | |
| "grad_norm": 5.226884365081787, | |
| "learning_rate": 1.7497846683893195e-05, | |
| "loss": 0.3684, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.2906976744186047, | |
| "grad_norm": 1.8360196352005005, | |
| "learning_rate": 1.7454780361757106e-05, | |
| "loss": 0.3172, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.302325581395349, | |
| "grad_norm": 5.412622451782227, | |
| "learning_rate": 1.7411714039621016e-05, | |
| "loss": 0.3139, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.3139534883720931, | |
| "grad_norm": 8.000215530395508, | |
| "learning_rate": 1.7368647717484927e-05, | |
| "loss": 0.333, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.3255813953488373, | |
| "grad_norm": 0.7105430364608765, | |
| "learning_rate": 1.7325581395348837e-05, | |
| "loss": 0.3729, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.3372093023255813, | |
| "grad_norm": 5.012469291687012, | |
| "learning_rate": 1.7282515073212748e-05, | |
| "loss": 0.4945, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.3488372093023255, | |
| "grad_norm": 5.788574695587158, | |
| "learning_rate": 1.7239448751076658e-05, | |
| "loss": 0.4288, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.3604651162790697, | |
| "grad_norm": 6.5563812255859375, | |
| "learning_rate": 1.719638242894057e-05, | |
| "loss": 0.5288, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.372093023255814, | |
| "grad_norm": 4.14633321762085, | |
| "learning_rate": 1.715331610680448e-05, | |
| "loss": 0.4037, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.3837209302325582, | |
| "grad_norm": 4.901910305023193, | |
| "learning_rate": 1.711024978466839e-05, | |
| "loss": 0.3318, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.3953488372093024, | |
| "grad_norm": 4.042469024658203, | |
| "learning_rate": 1.70671834625323e-05, | |
| "loss": 0.308, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.4069767441860466, | |
| "grad_norm": 4.500431060791016, | |
| "learning_rate": 1.702411714039621e-05, | |
| "loss": 0.3334, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.4186046511627908, | |
| "grad_norm": 5.609718322753906, | |
| "learning_rate": 1.6981050818260125e-05, | |
| "loss": 0.3684, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.4302325581395348, | |
| "grad_norm": 5.352325916290283, | |
| "learning_rate": 1.6937984496124032e-05, | |
| "loss": 0.5298, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.441860465116279, | |
| "grad_norm": 4.740088939666748, | |
| "learning_rate": 1.6894918173987942e-05, | |
| "loss": 0.326, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.4534883720930232, | |
| "grad_norm": 3.104138135910034, | |
| "learning_rate": 1.6851851851851853e-05, | |
| "loss": 0.2879, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.4651162790697674, | |
| "grad_norm": 5.4181928634643555, | |
| "learning_rate": 1.6808785529715763e-05, | |
| "loss": 0.2861, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.4767441860465116, | |
| "grad_norm": 7.169526100158691, | |
| "learning_rate": 1.6765719207579674e-05, | |
| "loss": 0.4213, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.4883720930232558, | |
| "grad_norm": 4.903182506561279, | |
| "learning_rate": 1.6722652885443584e-05, | |
| "loss": 0.4361, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 4.7495622634887695, | |
| "learning_rate": 1.6679586563307495e-05, | |
| "loss": 0.4048, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.5116279069767442, | |
| "grad_norm": 5.136496543884277, | |
| "learning_rate": 1.6636520241171406e-05, | |
| "loss": 0.4402, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.5232558139534884, | |
| "grad_norm": 7.539604663848877, | |
| "learning_rate": 1.6593453919035316e-05, | |
| "loss": 0.3299, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.5348837209302326, | |
| "grad_norm": 6.178093910217285, | |
| "learning_rate": 1.6550387596899227e-05, | |
| "loss": 0.494, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.5465116279069768, | |
| "grad_norm": 7.080124855041504, | |
| "learning_rate": 1.6507321274763137e-05, | |
| "loss": 0.4094, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.558139534883721, | |
| "grad_norm": 7.745635509490967, | |
| "learning_rate": 1.6464254952627048e-05, | |
| "loss": 0.4177, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.5697674418604652, | |
| "grad_norm": 3.9280893802642822, | |
| "learning_rate": 1.6421188630490958e-05, | |
| "loss": 0.3839, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.5813953488372094, | |
| "grad_norm": 5.974954605102539, | |
| "learning_rate": 1.6378122308354865e-05, | |
| "loss": 0.4142, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.5930232558139537, | |
| "grad_norm": 2.93672776222229, | |
| "learning_rate": 1.6335055986218776e-05, | |
| "loss": 0.3802, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.6046511627906976, | |
| "grad_norm": 3.1751911640167236, | |
| "learning_rate": 1.6291989664082686e-05, | |
| "loss": 0.3101, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.6162790697674418, | |
| "grad_norm": 6.783326625823975, | |
| "learning_rate": 1.62489233419466e-05, | |
| "loss": 0.3394, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.627906976744186, | |
| "grad_norm": 7.884994983673096, | |
| "learning_rate": 1.620585701981051e-05, | |
| "loss": 0.3857, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.6395348837209303, | |
| "grad_norm": 0.44594404101371765, | |
| "learning_rate": 1.616279069767442e-05, | |
| "loss": 0.2612, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.6511627906976745, | |
| "grad_norm": 7.8830766677856445, | |
| "learning_rate": 1.6119724375538332e-05, | |
| "loss": 0.2814, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.6627906976744184, | |
| "grad_norm": 6.6618452072143555, | |
| "learning_rate": 1.6076658053402242e-05, | |
| "loss": 0.4027, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.6744186046511627, | |
| "grad_norm": 6.068259239196777, | |
| "learning_rate": 1.6033591731266153e-05, | |
| "loss": 0.291, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.6860465116279069, | |
| "grad_norm": 4.270251750946045, | |
| "learning_rate": 1.5990525409130063e-05, | |
| "loss": 0.3398, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.697674418604651, | |
| "grad_norm": 5.579948902130127, | |
| "learning_rate": 1.5947459086993974e-05, | |
| "loss": 0.4223, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.7093023255813953, | |
| "grad_norm": 8.341035842895508, | |
| "learning_rate": 1.5904392764857884e-05, | |
| "loss": 0.3797, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.7209302325581395, | |
| "grad_norm": 5.616793155670166, | |
| "learning_rate": 1.5861326442721795e-05, | |
| "loss": 0.3325, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.7325581395348837, | |
| "grad_norm": 4.591681480407715, | |
| "learning_rate": 1.5818260120585702e-05, | |
| "loss": 0.2752, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.744186046511628, | |
| "grad_norm": 3.9187428951263428, | |
| "learning_rate": 1.5775193798449613e-05, | |
| "loss": 0.3654, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.755813953488372, | |
| "grad_norm": 3.269841194152832, | |
| "learning_rate": 1.5732127476313523e-05, | |
| "loss": 0.3589, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.7674418604651163, | |
| "grad_norm": 7.0987091064453125, | |
| "learning_rate": 1.5689061154177434e-05, | |
| "loss": 0.4337, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.7790697674418605, | |
| "grad_norm": 6.354150295257568, | |
| "learning_rate": 1.5645994832041344e-05, | |
| "loss": 0.3615, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.7906976744186047, | |
| "grad_norm": 4.65658712387085, | |
| "learning_rate": 1.5602928509905255e-05, | |
| "loss": 0.2358, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.802325581395349, | |
| "grad_norm": 6.8650407791137695, | |
| "learning_rate": 1.5559862187769165e-05, | |
| "loss": 0.3049, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.8139534883720931, | |
| "grad_norm": 7.136849880218506, | |
| "learning_rate": 1.5516795865633076e-05, | |
| "loss": 0.3295, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.8255813953488373, | |
| "grad_norm": 5.524831771850586, | |
| "learning_rate": 1.5473729543496986e-05, | |
| "loss": 0.3014, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.8372093023255816, | |
| "grad_norm": 3.4332170486450195, | |
| "learning_rate": 1.5430663221360897e-05, | |
| "loss": 0.3792, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.8488372093023255, | |
| "grad_norm": 3.5294277667999268, | |
| "learning_rate": 1.5387596899224807e-05, | |
| "loss": 0.3246, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.8604651162790697, | |
| "grad_norm": 4.115309715270996, | |
| "learning_rate": 1.5344530577088718e-05, | |
| "loss": 0.3287, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.872093023255814, | |
| "grad_norm": 2.1745433807373047, | |
| "learning_rate": 1.530146425495263e-05, | |
| "loss": 0.3469, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.8837209302325582, | |
| "grad_norm": 0.8755269646644592, | |
| "learning_rate": 1.5258397932816537e-05, | |
| "loss": 0.3019, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.8953488372093024, | |
| "grad_norm": 4.384407043457031, | |
| "learning_rate": 1.5215331610680448e-05, | |
| "loss": 0.3175, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.9069767441860463, | |
| "grad_norm": 6.729555130004883, | |
| "learning_rate": 1.5172265288544358e-05, | |
| "loss": 0.4233, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.9186046511627906, | |
| "grad_norm": 3.5528595447540283, | |
| "learning_rate": 1.5129198966408269e-05, | |
| "loss": 0.4227, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.9302325581395348, | |
| "grad_norm": 9.704643249511719, | |
| "learning_rate": 1.508613264427218e-05, | |
| "loss": 0.2756, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.941860465116279, | |
| "grad_norm": 8.284793853759766, | |
| "learning_rate": 1.504306632213609e-05, | |
| "loss": 0.4224, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.9534883720930232, | |
| "grad_norm": 3.9722352027893066, | |
| "learning_rate": 1.5000000000000002e-05, | |
| "loss": 0.3271, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.9651162790697674, | |
| "grad_norm": 3.740689754486084, | |
| "learning_rate": 1.4956933677863913e-05, | |
| "loss": 0.3578, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.9767441860465116, | |
| "grad_norm": 2.0497655868530273, | |
| "learning_rate": 1.4913867355727823e-05, | |
| "loss": 0.2528, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.9883720930232558, | |
| "grad_norm": 9.135187149047852, | |
| "learning_rate": 1.4870801033591734e-05, | |
| "loss": 0.3749, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 4.393891334533691, | |
| "learning_rate": 1.4827734711455644e-05, | |
| "loss": 0.3657, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 2.011627906976744, | |
| "grad_norm": 3.3051867485046387, | |
| "learning_rate": 1.4784668389319555e-05, | |
| "loss": 0.2316, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 2.0232558139534884, | |
| "grad_norm": 5.197268009185791, | |
| "learning_rate": 1.4741602067183465e-05, | |
| "loss": 0.2754, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 2.0348837209302326, | |
| "grad_norm": 0.7799841165542603, | |
| "learning_rate": 1.4698535745047372e-05, | |
| "loss": 0.2024, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 2.046511627906977, | |
| "grad_norm": 7.791285514831543, | |
| "learning_rate": 1.4655469422911283e-05, | |
| "loss": 0.1916, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 2.058139534883721, | |
| "grad_norm": 4.7339558601379395, | |
| "learning_rate": 1.4612403100775195e-05, | |
| "loss": 0.3417, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 2.0697674418604652, | |
| "grad_norm": 5.472916603088379, | |
| "learning_rate": 1.4569336778639106e-05, | |
| "loss": 0.2195, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 2.0813953488372094, | |
| "grad_norm": 5.680606365203857, | |
| "learning_rate": 1.4526270456503016e-05, | |
| "loss": 0.3598, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 2.0930232558139537, | |
| "grad_norm": 5.776130199432373, | |
| "learning_rate": 1.4483204134366927e-05, | |
| "loss": 0.3561, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 2.104651162790698, | |
| "grad_norm": 9.959355354309082, | |
| "learning_rate": 1.4440137812230837e-05, | |
| "loss": 0.291, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 2.116279069767442, | |
| "grad_norm": 0.20813125371932983, | |
| "learning_rate": 1.4397071490094748e-05, | |
| "loss": 0.2624, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 2.1279069767441863, | |
| "grad_norm": 2.536965847015381, | |
| "learning_rate": 1.4354005167958658e-05, | |
| "loss": 0.261, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 2.13953488372093, | |
| "grad_norm": 5.530265808105469, | |
| "learning_rate": 1.4310938845822569e-05, | |
| "loss": 0.3927, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 2.1511627906976742, | |
| "grad_norm": 9.419245719909668, | |
| "learning_rate": 1.426787252368648e-05, | |
| "loss": 0.4612, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 2.1627906976744184, | |
| "grad_norm": 6.30083703994751, | |
| "learning_rate": 1.422480620155039e-05, | |
| "loss": 0.4033, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 2.1744186046511627, | |
| "grad_norm": 5.586449146270752, | |
| "learning_rate": 1.41817398794143e-05, | |
| "loss": 0.45, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 2.186046511627907, | |
| "grad_norm": 3.4549663066864014, | |
| "learning_rate": 1.413867355727821e-05, | |
| "loss": 0.3375, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 2.197674418604651, | |
| "grad_norm": 3.5873939990997314, | |
| "learning_rate": 1.409560723514212e-05, | |
| "loss": 0.2418, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 2.2093023255813953, | |
| "grad_norm": 2.905010223388672, | |
| "learning_rate": 1.405254091300603e-05, | |
| "loss": 0.228, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 2.2209302325581395, | |
| "grad_norm": 7.639858722686768, | |
| "learning_rate": 1.400947459086994e-05, | |
| "loss": 0.2286, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 2.2325581395348837, | |
| "grad_norm": 4.643288612365723, | |
| "learning_rate": 1.3966408268733851e-05, | |
| "loss": 0.2854, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 2.244186046511628, | |
| "grad_norm": 3.442018985748291, | |
| "learning_rate": 1.3923341946597762e-05, | |
| "loss": 0.314, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 2.255813953488372, | |
| "grad_norm": 4.823495864868164, | |
| "learning_rate": 1.3880275624461672e-05, | |
| "loss": 0.2837, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 2.2674418604651163, | |
| "grad_norm": 8.560246467590332, | |
| "learning_rate": 1.3837209302325583e-05, | |
| "loss": 0.4313, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 2.2790697674418605, | |
| "grad_norm": 6.159987926483154, | |
| "learning_rate": 1.3794142980189493e-05, | |
| "loss": 0.3137, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 2.2906976744186047, | |
| "grad_norm": 4.31523323059082, | |
| "learning_rate": 1.3751076658053404e-05, | |
| "loss": 0.2022, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 2.302325581395349, | |
| "grad_norm": 2.9448819160461426, | |
| "learning_rate": 1.3708010335917314e-05, | |
| "loss": 0.2981, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 2.313953488372093, | |
| "grad_norm": 6.914911270141602, | |
| "learning_rate": 1.3664944013781225e-05, | |
| "loss": 0.225, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 2.3255813953488373, | |
| "grad_norm": 5.1811017990112305, | |
| "learning_rate": 1.3621877691645135e-05, | |
| "loss": 0.2588, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 2.3255813953488373, | |
| "eval_accuracy": 0.9744186046511628, | |
| "eval_loss": 0.10524194687604904, | |
| "eval_precision": 0.9734332599197442, | |
| "eval_recall": 0.9720720720720719, | |
| "eval_runtime": 55.6979, | |
| "eval_samples_per_second": 30.881, | |
| "eval_steps_per_second": 3.86, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 2.3372093023255816, | |
| "grad_norm": 0.10492008924484253, | |
| "learning_rate": 1.3578811369509044e-05, | |
| "loss": 0.2415, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 2.3488372093023258, | |
| "grad_norm": 4.322141647338867, | |
| "learning_rate": 1.3535745047372955e-05, | |
| "loss": 0.3838, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 2.3604651162790695, | |
| "grad_norm": 4.23949670791626, | |
| "learning_rate": 1.3492678725236865e-05, | |
| "loss": 0.3642, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 2.3720930232558137, | |
| "grad_norm": 0.4247274696826935, | |
| "learning_rate": 1.3449612403100776e-05, | |
| "loss": 0.2703, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 2.383720930232558, | |
| "grad_norm": 3.733469009399414, | |
| "learning_rate": 1.3406546080964686e-05, | |
| "loss": 0.2561, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 2.395348837209302, | |
| "grad_norm": 6.659445285797119, | |
| "learning_rate": 1.3363479758828597e-05, | |
| "loss": 0.5111, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 2.4069767441860463, | |
| "grad_norm": 8.265155792236328, | |
| "learning_rate": 1.3320413436692507e-05, | |
| "loss": 0.3314, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 2.4186046511627906, | |
| "grad_norm": 7.012650966644287, | |
| "learning_rate": 1.3277347114556418e-05, | |
| "loss": 0.2415, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 2.4302325581395348, | |
| "grad_norm": 6.1161885261535645, | |
| "learning_rate": 1.3234280792420329e-05, | |
| "loss": 0.3364, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 2.441860465116279, | |
| "grad_norm": 0.41343072056770325, | |
| "learning_rate": 1.3191214470284239e-05, | |
| "loss": 0.294, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 2.453488372093023, | |
| "grad_norm": 6.758556842803955, | |
| "learning_rate": 1.314814814814815e-05, | |
| "loss": 0.2794, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 2.4651162790697674, | |
| "grad_norm": 11.507908821105957, | |
| "learning_rate": 1.310508182601206e-05, | |
| "loss": 0.3635, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 2.4767441860465116, | |
| "grad_norm": 8.224679946899414, | |
| "learning_rate": 1.306201550387597e-05, | |
| "loss": 0.3312, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 2.488372093023256, | |
| "grad_norm": 7.767526149749756, | |
| "learning_rate": 1.301894918173988e-05, | |
| "loss": 0.1664, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 2.9291210174560547, | |
| "learning_rate": 1.297588285960379e-05, | |
| "loss": 0.3163, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 2.511627906976744, | |
| "grad_norm": 4.891172885894775, | |
| "learning_rate": 1.29328165374677e-05, | |
| "loss": 0.2852, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 2.5232558139534884, | |
| "grad_norm": 6.734064102172852, | |
| "learning_rate": 1.2889750215331611e-05, | |
| "loss": 0.2764, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 2.5348837209302326, | |
| "grad_norm": 4.138737201690674, | |
| "learning_rate": 1.2846683893195522e-05, | |
| "loss": 0.3058, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 2.546511627906977, | |
| "grad_norm": 8.400827407836914, | |
| "learning_rate": 1.2803617571059432e-05, | |
| "loss": 0.3334, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 2.558139534883721, | |
| "grad_norm": 4.968383312225342, | |
| "learning_rate": 1.2760551248923343e-05, | |
| "loss": 0.3415, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 2.5697674418604652, | |
| "grad_norm": 0.37472769618034363, | |
| "learning_rate": 1.2717484926787253e-05, | |
| "loss": 0.2979, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 2.5813953488372094, | |
| "grad_norm": 3.250365734100342, | |
| "learning_rate": 1.2674418604651164e-05, | |
| "loss": 0.2682, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 2.5930232558139537, | |
| "grad_norm": 5.750937461853027, | |
| "learning_rate": 1.2631352282515074e-05, | |
| "loss": 0.3554, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 2.604651162790698, | |
| "grad_norm": 6.268227577209473, | |
| "learning_rate": 1.2588285960378985e-05, | |
| "loss": 0.2666, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 2.616279069767442, | |
| "grad_norm": 3.9889657497406006, | |
| "learning_rate": 1.2545219638242895e-05, | |
| "loss": 0.1963, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 2.6279069767441863, | |
| "grad_norm": 7.115270614624023, | |
| "learning_rate": 1.2502153316106806e-05, | |
| "loss": 0.3314, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 2.6395348837209305, | |
| "grad_norm": 2.193861961364746, | |
| "learning_rate": 1.2459086993970715e-05, | |
| "loss": 0.2563, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 2.6511627906976747, | |
| "grad_norm": 5.2253947257995605, | |
| "learning_rate": 1.2416020671834625e-05, | |
| "loss": 0.2023, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 2.6627906976744184, | |
| "grad_norm": 7.278836250305176, | |
| "learning_rate": 1.2372954349698536e-05, | |
| "loss": 0.2424, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 2.6744186046511627, | |
| "grad_norm": 5.243718147277832, | |
| "learning_rate": 1.2329888027562446e-05, | |
| "loss": 0.3941, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 2.686046511627907, | |
| "grad_norm": 2.4259538650512695, | |
| "learning_rate": 1.2286821705426357e-05, | |
| "loss": 0.4413, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 2.697674418604651, | |
| "grad_norm": 5.592273235321045, | |
| "learning_rate": 1.2243755383290267e-05, | |
| "loss": 0.2828, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 2.7093023255813953, | |
| "grad_norm": 4.666529655456543, | |
| "learning_rate": 1.2200689061154178e-05, | |
| "loss": 0.31, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 2.7209302325581395, | |
| "grad_norm": 3.881617546081543, | |
| "learning_rate": 1.2157622739018088e-05, | |
| "loss": 0.2825, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 2.7325581395348837, | |
| "grad_norm": 6.821516036987305, | |
| "learning_rate": 1.2114556416881999e-05, | |
| "loss": 0.4878, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 2.744186046511628, | |
| "grad_norm": 2.946510076522827, | |
| "learning_rate": 1.2071490094745911e-05, | |
| "loss": 0.3473, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 2.755813953488372, | |
| "grad_norm": 5.623599529266357, | |
| "learning_rate": 1.2028423772609822e-05, | |
| "loss": 0.2614, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 2.7674418604651163, | |
| "grad_norm": 2.3572816848754883, | |
| "learning_rate": 1.1985357450473732e-05, | |
| "loss": 0.1801, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 2.7790697674418605, | |
| "grad_norm": 4.094923973083496, | |
| "learning_rate": 1.1942291128337643e-05, | |
| "loss": 0.2492, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 2.7906976744186047, | |
| "grad_norm": 2.7377874851226807, | |
| "learning_rate": 1.189922480620155e-05, | |
| "loss": 0.2706, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 2.802325581395349, | |
| "grad_norm": 5.991059303283691, | |
| "learning_rate": 1.185615848406546e-05, | |
| "loss": 0.3915, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 2.813953488372093, | |
| "grad_norm": 6.187289714813232, | |
| "learning_rate": 1.181309216192937e-05, | |
| "loss": 0.3056, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 2.8255813953488373, | |
| "grad_norm": 2.1323797702789307, | |
| "learning_rate": 1.1770025839793281e-05, | |
| "loss": 0.1805, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 2.8372093023255816, | |
| "grad_norm": 5.122096538543701, | |
| "learning_rate": 1.1726959517657192e-05, | |
| "loss": 0.2774, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 2.8488372093023253, | |
| "grad_norm": 11.290266036987305, | |
| "learning_rate": 1.1683893195521104e-05, | |
| "loss": 0.2419, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 2.8604651162790695, | |
| "grad_norm": 7.9658660888671875, | |
| "learning_rate": 1.1640826873385015e-05, | |
| "loss": 0.3104, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 2.8720930232558137, | |
| "grad_norm": 7.313680171966553, | |
| "learning_rate": 1.1597760551248925e-05, | |
| "loss": 0.1897, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 2.883720930232558, | |
| "grad_norm": 2.3773865699768066, | |
| "learning_rate": 1.1554694229112836e-05, | |
| "loss": 0.2036, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 2.895348837209302, | |
| "grad_norm": 6.803826332092285, | |
| "learning_rate": 1.1511627906976746e-05, | |
| "loss": 0.2791, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 2.9069767441860463, | |
| "grad_norm": 8.565348625183105, | |
| "learning_rate": 1.1468561584840657e-05, | |
| "loss": 0.2677, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.9186046511627906, | |
| "grad_norm": 2.4050967693328857, | |
| "learning_rate": 1.1425495262704567e-05, | |
| "loss": 0.3743, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 2.9302325581395348, | |
| "grad_norm": 4.761075496673584, | |
| "learning_rate": 1.1382428940568478e-05, | |
| "loss": 0.2059, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 2.941860465116279, | |
| "grad_norm": 7.1044487953186035, | |
| "learning_rate": 1.1339362618432387e-05, | |
| "loss": 0.3741, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 2.953488372093023, | |
| "grad_norm": 4.308391094207764, | |
| "learning_rate": 1.1296296296296297e-05, | |
| "loss": 0.2231, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 2.9651162790697674, | |
| "grad_norm": 6.208861351013184, | |
| "learning_rate": 1.1253229974160208e-05, | |
| "loss": 0.3116, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.9767441860465116, | |
| "grad_norm": 0.26213783025741577, | |
| "learning_rate": 1.1210163652024118e-05, | |
| "loss": 0.3295, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 2.988372093023256, | |
| "grad_norm": 7.947643756866455, | |
| "learning_rate": 1.1167097329888029e-05, | |
| "loss": 0.2413, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 5.538887977600098, | |
| "learning_rate": 1.112403100775194e-05, | |
| "loss": 0.1269, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 3.011627906976744, | |
| "grad_norm": 5.078567028045654, | |
| "learning_rate": 1.108096468561585e-05, | |
| "loss": 0.4048, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 3.0232558139534884, | |
| "grad_norm": 5.892114162445068, | |
| "learning_rate": 1.103789836347976e-05, | |
| "loss": 0.192, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 3.0348837209302326, | |
| "grad_norm": 5.286245822906494, | |
| "learning_rate": 1.099483204134367e-05, | |
| "loss": 0.3413, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 3.046511627906977, | |
| "grad_norm": 4.882418155670166, | |
| "learning_rate": 1.0951765719207581e-05, | |
| "loss": 0.3462, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 3.058139534883721, | |
| "grad_norm": 5.178489685058594, | |
| "learning_rate": 1.0908699397071492e-05, | |
| "loss": 0.3925, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 3.0697674418604652, | |
| "grad_norm": 0.4772948622703552, | |
| "learning_rate": 1.0865633074935402e-05, | |
| "loss": 0.3728, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 3.0813953488372094, | |
| "grad_norm": 4.9134979248046875, | |
| "learning_rate": 1.0822566752799313e-05, | |
| "loss": 0.257, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 3.0930232558139537, | |
| "grad_norm": 2.673166275024414, | |
| "learning_rate": 1.0779500430663222e-05, | |
| "loss": 0.2248, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 3.104651162790698, | |
| "grad_norm": 5.884793281555176, | |
| "learning_rate": 1.0736434108527132e-05, | |
| "loss": 0.3653, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 3.116279069767442, | |
| "grad_norm": 6.795538902282715, | |
| "learning_rate": 1.0693367786391043e-05, | |
| "loss": 0.2939, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 3.1279069767441863, | |
| "grad_norm": 4.707054615020752, | |
| "learning_rate": 1.0650301464254953e-05, | |
| "loss": 0.2882, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 3.13953488372093, | |
| "grad_norm": 5.53010368347168, | |
| "learning_rate": 1.0607235142118864e-05, | |
| "loss": 0.2095, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 3.1511627906976742, | |
| "grad_norm": 5.117499828338623, | |
| "learning_rate": 1.0564168819982774e-05, | |
| "loss": 0.2329, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 3.1627906976744184, | |
| "grad_norm": 6.0297722816467285, | |
| "learning_rate": 1.0521102497846685e-05, | |
| "loss": 0.2747, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 3.1744186046511627, | |
| "grad_norm": 3.6632094383239746, | |
| "learning_rate": 1.0478036175710595e-05, | |
| "loss": 0.2839, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 3.186046511627907, | |
| "grad_norm": 9.074336051940918, | |
| "learning_rate": 1.0434969853574506e-05, | |
| "loss": 0.3073, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 3.197674418604651, | |
| "grad_norm": 7.413300037384033, | |
| "learning_rate": 1.0391903531438416e-05, | |
| "loss": 0.2466, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 3.2093023255813953, | |
| "grad_norm": 5.545544624328613, | |
| "learning_rate": 1.0348837209302327e-05, | |
| "loss": 0.2604, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 3.2209302325581395, | |
| "grad_norm": 3.1789004802703857, | |
| "learning_rate": 1.0305770887166237e-05, | |
| "loss": 0.2464, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 3.2325581395348837, | |
| "grad_norm": 7.00383186340332, | |
| "learning_rate": 1.0262704565030148e-05, | |
| "loss": 0.2266, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 3.244186046511628, | |
| "grad_norm": 2.2015435695648193, | |
| "learning_rate": 1.0219638242894057e-05, | |
| "loss": 0.1769, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 3.255813953488372, | |
| "grad_norm": 4.523898124694824, | |
| "learning_rate": 1.0176571920757967e-05, | |
| "loss": 0.2129, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 3.2674418604651163, | |
| "grad_norm": 6.628675937652588, | |
| "learning_rate": 1.0133505598621878e-05, | |
| "loss": 0.439, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 3.2790697674418605, | |
| "grad_norm": 5.438514232635498, | |
| "learning_rate": 1.0090439276485788e-05, | |
| "loss": 0.2919, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 3.2906976744186047, | |
| "grad_norm": 6.8291120529174805, | |
| "learning_rate": 1.0047372954349699e-05, | |
| "loss": 0.2577, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 3.302325581395349, | |
| "grad_norm": 4.305720806121826, | |
| "learning_rate": 1.000430663221361e-05, | |
| "loss": 0.2261, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 3.313953488372093, | |
| "grad_norm": 3.586775064468384, | |
| "learning_rate": 9.96124031007752e-06, | |
| "loss": 0.3861, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 3.3255813953488373, | |
| "grad_norm": 7.542160987854004, | |
| "learning_rate": 9.91817398794143e-06, | |
| "loss": 0.2843, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 3.3372093023255816, | |
| "grad_norm": 5.423547744750977, | |
| "learning_rate": 9.875107665805341e-06, | |
| "loss": 0.4241, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 3.3488372093023258, | |
| "grad_norm": 3.778038501739502, | |
| "learning_rate": 9.832041343669252e-06, | |
| "loss": 0.1172, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 3.3604651162790695, | |
| "grad_norm": 7.731586933135986, | |
| "learning_rate": 9.788975021533162e-06, | |
| "loss": 0.3514, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 3.3720930232558137, | |
| "grad_norm": 9.345348358154297, | |
| "learning_rate": 9.745908699397073e-06, | |
| "loss": 0.35, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 3.383720930232558, | |
| "grad_norm": 5.361115455627441, | |
| "learning_rate": 9.702842377260983e-06, | |
| "loss": 0.2052, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 3.395348837209302, | |
| "grad_norm": 5.338897228240967, | |
| "learning_rate": 9.659776055124894e-06, | |
| "loss": 0.265, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 3.4069767441860463, | |
| "grad_norm": 6.550836086273193, | |
| "learning_rate": 9.616709732988804e-06, | |
| "loss": 0.1503, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 3.4186046511627906, | |
| "grad_norm": 4.383169651031494, | |
| "learning_rate": 9.573643410852715e-06, | |
| "loss": 0.1964, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 3.4302325581395348, | |
| "grad_norm": 0.05135205015540123, | |
| "learning_rate": 9.530577088716625e-06, | |
| "loss": 0.1701, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 3.441860465116279, | |
| "grad_norm": 0.8218990564346313, | |
| "learning_rate": 9.487510766580534e-06, | |
| "loss": 0.2921, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 3.453488372093023, | |
| "grad_norm": 6.9761643409729, | |
| "learning_rate": 9.444444444444445e-06, | |
| "loss": 0.258, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 3.4651162790697674, | |
| "grad_norm": 7.547414779663086, | |
| "learning_rate": 9.401378122308355e-06, | |
| "loss": 0.277, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 3.4767441860465116, | |
| "grad_norm": 3.0600359439849854, | |
| "learning_rate": 9.358311800172266e-06, | |
| "loss": 0.2429, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 3.488372093023256, | |
| "grad_norm": 8.364120483398438, | |
| "learning_rate": 9.315245478036176e-06, | |
| "loss": 0.3581, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 3.488372093023256, | |
| "eval_accuracy": 0.977906976744186, | |
| "eval_loss": 0.07633475214242935, | |
| "eval_precision": 0.9765496878642919, | |
| "eval_recall": 0.9761261261261261, | |
| "eval_runtime": 55.734, | |
| "eval_samples_per_second": 30.861, | |
| "eval_steps_per_second": 3.858, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 3.5, | |
| "grad_norm": 2.64277720451355, | |
| "learning_rate": 9.272179155900087e-06, | |
| "loss": 0.3382, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 3.511627906976744, | |
| "grad_norm": 4.21881628036499, | |
| "learning_rate": 9.229112833763997e-06, | |
| "loss": 0.3063, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 3.5232558139534884, | |
| "grad_norm": 4.8079681396484375, | |
| "learning_rate": 9.186046511627908e-06, | |
| "loss": 0.2596, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 3.5348837209302326, | |
| "grad_norm": 3.439565420150757, | |
| "learning_rate": 9.142980189491818e-06, | |
| "loss": 0.2461, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 3.546511627906977, | |
| "grad_norm": 1.834625005722046, | |
| "learning_rate": 9.099913867355729e-06, | |
| "loss": 0.2432, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 3.558139534883721, | |
| "grad_norm": 6.802725791931152, | |
| "learning_rate": 9.05684754521964e-06, | |
| "loss": 0.3319, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 3.5697674418604652, | |
| "grad_norm": 0.7873058319091797, | |
| "learning_rate": 9.01378122308355e-06, | |
| "loss": 0.2219, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 3.5813953488372094, | |
| "grad_norm": 2.9496591091156006, | |
| "learning_rate": 8.97071490094746e-06, | |
| "loss": 0.2262, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 3.5930232558139537, | |
| "grad_norm": 5.338105201721191, | |
| "learning_rate": 8.92764857881137e-06, | |
| "loss": 0.3029, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 3.604651162790698, | |
| "grad_norm": 9.626699447631836, | |
| "learning_rate": 8.88458225667528e-06, | |
| "loss": 0.223, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 3.616279069767442, | |
| "grad_norm": 3.6059839725494385, | |
| "learning_rate": 8.84151593453919e-06, | |
| "loss": 0.2861, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 3.6279069767441863, | |
| "grad_norm": 19.520587921142578, | |
| "learning_rate": 8.7984496124031e-06, | |
| "loss": 0.4198, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 3.6395348837209305, | |
| "grad_norm": 2.2791006565093994, | |
| "learning_rate": 8.755383290267013e-06, | |
| "loss": 0.1977, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 3.6511627906976747, | |
| "grad_norm": 1.9899883270263672, | |
| "learning_rate": 8.712316968130923e-06, | |
| "loss": 0.3211, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 3.6627906976744184, | |
| "grad_norm": 6.382236003875732, | |
| "learning_rate": 8.669250645994832e-06, | |
| "loss": 0.2108, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 3.6744186046511627, | |
| "grad_norm": 4.365346431732178, | |
| "learning_rate": 8.626184323858743e-06, | |
| "loss": 0.2534, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 3.686046511627907, | |
| "grad_norm": 5.100739002227783, | |
| "learning_rate": 8.583118001722653e-06, | |
| "loss": 0.2337, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 3.697674418604651, | |
| "grad_norm": 9.316176414489746, | |
| "learning_rate": 8.540051679586564e-06, | |
| "loss": 0.2106, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 3.7093023255813953, | |
| "grad_norm": 7.2974700927734375, | |
| "learning_rate": 8.496985357450474e-06, | |
| "loss": 0.3304, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 3.7209302325581395, | |
| "grad_norm": 3.294980049133301, | |
| "learning_rate": 8.453919035314385e-06, | |
| "loss": 0.2843, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 3.7325581395348837, | |
| "grad_norm": 7.945123672485352, | |
| "learning_rate": 8.410852713178295e-06, | |
| "loss": 0.2531, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 3.744186046511628, | |
| "grad_norm": 3.1718406677246094, | |
| "learning_rate": 8.367786391042206e-06, | |
| "loss": 0.226, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 3.755813953488372, | |
| "grad_norm": 6.0147809982299805, | |
| "learning_rate": 8.324720068906117e-06, | |
| "loss": 0.2268, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 3.7674418604651163, | |
| "grad_norm": 3.3491480350494385, | |
| "learning_rate": 8.281653746770027e-06, | |
| "loss": 0.2002, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 3.7790697674418605, | |
| "grad_norm": 3.9779930114746094, | |
| "learning_rate": 8.238587424633938e-06, | |
| "loss": 0.2724, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 3.7906976744186047, | |
| "grad_norm": 7.357614040374756, | |
| "learning_rate": 8.195521102497848e-06, | |
| "loss": 0.2564, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 3.802325581395349, | |
| "grad_norm": 4.795936584472656, | |
| "learning_rate": 8.152454780361759e-06, | |
| "loss": 0.2652, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 3.813953488372093, | |
| "grad_norm": 4.927507400512695, | |
| "learning_rate": 8.109388458225667e-06, | |
| "loss": 0.2766, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 3.8255813953488373, | |
| "grad_norm": 5.914189338684082, | |
| "learning_rate": 8.066322136089578e-06, | |
| "loss": 0.1539, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 3.8372093023255816, | |
| "grad_norm": 4.69040584564209, | |
| "learning_rate": 8.023255813953488e-06, | |
| "loss": 0.3369, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 3.8488372093023253, | |
| "grad_norm": 3.757734775543213, | |
| "learning_rate": 7.980189491817399e-06, | |
| "loss": 0.3506, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 3.8604651162790695, | |
| "grad_norm": 2.8368399143218994, | |
| "learning_rate": 7.93712316968131e-06, | |
| "loss": 0.2395, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 3.8720930232558137, | |
| "grad_norm": 2.709491014480591, | |
| "learning_rate": 7.89405684754522e-06, | |
| "loss": 0.1768, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 3.883720930232558, | |
| "grad_norm": 5.39201021194458, | |
| "learning_rate": 7.85099052540913e-06, | |
| "loss": 0.2179, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 3.895348837209302, | |
| "grad_norm": 4.398627281188965, | |
| "learning_rate": 7.807924203273041e-06, | |
| "loss": 0.2294, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 3.9069767441860463, | |
| "grad_norm": 0.8366562128067017, | |
| "learning_rate": 7.764857881136952e-06, | |
| "loss": 0.2408, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 3.9186046511627906, | |
| "grad_norm": 2.6247928142547607, | |
| "learning_rate": 7.721791559000862e-06, | |
| "loss": 0.142, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 3.9302325581395348, | |
| "grad_norm": 6.4815287590026855, | |
| "learning_rate": 7.678725236864773e-06, | |
| "loss": 0.2398, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 3.941860465116279, | |
| "grad_norm": 5.798953533172607, | |
| "learning_rate": 7.635658914728683e-06, | |
| "loss": 0.2441, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 3.953488372093023, | |
| "grad_norm": 7.927578926086426, | |
| "learning_rate": 7.592592592592594e-06, | |
| "loss": 0.187, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 3.9651162790697674, | |
| "grad_norm": 1.9918268918991089, | |
| "learning_rate": 7.5495262704565034e-06, | |
| "loss": 0.2368, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 3.9767441860465116, | |
| "grad_norm": 4.2533183097839355, | |
| "learning_rate": 7.506459948320414e-06, | |
| "loss": 0.3013, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 3.988372093023256, | |
| "grad_norm": 2.86790132522583, | |
| "learning_rate": 7.4633936261843245e-06, | |
| "loss": 0.2529, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 9.667749404907227, | |
| "learning_rate": 7.420327304048235e-06, | |
| "loss": 0.3019, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 4.011627906976744, | |
| "grad_norm": 4.353325843811035, | |
| "learning_rate": 7.3772609819121455e-06, | |
| "loss": 0.2089, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 4.023255813953488, | |
| "grad_norm": 9.287012100219727, | |
| "learning_rate": 7.334194659776056e-06, | |
| "loss": 0.2224, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 4.034883720930233, | |
| "grad_norm": 3.5896830558776855, | |
| "learning_rate": 7.2911283376399666e-06, | |
| "loss": 0.154, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 4.046511627906977, | |
| "grad_norm": 5.369675636291504, | |
| "learning_rate": 7.248062015503876e-06, | |
| "loss": 0.2273, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 4.058139534883721, | |
| "grad_norm": 6.626258373260498, | |
| "learning_rate": 7.204995693367787e-06, | |
| "loss": 0.2191, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 4.069767441860465, | |
| "grad_norm": 4.109323978424072, | |
| "learning_rate": 7.161929371231697e-06, | |
| "loss": 0.2996, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 4.0813953488372094, | |
| "grad_norm": 3.7047197818756104, | |
| "learning_rate": 7.118863049095608e-06, | |
| "loss": 0.2928, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 4.093023255813954, | |
| "grad_norm": 2.2483606338500977, | |
| "learning_rate": 7.075796726959518e-06, | |
| "loss": 0.2117, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 4.104651162790698, | |
| "grad_norm": 5.811885833740234, | |
| "learning_rate": 7.032730404823429e-06, | |
| "loss": 0.1696, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 4.116279069767442, | |
| "grad_norm": 7.897636413574219, | |
| "learning_rate": 6.9896640826873386e-06, | |
| "loss": 0.3472, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 4.127906976744186, | |
| "grad_norm": 3.9192512035369873, | |
| "learning_rate": 6.946597760551249e-06, | |
| "loss": 0.2767, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 4.1395348837209305, | |
| "grad_norm": 2.2177700996398926, | |
| "learning_rate": 6.90353143841516e-06, | |
| "loss": 0.3494, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 4.151162790697675, | |
| "grad_norm": 4.780468940734863, | |
| "learning_rate": 6.86046511627907e-06, | |
| "loss": 0.2406, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 4.162790697674419, | |
| "grad_norm": 4.2417144775390625, | |
| "learning_rate": 6.817398794142981e-06, | |
| "loss": 0.3352, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 4.174418604651163, | |
| "grad_norm": 0.6835489869117737, | |
| "learning_rate": 6.774332472006891e-06, | |
| "loss": 0.3347, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 4.186046511627907, | |
| "grad_norm": 3.9152488708496094, | |
| "learning_rate": 6.731266149870802e-06, | |
| "loss": 0.2163, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 4.1976744186046515, | |
| "grad_norm": 2.834073781967163, | |
| "learning_rate": 6.688199827734711e-06, | |
| "loss": 0.1722, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 4.209302325581396, | |
| "grad_norm": 5.854266166687012, | |
| "learning_rate": 6.645133505598622e-06, | |
| "loss": 0.1984, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 4.22093023255814, | |
| "grad_norm": 0.08869998902082443, | |
| "learning_rate": 6.6020671834625324e-06, | |
| "loss": 0.2, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 4.232558139534884, | |
| "grad_norm": 3.8255410194396973, | |
| "learning_rate": 6.559000861326443e-06, | |
| "loss": 0.1593, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 4.2441860465116275, | |
| "grad_norm": 4.605363845825195, | |
| "learning_rate": 6.5159345391903535e-06, | |
| "loss": 0.1738, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 4.2558139534883725, | |
| "grad_norm": 3.014634609222412, | |
| "learning_rate": 6.472868217054265e-06, | |
| "loss": 0.2175, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 4.267441860465116, | |
| "grad_norm": 2.7309672832489014, | |
| "learning_rate": 6.429801894918174e-06, | |
| "loss": 0.2626, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 4.27906976744186, | |
| "grad_norm": 1.9590706825256348, | |
| "learning_rate": 6.386735572782084e-06, | |
| "loss": 0.2268, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 4.290697674418604, | |
| "grad_norm": 5.337926864624023, | |
| "learning_rate": 6.343669250645995e-06, | |
| "loss": 0.3317, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 4.3023255813953485, | |
| "grad_norm": 12.289432525634766, | |
| "learning_rate": 6.300602928509906e-06, | |
| "loss": 0.3473, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 4.313953488372093, | |
| "grad_norm": 14.134299278259277, | |
| "learning_rate": 6.257536606373817e-06, | |
| "loss": 0.3093, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 4.325581395348837, | |
| "grad_norm": 2.1748945713043213, | |
| "learning_rate": 6.214470284237727e-06, | |
| "loss": 0.1843, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 4.337209302325581, | |
| "grad_norm": 3.492067813873291, | |
| "learning_rate": 6.171403962101638e-06, | |
| "loss": 0.2553, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 4.348837209302325, | |
| "grad_norm": 1.5793895721435547, | |
| "learning_rate": 6.128337639965547e-06, | |
| "loss": 0.3061, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 4.3604651162790695, | |
| "grad_norm": 1.2815021276474, | |
| "learning_rate": 6.085271317829458e-06, | |
| "loss": 0.1954, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 4.372093023255814, | |
| "grad_norm": 1.3373526334762573, | |
| "learning_rate": 6.042204995693368e-06, | |
| "loss": 0.22, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 4.383720930232558, | |
| "grad_norm": 6.806828022003174, | |
| "learning_rate": 5.999138673557279e-06, | |
| "loss": 0.215, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 4.395348837209302, | |
| "grad_norm": 0.1644594818353653, | |
| "learning_rate": 5.9560723514211895e-06, | |
| "loss": 0.2546, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 4.406976744186046, | |
| "grad_norm": 5.289101600646973, | |
| "learning_rate": 5.9130060292851e-06, | |
| "loss": 0.2607, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 4.4186046511627906, | |
| "grad_norm": 2.28316330909729, | |
| "learning_rate": 5.86993970714901e-06, | |
| "loss": 0.3033, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 4.430232558139535, | |
| "grad_norm": 2.748676061630249, | |
| "learning_rate": 5.82687338501292e-06, | |
| "loss": 0.2547, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 4.441860465116279, | |
| "grad_norm": 3.8468544483184814, | |
| "learning_rate": 5.783807062876831e-06, | |
| "loss": 0.2835, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 4.453488372093023, | |
| "grad_norm": 0.5033296942710876, | |
| "learning_rate": 5.740740740740741e-06, | |
| "loss": 0.2592, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 4.465116279069767, | |
| "grad_norm": 5.293764591217041, | |
| "learning_rate": 5.697674418604652e-06, | |
| "loss": 0.3732, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 4.476744186046512, | |
| "grad_norm": 5.627339839935303, | |
| "learning_rate": 5.654608096468562e-06, | |
| "loss": 0.2281, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 4.488372093023256, | |
| "grad_norm": 5.63222074508667, | |
| "learning_rate": 5.611541774332473e-06, | |
| "loss": 0.2226, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 4.5, | |
| "grad_norm": 7.6467366218566895, | |
| "learning_rate": 5.572782084409992e-06, | |
| "loss": 0.3368, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 4.511627906976744, | |
| "grad_norm": 7.069427967071533, | |
| "learning_rate": 5.529715762273902e-06, | |
| "loss": 0.2958, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 4.523255813953488, | |
| "grad_norm": 4.929359436035156, | |
| "learning_rate": 5.486649440137812e-06, | |
| "loss": 0.1646, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 4.534883720930233, | |
| "grad_norm": 6.314108848571777, | |
| "learning_rate": 5.443583118001723e-06, | |
| "loss": 0.2944, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 4.546511627906977, | |
| "grad_norm": 6.466343402862549, | |
| "learning_rate": 5.400516795865633e-06, | |
| "loss": 0.1311, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 4.558139534883721, | |
| "grad_norm": 7.1072163581848145, | |
| "learning_rate": 5.357450473729544e-06, | |
| "loss": 0.2246, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 4.569767441860465, | |
| "grad_norm": 3.7222752571105957, | |
| "learning_rate": 5.314384151593454e-06, | |
| "loss": 0.3476, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 4.5813953488372094, | |
| "grad_norm": 8.975520133972168, | |
| "learning_rate": 5.271317829457366e-06, | |
| "loss": 0.2638, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 4.593023255813954, | |
| "grad_norm": 6.1978864669799805, | |
| "learning_rate": 5.2282515073212745e-06, | |
| "loss": 0.2904, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 4.604651162790698, | |
| "grad_norm": 1.6185115575790405, | |
| "learning_rate": 5.185185185185185e-06, | |
| "loss": 0.258, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 4.616279069767442, | |
| "grad_norm": 1.113917589187622, | |
| "learning_rate": 5.1421188630490955e-06, | |
| "loss": 0.2658, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 4.627906976744186, | |
| "grad_norm": 2.1229023933410645, | |
| "learning_rate": 5.099052540913007e-06, | |
| "loss": 0.1123, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 4.6395348837209305, | |
| "grad_norm": 5.237680435180664, | |
| "learning_rate": 5.055986218776917e-06, | |
| "loss": 0.2383, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 4.651162790697675, | |
| "grad_norm": 5.1341938972473145, | |
| "learning_rate": 5.012919896640828e-06, | |
| "loss": 0.2337, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 4.651162790697675, | |
| "eval_accuracy": 0.9796511627906976, | |
| "eval_loss": 0.0700233206152916, | |
| "eval_precision": 0.978001140295525, | |
| "eval_recall": 0.978490990990991, | |
| "eval_runtime": 54.9842, | |
| "eval_samples_per_second": 31.282, | |
| "eval_steps_per_second": 3.91, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 4.662790697674419, | |
| "grad_norm": 4.922736167907715, | |
| "learning_rate": 4.969853574504738e-06, | |
| "loss": 0.1442, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 4.674418604651163, | |
| "grad_norm": 0.1979890614748001, | |
| "learning_rate": 4.926787252368648e-06, | |
| "loss": 0.2224, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 4.686046511627907, | |
| "grad_norm": 5.6324286460876465, | |
| "learning_rate": 4.883720930232559e-06, | |
| "loss": 0.309, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 4.6976744186046515, | |
| "grad_norm": 5.676214694976807, | |
| "learning_rate": 4.840654608096469e-06, | |
| "loss": 0.2268, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 4.709302325581396, | |
| "grad_norm": 2.2974443435668945, | |
| "learning_rate": 4.79758828596038e-06, | |
| "loss": 0.2963, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 4.720930232558139, | |
| "grad_norm": 3.438241958618164, | |
| "learning_rate": 4.754521963824289e-06, | |
| "loss": 0.266, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 4.732558139534884, | |
| "grad_norm": 8.339725494384766, | |
| "learning_rate": 4.7114556416882e-06, | |
| "loss": 0.2001, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 4.7441860465116275, | |
| "grad_norm": 11.54576587677002, | |
| "learning_rate": 4.6683893195521104e-06, | |
| "loss": 0.2883, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 4.7558139534883725, | |
| "grad_norm": 4.653608798980713, | |
| "learning_rate": 4.625322997416021e-06, | |
| "loss": 0.2116, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 4.767441860465116, | |
| "grad_norm": 3.2874107360839844, | |
| "learning_rate": 4.5822566752799315e-06, | |
| "loss": 0.1882, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 4.779069767441861, | |
| "grad_norm": 0.14985981583595276, | |
| "learning_rate": 4.539190353143842e-06, | |
| "loss": 0.193, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 4.790697674418604, | |
| "grad_norm": 5.176823139190674, | |
| "learning_rate": 4.4961240310077525e-06, | |
| "loss": 0.1616, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 4.8023255813953485, | |
| "grad_norm": 8.732250213623047, | |
| "learning_rate": 4.453057708871662e-06, | |
| "loss": 0.2681, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 4.813953488372093, | |
| "grad_norm": 4.840031147003174, | |
| "learning_rate": 4.409991386735573e-06, | |
| "loss": 0.2387, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 4.825581395348837, | |
| "grad_norm": 2.597043991088867, | |
| "learning_rate": 4.366925064599484e-06, | |
| "loss": 0.2194, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 4.837209302325581, | |
| "grad_norm": 1.7233494520187378, | |
| "learning_rate": 4.323858742463394e-06, | |
| "loss": 0.267, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 4.848837209302325, | |
| "grad_norm": 0.5028674006462097, | |
| "learning_rate": 4.280792420327304e-06, | |
| "loss": 0.181, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 4.8604651162790695, | |
| "grad_norm": 3.6632041931152344, | |
| "learning_rate": 4.237726098191215e-06, | |
| "loss": 0.1867, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 4.872093023255814, | |
| "grad_norm": 0.21562618017196655, | |
| "learning_rate": 4.194659776055125e-06, | |
| "loss": 0.2073, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 4.883720930232558, | |
| "grad_norm": 5.712989330291748, | |
| "learning_rate": 4.151593453919036e-06, | |
| "loss": 0.1728, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 4.895348837209302, | |
| "grad_norm": 7.611968994140625, | |
| "learning_rate": 4.108527131782946e-06, | |
| "loss": 0.2316, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 4.906976744186046, | |
| "grad_norm": 0.3871056139469147, | |
| "learning_rate": 4.065460809646857e-06, | |
| "loss": 0.2208, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 4.9186046511627906, | |
| "grad_norm": 4.011594772338867, | |
| "learning_rate": 4.022394487510767e-06, | |
| "loss": 0.2032, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 4.930232558139535, | |
| "grad_norm": 0.46288034319877625, | |
| "learning_rate": 3.979328165374677e-06, | |
| "loss": 0.0885, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 4.941860465116279, | |
| "grad_norm": 4.409682750701904, | |
| "learning_rate": 3.936261843238588e-06, | |
| "loss": 0.3538, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 4.953488372093023, | |
| "grad_norm": 15.171357154846191, | |
| "learning_rate": 3.893195521102498e-06, | |
| "loss": 0.3007, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 4.965116279069767, | |
| "grad_norm": 3.006840229034424, | |
| "learning_rate": 3.850129198966409e-06, | |
| "loss": 0.1961, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 4.976744186046512, | |
| "grad_norm": 3.0607519149780273, | |
| "learning_rate": 3.8070628768303192e-06, | |
| "loss": 0.3155, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 4.988372093023256, | |
| "grad_norm": 1.423635482788086, | |
| "learning_rate": 3.7639965546942293e-06, | |
| "loss": 0.1748, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 3.1234054565429688, | |
| "learning_rate": 3.72093023255814e-06, | |
| "loss": 0.273, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 5.011627906976744, | |
| "grad_norm": 5.210692405700684, | |
| "learning_rate": 3.6778639104220504e-06, | |
| "loss": 0.2768, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 5.023255813953488, | |
| "grad_norm": 7.260842800140381, | |
| "learning_rate": 3.6347975882859605e-06, | |
| "loss": 0.2462, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 5.034883720930233, | |
| "grad_norm": 4.551229953765869, | |
| "learning_rate": 3.591731266149871e-06, | |
| "loss": 0.2205, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 5.046511627906977, | |
| "grad_norm": 2.6683349609375, | |
| "learning_rate": 3.5486649440137815e-06, | |
| "loss": 0.2263, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 5.058139534883721, | |
| "grad_norm": 3.597114086151123, | |
| "learning_rate": 3.505598621877692e-06, | |
| "loss": 0.2586, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 5.069767441860465, | |
| "grad_norm": 3.4024291038513184, | |
| "learning_rate": 3.462532299741602e-06, | |
| "loss": 0.0955, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 5.0813953488372094, | |
| "grad_norm": 7.212930679321289, | |
| "learning_rate": 3.4194659776055127e-06, | |
| "loss": 0.2704, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 5.093023255813954, | |
| "grad_norm": 5.682826995849609, | |
| "learning_rate": 3.3763996554694232e-06, | |
| "loss": 0.1855, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 5.104651162790698, | |
| "grad_norm": 0.4372553825378418, | |
| "learning_rate": 3.3333333333333333e-06, | |
| "loss": 0.1976, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 5.116279069767442, | |
| "grad_norm": 4.182933330535889, | |
| "learning_rate": 3.290267011197244e-06, | |
| "loss": 0.2642, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 5.127906976744186, | |
| "grad_norm": 2.625610828399658, | |
| "learning_rate": 3.247200689061155e-06, | |
| "loss": 0.1924, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 5.1395348837209305, | |
| "grad_norm": 4.952889442443848, | |
| "learning_rate": 3.2041343669250645e-06, | |
| "loss": 0.2458, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 5.151162790697675, | |
| "grad_norm": 2.9454457759857178, | |
| "learning_rate": 3.1610680447889754e-06, | |
| "loss": 0.2049, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 5.162790697674419, | |
| "grad_norm": 1.9158250093460083, | |
| "learning_rate": 3.118001722652886e-06, | |
| "loss": 0.2577, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 5.174418604651163, | |
| "grad_norm": 3.5840890407562256, | |
| "learning_rate": 3.074935400516796e-06, | |
| "loss": 0.2404, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 5.186046511627907, | |
| "grad_norm": 8.794764518737793, | |
| "learning_rate": 3.0318690783807066e-06, | |
| "loss": 0.1885, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 5.1976744186046515, | |
| "grad_norm": 3.7340569496154785, | |
| "learning_rate": 2.988802756244617e-06, | |
| "loss": 0.1926, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 5.209302325581396, | |
| "grad_norm": 5.893591403961182, | |
| "learning_rate": 2.9457364341085276e-06, | |
| "loss": 0.314, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 5.22093023255814, | |
| "grad_norm": 1.9155763387680054, | |
| "learning_rate": 2.9026701119724377e-06, | |
| "loss": 0.1849, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 5.232558139534884, | |
| "grad_norm": 2.542473077774048, | |
| "learning_rate": 2.8596037898363483e-06, | |
| "loss": 0.1828, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 5.2441860465116275, | |
| "grad_norm": 2.4870667457580566, | |
| "learning_rate": 2.8165374677002588e-06, | |
| "loss": 0.1787, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 5.2558139534883725, | |
| "grad_norm": 4.764167308807373, | |
| "learning_rate": 2.773471145564169e-06, | |
| "loss": 0.2819, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 5.267441860465116, | |
| "grad_norm": 0.5642258524894714, | |
| "learning_rate": 2.7304048234280794e-06, | |
| "loss": 0.1526, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 5.27906976744186, | |
| "grad_norm": 7.091589450836182, | |
| "learning_rate": 2.68733850129199e-06, | |
| "loss": 0.2592, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 5.290697674418604, | |
| "grad_norm": 6.385696887969971, | |
| "learning_rate": 2.6442721791559e-06, | |
| "loss": 0.2047, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 5.3023255813953485, | |
| "grad_norm": 9.775449752807617, | |
| "learning_rate": 2.6012058570198106e-06, | |
| "loss": 0.3159, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 5.313953488372093, | |
| "grad_norm": 7.352219581604004, | |
| "learning_rate": 2.558139534883721e-06, | |
| "loss": 0.2154, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 5.325581395348837, | |
| "grad_norm": 5.333184719085693, | |
| "learning_rate": 2.515073212747632e-06, | |
| "loss": 0.1613, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 5.337209302325581, | |
| "grad_norm": 5.456418037414551, | |
| "learning_rate": 2.4720068906115417e-06, | |
| "loss": 0.2579, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 5.348837209302325, | |
| "grad_norm": 0.07044707238674164, | |
| "learning_rate": 2.4289405684754527e-06, | |
| "loss": 0.1915, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 5.3604651162790695, | |
| "grad_norm": 6.497509479522705, | |
| "learning_rate": 2.3858742463393628e-06, | |
| "loss": 0.2478, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 5.372093023255814, | |
| "grad_norm": 5.096982479095459, | |
| "learning_rate": 2.3428079242032733e-06, | |
| "loss": 0.3055, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 5.383720930232558, | |
| "grad_norm": 7.590965747833252, | |
| "learning_rate": 2.299741602067184e-06, | |
| "loss": 0.184, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 5.395348837209302, | |
| "grad_norm": 2.4741833209991455, | |
| "learning_rate": 2.256675279931094e-06, | |
| "loss": 0.1963, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 5.406976744186046, | |
| "grad_norm": 9.394286155700684, | |
| "learning_rate": 2.2136089577950044e-06, | |
| "loss": 0.3012, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 5.4186046511627906, | |
| "grad_norm": 0.2691817581653595, | |
| "learning_rate": 2.170542635658915e-06, | |
| "loss": 0.2925, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 5.430232558139535, | |
| "grad_norm": 5.20473575592041, | |
| "learning_rate": 2.127476313522825e-06, | |
| "loss": 0.2325, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 5.441860465116279, | |
| "grad_norm": 3.867701530456543, | |
| "learning_rate": 2.0844099913867356e-06, | |
| "loss": 0.2992, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 5.453488372093023, | |
| "grad_norm": 0.7283692955970764, | |
| "learning_rate": 2.041343669250646e-06, | |
| "loss": 0.1842, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 5.465116279069767, | |
| "grad_norm": 4.095493793487549, | |
| "learning_rate": 1.9982773471145566e-06, | |
| "loss": 0.1471, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 5.476744186046512, | |
| "grad_norm": 6.952198505401611, | |
| "learning_rate": 1.955211024978467e-06, | |
| "loss": 0.3796, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 5.488372093023256, | |
| "grad_norm": 2.4238672256469727, | |
| "learning_rate": 1.9121447028423773e-06, | |
| "loss": 0.2519, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 5.5, | |
| "grad_norm": 1.495030403137207, | |
| "learning_rate": 1.8690783807062878e-06, | |
| "loss": 0.1058, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 5.511627906976744, | |
| "grad_norm": 3.9967563152313232, | |
| "learning_rate": 1.826012058570198e-06, | |
| "loss": 0.2686, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 5.523255813953488, | |
| "grad_norm": 2.738013505935669, | |
| "learning_rate": 1.7829457364341088e-06, | |
| "loss": 0.1865, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 5.534883720930233, | |
| "grad_norm": 8.840526580810547, | |
| "learning_rate": 1.7398794142980192e-06, | |
| "loss": 0.2717, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 5.546511627906977, | |
| "grad_norm": 3.8679466247558594, | |
| "learning_rate": 1.6968130921619295e-06, | |
| "loss": 0.2082, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 5.558139534883721, | |
| "grad_norm": 4.620917797088623, | |
| "learning_rate": 1.65374677002584e-06, | |
| "loss": 0.2089, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 5.569767441860465, | |
| "grad_norm": 1.7068657875061035, | |
| "learning_rate": 1.6106804478897503e-06, | |
| "loss": 0.2387, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 5.5813953488372094, | |
| "grad_norm": 6.4312896728515625, | |
| "learning_rate": 1.5676141257536608e-06, | |
| "loss": 0.1885, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 5.593023255813954, | |
| "grad_norm": 4.719575881958008, | |
| "learning_rate": 1.5245478036175711e-06, | |
| "loss": 0.2277, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 5.604651162790698, | |
| "grad_norm": 2.6920177936553955, | |
| "learning_rate": 1.4814814814814815e-06, | |
| "loss": 0.2396, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 5.616279069767442, | |
| "grad_norm": 3.6449053287506104, | |
| "learning_rate": 1.4384151593453922e-06, | |
| "loss": 0.1681, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 5.627906976744186, | |
| "grad_norm": 7.3726325035095215, | |
| "learning_rate": 1.3953488372093025e-06, | |
| "loss": 0.2193, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 5.6395348837209305, | |
| "grad_norm": 8.410340309143066, | |
| "learning_rate": 1.3522825150732128e-06, | |
| "loss": 0.1999, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 5.651162790697675, | |
| "grad_norm": 1.85641348361969, | |
| "learning_rate": 1.3092161929371233e-06, | |
| "loss": 0.2887, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 5.662790697674419, | |
| "grad_norm": 4.673345565795898, | |
| "learning_rate": 1.2661498708010337e-06, | |
| "loss": 0.3156, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 5.674418604651163, | |
| "grad_norm": 7.408679008483887, | |
| "learning_rate": 1.2230835486649442e-06, | |
| "loss": 0.1357, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 5.686046511627907, | |
| "grad_norm": 1.3766252994537354, | |
| "learning_rate": 1.1800172265288545e-06, | |
| "loss": 0.2649, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 5.6976744186046515, | |
| "grad_norm": 3.031282663345337, | |
| "learning_rate": 1.136950904392765e-06, | |
| "loss": 0.2099, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 5.709302325581396, | |
| "grad_norm": 7.257900714874268, | |
| "learning_rate": 1.0938845822566753e-06, | |
| "loss": 0.2836, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 5.720930232558139, | |
| "grad_norm": 6.421561241149902, | |
| "learning_rate": 1.0508182601205856e-06, | |
| "loss": 0.277, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 5.732558139534884, | |
| "grad_norm": 9.094550132751465, | |
| "learning_rate": 1.0077519379844962e-06, | |
| "loss": 0.2506, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 5.7441860465116275, | |
| "grad_norm": 3.772926092147827, | |
| "learning_rate": 9.646856158484067e-07, | |
| "loss": 0.2569, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 5.7558139534883725, | |
| "grad_norm": 3.3292300701141357, | |
| "learning_rate": 9.216192937123171e-07, | |
| "loss": 0.2032, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 5.767441860465116, | |
| "grad_norm": 4.869121074676514, | |
| "learning_rate": 8.785529715762274e-07, | |
| "loss": 0.1975, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 5.779069767441861, | |
| "grad_norm": 0.4935530424118042, | |
| "learning_rate": 8.354866494401378e-07, | |
| "loss": 0.1365, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 5.790697674418604, | |
| "grad_norm": 1.3626408576965332, | |
| "learning_rate": 7.924203273040483e-07, | |
| "loss": 0.1084, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 5.8023255813953485, | |
| "grad_norm": 9.085593223571777, | |
| "learning_rate": 7.493540051679588e-07, | |
| "loss": 0.3171, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 5.813953488372093, | |
| "grad_norm": 0.36338621377944946, | |
| "learning_rate": 7.062876830318692e-07, | |
| "loss": 0.1801, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 5.813953488372093, | |
| "eval_accuracy": 0.9837209302325581, | |
| "eval_loss": 0.05773168057203293, | |
| "eval_precision": 0.9820503676194734, | |
| "eval_recall": 0.982545045045045, | |
| "eval_runtime": 55.3139, | |
| "eval_samples_per_second": 31.095, | |
| "eval_steps_per_second": 3.887, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 5.825581395348837, | |
| "grad_norm": 2.7037956714630127, | |
| "learning_rate": 6.632213608957795e-07, | |
| "loss": 0.2007, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 5.837209302325581, | |
| "grad_norm": 11.454206466674805, | |
| "learning_rate": 6.201550387596899e-07, | |
| "loss": 0.1942, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 5.848837209302325, | |
| "grad_norm": 3.066746711730957, | |
| "learning_rate": 5.770887166236004e-07, | |
| "loss": 0.2566, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 5.8604651162790695, | |
| "grad_norm": 3.7300071716308594, | |
| "learning_rate": 5.340223944875108e-07, | |
| "loss": 0.2125, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 5.872093023255814, | |
| "grad_norm": 2.9397366046905518, | |
| "learning_rate": 4.909560723514212e-07, | |
| "loss": 0.2984, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 5.883720930232558, | |
| "grad_norm": 12.143790245056152, | |
| "learning_rate": 4.478897502153316e-07, | |
| "loss": 0.2164, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 5.895348837209302, | |
| "grad_norm": 3.810537099838257, | |
| "learning_rate": 4.048234280792421e-07, | |
| "loss": 0.2565, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 5.906976744186046, | |
| "grad_norm": 2.6605870723724365, | |
| "learning_rate": 3.6175710594315246e-07, | |
| "loss": 0.1347, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 5.9186046511627906, | |
| "grad_norm": 0.5655523538589478, | |
| "learning_rate": 3.1869078380706293e-07, | |
| "loss": 0.2056, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 5.930232558139535, | |
| "grad_norm": 5.374477863311768, | |
| "learning_rate": 2.756244616709733e-07, | |
| "loss": 0.2222, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 5.941860465116279, | |
| "grad_norm": 2.566845417022705, | |
| "learning_rate": 2.3255813953488374e-07, | |
| "loss": 0.105, | |
| "step": 5110 | |
| }, | |
| { | |
| "epoch": 5.953488372093023, | |
| "grad_norm": 4.244629383087158, | |
| "learning_rate": 1.8949181739879416e-07, | |
| "loss": 0.1932, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 5.965116279069767, | |
| "grad_norm": 8.944864273071289, | |
| "learning_rate": 1.4642549526270458e-07, | |
| "loss": 0.2709, | |
| "step": 5130 | |
| }, | |
| { | |
| "epoch": 5.976744186046512, | |
| "grad_norm": 4.647603988647461, | |
| "learning_rate": 1.03359173126615e-07, | |
| "loss": 0.1881, | |
| "step": 5140 | |
| }, | |
| { | |
| "epoch": 5.988372093023256, | |
| "grad_norm": 6.423856258392334, | |
| "learning_rate": 6.029285099052542e-08, | |
| "loss": 0.2541, | |
| "step": 5150 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "grad_norm": 5.573716163635254, | |
| "learning_rate": 1.722652885443583e-08, | |
| "loss": 0.2445, | |
| "step": 5160 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "step": 5160, | |
| "total_flos": 6.399745594361119e+18, | |
| "train_loss": 0.5480184263730234, | |
| "train_runtime": 2905.2086, | |
| "train_samples_per_second": 28.418, | |
| "train_steps_per_second": 1.776 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 5160, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 6, | |
| "save_steps": 1000, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 6.399745594361119e+18, | |
| "train_batch_size": 16, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |