| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 3.0, | |
| "eval_steps": 500, | |
| "global_step": 3000, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 31.807226181030273, | |
| "learning_rate": 4.5e-06, | |
| "loss": 10.2322, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 14.19406795501709, | |
| "learning_rate": 9.5e-06, | |
| "loss": 8.6686, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 8.947138786315918, | |
| "learning_rate": 1.45e-05, | |
| "loss": 6.6091, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 7.508649826049805, | |
| "learning_rate": 1.9500000000000003e-05, | |
| "loss": 5.1112, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 5.768527030944824, | |
| "learning_rate": 2.45e-05, | |
| "loss": 4.4247, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 5.962294578552246, | |
| "learning_rate": 2.95e-05, | |
| "loss": 3.7838, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 7.343990802764893, | |
| "learning_rate": 3.45e-05, | |
| "loss": 3.3172, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 5.001032829284668, | |
| "learning_rate": 3.9500000000000005e-05, | |
| "loss": 3.1734, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 3.4707157611846924, | |
| "learning_rate": 4.4500000000000004e-05, | |
| "loss": 2.9784, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 4.321723461151123, | |
| "learning_rate": 4.9500000000000004e-05, | |
| "loss": 2.6301, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 4.037026405334473, | |
| "learning_rate": 4.98448275862069e-05, | |
| "loss": 2.8441, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 6.042785167694092, | |
| "learning_rate": 4.967241379310345e-05, | |
| "loss": 2.5558, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 7.331098556518555, | |
| "learning_rate": 4.9500000000000004e-05, | |
| "loss": 2.5997, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 5.285608291625977, | |
| "learning_rate": 4.932758620689655e-05, | |
| "loss": 2.3912, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 5.633673191070557, | |
| "learning_rate": 4.915517241379311e-05, | |
| "loss": 2.3576, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 5.462876796722412, | |
| "learning_rate": 4.898275862068966e-05, | |
| "loss": 2.3879, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 5.2996015548706055, | |
| "learning_rate": 4.8810344827586205e-05, | |
| "loss": 2.6634, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 4.960510730743408, | |
| "learning_rate": 4.863793103448276e-05, | |
| "loss": 2.3228, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 5.18018913269043, | |
| "learning_rate": 4.846551724137931e-05, | |
| "loss": 2.4035, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 3.5356884002685547, | |
| "learning_rate": 4.8293103448275865e-05, | |
| "loss": 1.9488, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 2.8790781497955322, | |
| "learning_rate": 4.812068965517242e-05, | |
| "loss": 2.2974, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 3.502866506576538, | |
| "learning_rate": 4.794827586206897e-05, | |
| "loss": 2.1646, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 2.9694302082061768, | |
| "learning_rate": 4.777586206896552e-05, | |
| "loss": 2.1186, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 2.986870050430298, | |
| "learning_rate": 4.7603448275862073e-05, | |
| "loss": 2.1907, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 7.217125415802002, | |
| "learning_rate": 4.743103448275862e-05, | |
| "loss": 2.0278, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 2.887138843536377, | |
| "learning_rate": 4.725862068965518e-05, | |
| "loss": 2.269, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 14.621356964111328, | |
| "learning_rate": 4.7086206896551726e-05, | |
| "loss": 2.0509, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 2.5131711959838867, | |
| "learning_rate": 4.6913793103448275e-05, | |
| "loss": 2.2239, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 4.753391265869141, | |
| "learning_rate": 4.674137931034483e-05, | |
| "loss": 2.0629, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 4.106902599334717, | |
| "learning_rate": 4.656896551724138e-05, | |
| "loss": 2.0625, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 4.522510528564453, | |
| "learning_rate": 4.6396551724137935e-05, | |
| "loss": 1.9146, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 3.454258441925049, | |
| "learning_rate": 4.622413793103448e-05, | |
| "loss": 1.7679, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 12.172571182250977, | |
| "learning_rate": 4.605172413793103e-05, | |
| "loss": 1.762, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 5.5643630027771, | |
| "learning_rate": 4.587931034482759e-05, | |
| "loss": 1.9651, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 5.371229648590088, | |
| "learning_rate": 4.5706896551724136e-05, | |
| "loss": 2.1326, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 2.6412742137908936, | |
| "learning_rate": 4.553448275862069e-05, | |
| "loss": 1.955, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 5.9563775062561035, | |
| "learning_rate": 4.536206896551725e-05, | |
| "loss": 1.8654, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 5.498210906982422, | |
| "learning_rate": 4.5189655172413796e-05, | |
| "loss": 2.0221, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 1.9855848550796509, | |
| "learning_rate": 4.501724137931035e-05, | |
| "loss": 1.9202, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 6.402889728546143, | |
| "learning_rate": 4.48448275862069e-05, | |
| "loss": 1.6582, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 1.9488422870635986, | |
| "learning_rate": 4.467241379310345e-05, | |
| "loss": 1.9576, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 1.99562406539917, | |
| "learning_rate": 4.4500000000000004e-05, | |
| "loss": 1.6823, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 3.2172203063964844, | |
| "learning_rate": 4.432758620689655e-05, | |
| "loss": 1.7448, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 1.7004073858261108, | |
| "learning_rate": 4.415517241379311e-05, | |
| "loss": 1.687, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 2.6527938842773438, | |
| "learning_rate": 4.398275862068966e-05, | |
| "loss": 1.6152, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 2.223691463470459, | |
| "learning_rate": 4.3810344827586206e-05, | |
| "loss": 1.7911, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 2.228793144226074, | |
| "learning_rate": 4.363793103448276e-05, | |
| "loss": 1.9451, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 2.724231243133545, | |
| "learning_rate": 4.346551724137931e-05, | |
| "loss": 1.791, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 3.219877243041992, | |
| "learning_rate": 4.3293103448275865e-05, | |
| "loss": 1.692, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 2.977465867996216, | |
| "learning_rate": 4.3120689655172414e-05, | |
| "loss": 1.886, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 2.566962242126465, | |
| "learning_rate": 4.294827586206896e-05, | |
| "loss": 1.802, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.52, | |
| "grad_norm": 3.129835844039917, | |
| "learning_rate": 4.2775862068965525e-05, | |
| "loss": 1.659, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.53, | |
| "grad_norm": 3.913256883621216, | |
| "learning_rate": 4.2603448275862074e-05, | |
| "loss": 1.5919, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.54, | |
| "grad_norm": 1.7998114824295044, | |
| "learning_rate": 4.243103448275862e-05, | |
| "loss": 1.7564, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.55, | |
| "grad_norm": 2.310154676437378, | |
| "learning_rate": 4.225862068965518e-05, | |
| "loss": 1.7347, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.56, | |
| "grad_norm": 1.835707426071167, | |
| "learning_rate": 4.208620689655173e-05, | |
| "loss": 1.802, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.57, | |
| "grad_norm": 1.9488534927368164, | |
| "learning_rate": 4.191379310344828e-05, | |
| "loss": 1.9981, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.58, | |
| "grad_norm": 5.861155986785889, | |
| "learning_rate": 4.174137931034483e-05, | |
| "loss": 1.9519, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.59, | |
| "grad_norm": 2.403691053390503, | |
| "learning_rate": 4.156896551724138e-05, | |
| "loss": 1.6711, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.6, | |
| "grad_norm": 1.9013047218322754, | |
| "learning_rate": 4.1396551724137935e-05, | |
| "loss": 1.6679, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.61, | |
| "grad_norm": 2.6646413803100586, | |
| "learning_rate": 4.1224137931034484e-05, | |
| "loss": 1.7117, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.62, | |
| "grad_norm": 2.6011579036712646, | |
| "learning_rate": 4.105172413793103e-05, | |
| "loss": 1.6351, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.63, | |
| "grad_norm": 1.9296406507492065, | |
| "learning_rate": 4.087931034482759e-05, | |
| "loss": 1.8179, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.64, | |
| "grad_norm": 2.2965588569641113, | |
| "learning_rate": 4.070689655172414e-05, | |
| "loss": 1.7164, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.65, | |
| "grad_norm": 1.7480602264404297, | |
| "learning_rate": 4.053448275862069e-05, | |
| "loss": 1.5552, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.66, | |
| "grad_norm": 2.9659781455993652, | |
| "learning_rate": 4.036206896551724e-05, | |
| "loss": 1.79, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.67, | |
| "grad_norm": 1.5911238193511963, | |
| "learning_rate": 4.018965517241379e-05, | |
| "loss": 1.7139, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.68, | |
| "grad_norm": 1.8339680433273315, | |
| "learning_rate": 4.0017241379310345e-05, | |
| "loss": 1.6479, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.69, | |
| "grad_norm": 1.8804699182510376, | |
| "learning_rate": 3.98448275862069e-05, | |
| "loss": 1.6501, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.7, | |
| "grad_norm": 5.505954742431641, | |
| "learning_rate": 3.967241379310345e-05, | |
| "loss": 1.549, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.71, | |
| "grad_norm": 3.225405693054199, | |
| "learning_rate": 3.9500000000000005e-05, | |
| "loss": 1.5231, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.72, | |
| "grad_norm": 2.4233756065368652, | |
| "learning_rate": 3.932758620689655e-05, | |
| "loss": 1.5577, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.73, | |
| "grad_norm": 1.9218379259109497, | |
| "learning_rate": 3.915517241379311e-05, | |
| "loss": 1.8394, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.74, | |
| "grad_norm": 1.7937753200531006, | |
| "learning_rate": 3.898275862068966e-05, | |
| "loss": 1.4735, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.75, | |
| "grad_norm": 2.696561336517334, | |
| "learning_rate": 3.8810344827586206e-05, | |
| "loss": 1.7371, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.76, | |
| "grad_norm": 1.6280263662338257, | |
| "learning_rate": 3.863793103448276e-05, | |
| "loss": 1.6161, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.77, | |
| "grad_norm": 1.7006651163101196, | |
| "learning_rate": 3.846551724137931e-05, | |
| "loss": 1.4593, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.78, | |
| "grad_norm": 3.765646457672119, | |
| "learning_rate": 3.8293103448275866e-05, | |
| "loss": 1.5506, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.79, | |
| "grad_norm": 2.9314448833465576, | |
| "learning_rate": 3.8120689655172415e-05, | |
| "loss": 1.5805, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.8, | |
| "grad_norm": 1.736654281616211, | |
| "learning_rate": 3.794827586206896e-05, | |
| "loss": 1.5675, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.81, | |
| "grad_norm": 2.3542542457580566, | |
| "learning_rate": 3.777586206896552e-05, | |
| "loss": 1.5955, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.82, | |
| "grad_norm": 2.1898491382598877, | |
| "learning_rate": 3.760344827586207e-05, | |
| "loss": 1.3708, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.83, | |
| "grad_norm": 2.2105820178985596, | |
| "learning_rate": 3.743103448275862e-05, | |
| "loss": 1.6554, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.84, | |
| "grad_norm": 4.901267051696777, | |
| "learning_rate": 3.725862068965517e-05, | |
| "loss": 1.3071, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.85, | |
| "grad_norm": 1.991741418838501, | |
| "learning_rate": 3.708620689655173e-05, | |
| "loss": 1.3876, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.86, | |
| "grad_norm": 2.1567299365997314, | |
| "learning_rate": 3.691379310344828e-05, | |
| "loss": 1.5718, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.87, | |
| "grad_norm": 1.741442084312439, | |
| "learning_rate": 3.674137931034483e-05, | |
| "loss": 1.3982, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.88, | |
| "grad_norm": 2.704728364944458, | |
| "learning_rate": 3.656896551724138e-05, | |
| "loss": 1.3117, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.89, | |
| "grad_norm": 1.5768979787826538, | |
| "learning_rate": 3.6396551724137936e-05, | |
| "loss": 1.6102, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.9, | |
| "grad_norm": 1.6916723251342773, | |
| "learning_rate": 3.6224137931034484e-05, | |
| "loss": 1.6576, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.91, | |
| "grad_norm": 2.1157259941101074, | |
| "learning_rate": 3.605172413793104e-05, | |
| "loss": 1.4144, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.92, | |
| "grad_norm": 1.7090058326721191, | |
| "learning_rate": 3.587931034482759e-05, | |
| "loss": 1.3267, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.93, | |
| "grad_norm": 2.5148727893829346, | |
| "learning_rate": 3.570689655172414e-05, | |
| "loss": 1.7362, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.94, | |
| "grad_norm": 2.4861910343170166, | |
| "learning_rate": 3.553448275862069e-05, | |
| "loss": 1.4764, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.95, | |
| "grad_norm": 1.909857988357544, | |
| "learning_rate": 3.536206896551724e-05, | |
| "loss": 1.8828, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.96, | |
| "grad_norm": 1.5957543849945068, | |
| "learning_rate": 3.51896551724138e-05, | |
| "loss": 1.4019, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.97, | |
| "grad_norm": 2.6423611640930176, | |
| "learning_rate": 3.5017241379310345e-05, | |
| "loss": 1.2464, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.98, | |
| "grad_norm": 1.7054439783096313, | |
| "learning_rate": 3.4844827586206894e-05, | |
| "loss": 1.3908, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.99, | |
| "grad_norm": 1.5509027242660522, | |
| "learning_rate": 3.467241379310345e-05, | |
| "loss": 1.6855, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 2.0527048110961914, | |
| "learning_rate": 3.45e-05, | |
| "loss": 1.41, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 1.01, | |
| "grad_norm": 1.912597417831421, | |
| "learning_rate": 3.4327586206896554e-05, | |
| "loss": 1.696, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 1.02, | |
| "grad_norm": 2.4118075370788574, | |
| "learning_rate": 3.415517241379311e-05, | |
| "loss": 1.4162, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 1.03, | |
| "grad_norm": 1.8533530235290527, | |
| "learning_rate": 3.398275862068966e-05, | |
| "loss": 1.3809, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 1.04, | |
| "grad_norm": 1.950042724609375, | |
| "learning_rate": 3.381034482758621e-05, | |
| "loss": 1.3845, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 1.05, | |
| "grad_norm": 2.852569341659546, | |
| "learning_rate": 3.363793103448276e-05, | |
| "loss": 1.3189, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 1.06, | |
| "grad_norm": 1.6454888582229614, | |
| "learning_rate": 3.346551724137931e-05, | |
| "loss": 1.4498, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 1.07, | |
| "grad_norm": 2.481353759765625, | |
| "learning_rate": 3.3293103448275866e-05, | |
| "loss": 1.4578, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 1.08, | |
| "grad_norm": 2.739814519882202, | |
| "learning_rate": 3.3120689655172415e-05, | |
| "loss": 1.4949, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 1.09, | |
| "grad_norm": 2.799699068069458, | |
| "learning_rate": 3.2948275862068964e-05, | |
| "loss": 1.1835, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 1.1, | |
| "grad_norm": 1.6168289184570312, | |
| "learning_rate": 3.277586206896552e-05, | |
| "loss": 1.4057, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 1.11, | |
| "grad_norm": 3.1404354572296143, | |
| "learning_rate": 3.260344827586207e-05, | |
| "loss": 1.2438, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 1.12, | |
| "grad_norm": 2.4057466983795166, | |
| "learning_rate": 3.2431034482758623e-05, | |
| "loss": 1.5332, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.13, | |
| "grad_norm": 1.1605113744735718, | |
| "learning_rate": 3.225862068965517e-05, | |
| "loss": 1.3803, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 1.1400000000000001, | |
| "grad_norm": 2.467651605606079, | |
| "learning_rate": 3.208620689655172e-05, | |
| "loss": 1.3769, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 1.15, | |
| "grad_norm": 2.0736987590789795, | |
| "learning_rate": 3.1913793103448276e-05, | |
| "loss": 1.4032, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 1.16, | |
| "grad_norm": 2.9988651275634766, | |
| "learning_rate": 3.1741379310344825e-05, | |
| "loss": 1.4226, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 1.17, | |
| "grad_norm": 2.2302157878875732, | |
| "learning_rate": 3.156896551724138e-05, | |
| "loss": 1.5159, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 1.18, | |
| "grad_norm": 1.2775397300720215, | |
| "learning_rate": 3.1396551724137936e-05, | |
| "loss": 1.2539, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 1.19, | |
| "grad_norm": 3.2107908725738525, | |
| "learning_rate": 3.1224137931034485e-05, | |
| "loss": 1.3425, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 1.2, | |
| "grad_norm": 1.783532977104187, | |
| "learning_rate": 3.105172413793104e-05, | |
| "loss": 1.2874, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 1.21, | |
| "grad_norm": 2.1126065254211426, | |
| "learning_rate": 3.087931034482759e-05, | |
| "loss": 1.5065, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 1.22, | |
| "grad_norm": 2.010965347290039, | |
| "learning_rate": 3.070689655172414e-05, | |
| "loss": 1.3932, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 1.23, | |
| "grad_norm": 5.611945152282715, | |
| "learning_rate": 3.053448275862069e-05, | |
| "loss": 1.4646, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 1.24, | |
| "grad_norm": 1.9755380153656006, | |
| "learning_rate": 3.0362068965517242e-05, | |
| "loss": 1.1764, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "grad_norm": 2.106905221939087, | |
| "learning_rate": 3.0189655172413794e-05, | |
| "loss": 1.3547, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 1.26, | |
| "grad_norm": 1.7636513710021973, | |
| "learning_rate": 3.0017241379310346e-05, | |
| "loss": 1.2024, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 1.27, | |
| "grad_norm": 2.3909709453582764, | |
| "learning_rate": 2.9844827586206898e-05, | |
| "loss": 1.2256, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 1.28, | |
| "grad_norm": 1.8174681663513184, | |
| "learning_rate": 2.967241379310345e-05, | |
| "loss": 1.4106, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.29, | |
| "grad_norm": 1.857674241065979, | |
| "learning_rate": 2.95e-05, | |
| "loss": 1.1851, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 1.3, | |
| "grad_norm": 1.9900033473968506, | |
| "learning_rate": 2.932758620689655e-05, | |
| "loss": 1.2333, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 1.31, | |
| "grad_norm": 1.5615085363388062, | |
| "learning_rate": 2.9155172413793103e-05, | |
| "loss": 1.3969, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 1.32, | |
| "grad_norm": 1.7219326496124268, | |
| "learning_rate": 2.8982758620689655e-05, | |
| "loss": 1.2629, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 1.33, | |
| "grad_norm": 1.7303168773651123, | |
| "learning_rate": 2.881034482758621e-05, | |
| "loss": 1.4391, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 1.34, | |
| "grad_norm": 1.7237037420272827, | |
| "learning_rate": 2.8637931034482763e-05, | |
| "loss": 1.3606, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 1.35, | |
| "grad_norm": 1.575646162033081, | |
| "learning_rate": 2.8465517241379315e-05, | |
| "loss": 1.1907, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 1.3599999999999999, | |
| "grad_norm": 2.1866729259490967, | |
| "learning_rate": 2.8293103448275863e-05, | |
| "loss": 1.3406, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 1.37, | |
| "grad_norm": 1.5900951623916626, | |
| "learning_rate": 2.8120689655172415e-05, | |
| "loss": 1.3234, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 1.38, | |
| "grad_norm": 2.729886293411255, | |
| "learning_rate": 2.7948275862068968e-05, | |
| "loss": 1.3068, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 1.3900000000000001, | |
| "grad_norm": 7.489139556884766, | |
| "learning_rate": 2.777586206896552e-05, | |
| "loss": 1.2874, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 1.4, | |
| "grad_norm": 1.6766670942306519, | |
| "learning_rate": 2.7603448275862072e-05, | |
| "loss": 1.369, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 1.41, | |
| "grad_norm": 1.6040589809417725, | |
| "learning_rate": 2.743103448275862e-05, | |
| "loss": 1.4296, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 1.42, | |
| "grad_norm": 2.0411570072174072, | |
| "learning_rate": 2.7258620689655173e-05, | |
| "loss": 1.2803, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 1.43, | |
| "grad_norm": 2.604081869125366, | |
| "learning_rate": 2.7086206896551725e-05, | |
| "loss": 1.2784, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 1.44, | |
| "grad_norm": 1.4787765741348267, | |
| "learning_rate": 2.6913793103448277e-05, | |
| "loss": 1.3021, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.45, | |
| "grad_norm": 1.665315866470337, | |
| "learning_rate": 2.674137931034483e-05, | |
| "loss": 1.1502, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 1.46, | |
| "grad_norm": 2.4420969486236572, | |
| "learning_rate": 2.6568965517241378e-05, | |
| "loss": 1.3531, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 1.47, | |
| "grad_norm": 1.961679220199585, | |
| "learning_rate": 2.639655172413793e-05, | |
| "loss": 1.413, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 1.48, | |
| "grad_norm": 1.6548084020614624, | |
| "learning_rate": 2.6224137931034482e-05, | |
| "loss": 1.2579, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 1.49, | |
| "grad_norm": 1.22232985496521, | |
| "learning_rate": 2.6051724137931037e-05, | |
| "loss": 1.4794, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 1.5, | |
| "grad_norm": 2.133786916732788, | |
| "learning_rate": 2.587931034482759e-05, | |
| "loss": 1.4226, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 1.51, | |
| "grad_norm": 1.7223315238952637, | |
| "learning_rate": 2.570689655172414e-05, | |
| "loss": 1.3799, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 1.52, | |
| "grad_norm": 2.187749147415161, | |
| "learning_rate": 2.5534482758620693e-05, | |
| "loss": 1.3415, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 1.53, | |
| "grad_norm": 1.8895548582077026, | |
| "learning_rate": 2.5362068965517246e-05, | |
| "loss": 1.2736, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 1.54, | |
| "grad_norm": 2.2825796604156494, | |
| "learning_rate": 2.5189655172413794e-05, | |
| "loss": 1.4374, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 1.55, | |
| "grad_norm": 1.5137097835540771, | |
| "learning_rate": 2.5017241379310346e-05, | |
| "loss": 1.1751, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 1.56, | |
| "grad_norm": 1.8762792348861694, | |
| "learning_rate": 2.48448275862069e-05, | |
| "loss": 1.2994, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 1.5699999999999998, | |
| "grad_norm": 1.4601935148239136, | |
| "learning_rate": 2.467241379310345e-05, | |
| "loss": 1.6414, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 1.58, | |
| "grad_norm": 1.7618972063064575, | |
| "learning_rate": 2.45e-05, | |
| "loss": 1.3107, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 1.5899999999999999, | |
| "grad_norm": 3.77205753326416, | |
| "learning_rate": 2.432758620689655e-05, | |
| "loss": 1.4596, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 1.6, | |
| "grad_norm": 1.3197537660598755, | |
| "learning_rate": 2.4155172413793103e-05, | |
| "loss": 1.2752, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.6099999999999999, | |
| "grad_norm": 1.2479321956634521, | |
| "learning_rate": 2.398275862068966e-05, | |
| "loss": 1.2671, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 1.62, | |
| "grad_norm": 1.6044553518295288, | |
| "learning_rate": 2.3810344827586208e-05, | |
| "loss": 1.4017, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 1.63, | |
| "grad_norm": 1.746628999710083, | |
| "learning_rate": 2.363793103448276e-05, | |
| "loss": 1.1989, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 1.6400000000000001, | |
| "grad_norm": 1.4533942937850952, | |
| "learning_rate": 2.3465517241379312e-05, | |
| "loss": 1.3846, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 1.65, | |
| "grad_norm": 1.4432073831558228, | |
| "learning_rate": 2.3293103448275864e-05, | |
| "loss": 1.385, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 1.6600000000000001, | |
| "grad_norm": 1.7605477571487427, | |
| "learning_rate": 2.3120689655172416e-05, | |
| "loss": 1.3516, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 1.67, | |
| "grad_norm": 2.024019718170166, | |
| "learning_rate": 2.2948275862068965e-05, | |
| "loss": 1.2414, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 1.6800000000000002, | |
| "grad_norm": 1.9421080350875854, | |
| "learning_rate": 2.2775862068965517e-05, | |
| "loss": 1.3778, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 1.69, | |
| "grad_norm": 2.7231059074401855, | |
| "learning_rate": 2.2603448275862072e-05, | |
| "loss": 1.3709, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 1.7, | |
| "grad_norm": 4.394734859466553, | |
| "learning_rate": 2.2431034482758624e-05, | |
| "loss": 1.0418, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 1.71, | |
| "grad_norm": 1.7055786848068237, | |
| "learning_rate": 2.2258620689655173e-05, | |
| "loss": 1.3787, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 1.72, | |
| "grad_norm": 2.871880292892456, | |
| "learning_rate": 2.2086206896551725e-05, | |
| "loss": 1.2104, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 1.73, | |
| "grad_norm": 1.9549953937530518, | |
| "learning_rate": 2.1913793103448277e-05, | |
| "loss": 1.3271, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 1.74, | |
| "grad_norm": 1.724934458732605, | |
| "learning_rate": 2.174137931034483e-05, | |
| "loss": 1.1641, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 1.75, | |
| "grad_norm": 1.8861603736877441, | |
| "learning_rate": 2.1568965517241378e-05, | |
| "loss": 1.3302, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 1.76, | |
| "grad_norm": 1.9307113885879517, | |
| "learning_rate": 2.139655172413793e-05, | |
| "loss": 1.4253, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.77, | |
| "grad_norm": 1.860857605934143, | |
| "learning_rate": 2.1224137931034486e-05, | |
| "loss": 1.4241, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 1.78, | |
| "grad_norm": 2.8345351219177246, | |
| "learning_rate": 2.1051724137931038e-05, | |
| "loss": 1.2214, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 1.79, | |
| "grad_norm": 1.1715213060379028, | |
| "learning_rate": 2.0879310344827586e-05, | |
| "loss": 1.3256, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 1.8, | |
| "grad_norm": 1.8184287548065186, | |
| "learning_rate": 2.070689655172414e-05, | |
| "loss": 1.1173, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 1.81, | |
| "grad_norm": 1.9528814554214478, | |
| "learning_rate": 2.053448275862069e-05, | |
| "loss": 1.1872, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 1.8199999999999998, | |
| "grad_norm": 2.1913580894470215, | |
| "learning_rate": 2.0362068965517243e-05, | |
| "loss": 1.2677, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 1.83, | |
| "grad_norm": 1.9390571117401123, | |
| "learning_rate": 2.0189655172413795e-05, | |
| "loss": 1.2469, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 1.8399999999999999, | |
| "grad_norm": 1.8685649633407593, | |
| "learning_rate": 2.0017241379310343e-05, | |
| "loss": 1.2809, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 1.85, | |
| "grad_norm": 2.4311530590057373, | |
| "learning_rate": 1.98448275862069e-05, | |
| "loss": 1.2982, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 1.8599999999999999, | |
| "grad_norm": 1.5639426708221436, | |
| "learning_rate": 1.967241379310345e-05, | |
| "loss": 1.2996, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 1.87, | |
| "grad_norm": 1.2009373903274536, | |
| "learning_rate": 1.9500000000000003e-05, | |
| "loss": 1.2812, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 1.88, | |
| "grad_norm": 1.017490267753601, | |
| "learning_rate": 1.9327586206896552e-05, | |
| "loss": 1.1055, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 1.8900000000000001, | |
| "grad_norm": 2.3106815814971924, | |
| "learning_rate": 1.9155172413793104e-05, | |
| "loss": 1.2474, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 1.9, | |
| "grad_norm": 1.609084129333496, | |
| "learning_rate": 1.8982758620689656e-05, | |
| "loss": 1.1917, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 1.9100000000000001, | |
| "grad_norm": 1.7436444759368896, | |
| "learning_rate": 1.8810344827586208e-05, | |
| "loss": 1.3704, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 1.92, | |
| "grad_norm": 1.7709165811538696, | |
| "learning_rate": 1.863793103448276e-05, | |
| "loss": 1.3904, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.9300000000000002, | |
| "grad_norm": 1.775133490562439, | |
| "learning_rate": 1.8465517241379312e-05, | |
| "loss": 1.2368, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 1.94, | |
| "grad_norm": 2.224012613296509, | |
| "learning_rate": 1.8293103448275864e-05, | |
| "loss": 1.1834, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 1.95, | |
| "grad_norm": 1.8021095991134644, | |
| "learning_rate": 1.8120689655172416e-05, | |
| "loss": 1.2385, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 1.96, | |
| "grad_norm": 2.179215431213379, | |
| "learning_rate": 1.7948275862068965e-05, | |
| "loss": 1.4039, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 1.97, | |
| "grad_norm": 1.850145697593689, | |
| "learning_rate": 1.7775862068965517e-05, | |
| "loss": 1.1953, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 1.98, | |
| "grad_norm": 1.6872076988220215, | |
| "learning_rate": 1.760344827586207e-05, | |
| "loss": 1.2247, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 1.99, | |
| "grad_norm": 2.2350351810455322, | |
| "learning_rate": 1.743103448275862e-05, | |
| "loss": 1.0938, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 1.5278306007385254, | |
| "learning_rate": 1.7258620689655173e-05, | |
| "loss": 1.216, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 2.01, | |
| "grad_norm": 1.9336029291152954, | |
| "learning_rate": 1.7086206896551722e-05, | |
| "loss": 1.1958, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 2.02, | |
| "grad_norm": 1.2718322277069092, | |
| "learning_rate": 1.6913793103448278e-05, | |
| "loss": 1.047, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 2.03, | |
| "grad_norm": 1.8943592309951782, | |
| "learning_rate": 1.674137931034483e-05, | |
| "loss": 1.3197, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 2.04, | |
| "grad_norm": 1.7184983491897583, | |
| "learning_rate": 1.6568965517241382e-05, | |
| "loss": 1.1611, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 2.05, | |
| "grad_norm": 2.1850178241729736, | |
| "learning_rate": 1.639655172413793e-05, | |
| "loss": 1.2405, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 2.06, | |
| "grad_norm": 1.6850008964538574, | |
| "learning_rate": 1.6224137931034483e-05, | |
| "loss": 1.119, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 2.07, | |
| "grad_norm": 1.895009160041809, | |
| "learning_rate": 1.6051724137931035e-05, | |
| "loss": 1.3074, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 2.08, | |
| "grad_norm": 2.0490798950195312, | |
| "learning_rate": 1.5879310344827587e-05, | |
| "loss": 1.0361, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 2.09, | |
| "grad_norm": 2.465575695037842, | |
| "learning_rate": 1.570689655172414e-05, | |
| "loss": 1.1015, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 2.1, | |
| "grad_norm": 1.3066885471343994, | |
| "learning_rate": 1.553448275862069e-05, | |
| "loss": 1.1267, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 2.11, | |
| "grad_norm": 1.7399309873580933, | |
| "learning_rate": 1.5362068965517243e-05, | |
| "loss": 1.203, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 2.12, | |
| "grad_norm": 2.804007053375244, | |
| "learning_rate": 1.5189655172413793e-05, | |
| "loss": 1.0919, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 2.13, | |
| "grad_norm": 1.44864022731781, | |
| "learning_rate": 1.5017241379310346e-05, | |
| "loss": 1.4365, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 2.14, | |
| "grad_norm": 1.5640863180160522, | |
| "learning_rate": 1.4844827586206898e-05, | |
| "loss": 1.1694, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 2.15, | |
| "grad_norm": 1.847272276878357, | |
| "learning_rate": 1.4672413793103448e-05, | |
| "loss": 1.1713, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 2.16, | |
| "grad_norm": 1.4313805103302002, | |
| "learning_rate": 1.45e-05, | |
| "loss": 1.108, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 2.17, | |
| "grad_norm": 1.906491994857788, | |
| "learning_rate": 1.432758620689655e-05, | |
| "loss": 1.2404, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 2.18, | |
| "grad_norm": 2.265270948410034, | |
| "learning_rate": 1.4155172413793106e-05, | |
| "loss": 1.1744, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 2.19, | |
| "grad_norm": 1.927332878112793, | |
| "learning_rate": 1.3982758620689656e-05, | |
| "loss": 1.1907, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 2.2, | |
| "grad_norm": 1.7866955995559692, | |
| "learning_rate": 1.3810344827586208e-05, | |
| "loss": 1.1983, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 2.21, | |
| "grad_norm": 1.738226056098938, | |
| "learning_rate": 1.3637931034482759e-05, | |
| "loss": 1.2034, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 2.22, | |
| "grad_norm": 1.8509528636932373, | |
| "learning_rate": 1.3465517241379311e-05, | |
| "loss": 1.2246, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 2.23, | |
| "grad_norm": 2.776275634765625, | |
| "learning_rate": 1.3293103448275861e-05, | |
| "loss": 1.0615, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 2.24, | |
| "grad_norm": 1.994899868965149, | |
| "learning_rate": 1.3120689655172413e-05, | |
| "loss": 1.0704, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 2.25, | |
| "grad_norm": 1.3992385864257812, | |
| "learning_rate": 1.2948275862068966e-05, | |
| "loss": 1.1459, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 2.26, | |
| "grad_norm": 1.4067262411117554, | |
| "learning_rate": 1.277586206896552e-05, | |
| "loss": 1.104, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 2.27, | |
| "grad_norm": 2.0203895568847656, | |
| "learning_rate": 1.260344827586207e-05, | |
| "loss": 1.2588, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 2.2800000000000002, | |
| "grad_norm": 2.0324246883392334, | |
| "learning_rate": 1.2431034482758622e-05, | |
| "loss": 1.2133, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 2.29, | |
| "grad_norm": 1.5362598896026611, | |
| "learning_rate": 1.2258620689655174e-05, | |
| "loss": 1.2062, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 2.3, | |
| "grad_norm": 1.9725372791290283, | |
| "learning_rate": 1.2086206896551724e-05, | |
| "loss": 1.2211, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 2.31, | |
| "grad_norm": 2.6382522583007812, | |
| "learning_rate": 1.1913793103448276e-05, | |
| "loss": 1.3856, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 2.32, | |
| "grad_norm": 0.9707180857658386, | |
| "learning_rate": 1.1741379310344828e-05, | |
| "loss": 1.1416, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 2.33, | |
| "grad_norm": 2.1676788330078125, | |
| "learning_rate": 1.156896551724138e-05, | |
| "loss": 1.2297, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 2.34, | |
| "grad_norm": 1.6406278610229492, | |
| "learning_rate": 1.1396551724137931e-05, | |
| "loss": 0.9923, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 2.35, | |
| "grad_norm": 1.8479793071746826, | |
| "learning_rate": 1.1224137931034483e-05, | |
| "loss": 1.0925, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 2.36, | |
| "grad_norm": 1.7905402183532715, | |
| "learning_rate": 1.1051724137931035e-05, | |
| "loss": 1.2361, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 2.37, | |
| "grad_norm": 2.006901979446411, | |
| "learning_rate": 1.0879310344827587e-05, | |
| "loss": 1.144, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 2.38, | |
| "grad_norm": 1.4485629796981812, | |
| "learning_rate": 1.0706896551724138e-05, | |
| "loss": 1.1374, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 2.39, | |
| "grad_norm": 1.953250765800476, | |
| "learning_rate": 1.053448275862069e-05, | |
| "loss": 1.222, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 2.4, | |
| "grad_norm": 1.4904489517211914, | |
| "learning_rate": 1.0362068965517242e-05, | |
| "loss": 1.1942, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 2.41, | |
| "grad_norm": 1.9679901599884033, | |
| "learning_rate": 1.0189655172413794e-05, | |
| "loss": 1.2024, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 2.42, | |
| "grad_norm": 1.6501275300979614, | |
| "learning_rate": 1.0017241379310344e-05, | |
| "loss": 1.2754, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 2.43, | |
| "grad_norm": 1.617769479751587, | |
| "learning_rate": 9.844827586206896e-06, | |
| "loss": 1.2086, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 2.44, | |
| "grad_norm": 1.9600815773010254, | |
| "learning_rate": 9.672413793103448e-06, | |
| "loss": 1.1549, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 2.45, | |
| "grad_norm": 1.5949968099594116, | |
| "learning_rate": 9.5e-06, | |
| "loss": 1.2698, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 2.46, | |
| "grad_norm": 1.2073543071746826, | |
| "learning_rate": 9.327586206896553e-06, | |
| "loss": 1.1885, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 2.4699999999999998, | |
| "grad_norm": 1.0354326963424683, | |
| "learning_rate": 9.155172413793103e-06, | |
| "loss": 1.2975, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 2.48, | |
| "grad_norm": 4.918059825897217, | |
| "learning_rate": 8.982758620689657e-06, | |
| "loss": 1.161, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 2.49, | |
| "grad_norm": 2.176772117614746, | |
| "learning_rate": 8.810344827586207e-06, | |
| "loss": 1.1096, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 1.9638526439666748, | |
| "learning_rate": 8.63793103448276e-06, | |
| "loss": 1.1893, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 2.51, | |
| "grad_norm": 1.3812869787216187, | |
| "learning_rate": 8.46551724137931e-06, | |
| "loss": 1.1958, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 2.52, | |
| "grad_norm": 2.1832292079925537, | |
| "learning_rate": 8.293103448275863e-06, | |
| "loss": 1.1582, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 2.5300000000000002, | |
| "grad_norm": 1.416548728942871, | |
| "learning_rate": 8.120689655172414e-06, | |
| "loss": 0.9992, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 2.54, | |
| "grad_norm": 1.465884804725647, | |
| "learning_rate": 7.948275862068966e-06, | |
| "loss": 1.1773, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 2.55, | |
| "grad_norm": 1.754629373550415, | |
| "learning_rate": 7.775862068965516e-06, | |
| "loss": 1.2893, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 2.56, | |
| "grad_norm": 1.3299590349197388, | |
| "learning_rate": 7.60344827586207e-06, | |
| "loss": 1.1175, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 2.57, | |
| "grad_norm": 1.6379165649414062, | |
| "learning_rate": 7.431034482758621e-06, | |
| "loss": 1.1522, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 2.58, | |
| "grad_norm": 1.038396954536438, | |
| "learning_rate": 7.258620689655173e-06, | |
| "loss": 1.1105, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 2.59, | |
| "grad_norm": 3.7200427055358887, | |
| "learning_rate": 7.086206896551724e-06, | |
| "loss": 1.2392, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 2.6, | |
| "grad_norm": 1.996336817741394, | |
| "learning_rate": 6.913793103448277e-06, | |
| "loss": 1.2586, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 2.61, | |
| "grad_norm": 2.2930266857147217, | |
| "learning_rate": 6.741379310344828e-06, | |
| "loss": 1.2299, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 2.62, | |
| "grad_norm": 1.5807350873947144, | |
| "learning_rate": 6.568965517241379e-06, | |
| "loss": 1.1705, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 2.63, | |
| "grad_norm": 2.1657156944274902, | |
| "learning_rate": 6.3965517241379305e-06, | |
| "loss": 1.008, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 2.64, | |
| "grad_norm": 2.036822557449341, | |
| "learning_rate": 6.224137931034483e-06, | |
| "loss": 1.36, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 2.65, | |
| "grad_norm": 1.8225266933441162, | |
| "learning_rate": 6.051724137931035e-06, | |
| "loss": 1.0531, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 2.66, | |
| "grad_norm": 1.9631621837615967, | |
| "learning_rate": 5.879310344827586e-06, | |
| "loss": 1.1185, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 2.67, | |
| "grad_norm": 1.3313068151474, | |
| "learning_rate": 5.706896551724138e-06, | |
| "loss": 1.0997, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 2.68, | |
| "grad_norm": 1.7591108083724976, | |
| "learning_rate": 5.53448275862069e-06, | |
| "loss": 1.2264, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 2.69, | |
| "grad_norm": 1.669311761856079, | |
| "learning_rate": 5.362068965517242e-06, | |
| "loss": 1.0538, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 2.7, | |
| "grad_norm": 1.9697110652923584, | |
| "learning_rate": 5.1896551724137934e-06, | |
| "loss": 1.2217, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 2.71, | |
| "grad_norm": 1.5783755779266357, | |
| "learning_rate": 5.0172413793103455e-06, | |
| "loss": 1.1666, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 2.7199999999999998, | |
| "grad_norm": 2.451660394668579, | |
| "learning_rate": 4.844827586206897e-06, | |
| "loss": 1.2254, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 2.73, | |
| "grad_norm": 1.6855318546295166, | |
| "learning_rate": 4.672413793103449e-06, | |
| "loss": 1.1644, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 2.74, | |
| "grad_norm": 1.955734372138977, | |
| "learning_rate": 4.5e-06, | |
| "loss": 1.2558, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 2.75, | |
| "grad_norm": 1.5705198049545288, | |
| "learning_rate": 4.327586206896552e-06, | |
| "loss": 1.3614, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 2.76, | |
| "grad_norm": 1.878604769706726, | |
| "learning_rate": 4.1551724137931034e-06, | |
| "loss": 1.1227, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 2.77, | |
| "grad_norm": 1.3818405866622925, | |
| "learning_rate": 3.9827586206896555e-06, | |
| "loss": 1.0261, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 2.7800000000000002, | |
| "grad_norm": 1.960067629814148, | |
| "learning_rate": 3.8103448275862068e-06, | |
| "loss": 1.2194, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 2.79, | |
| "grad_norm": 1.6252312660217285, | |
| "learning_rate": 3.637931034482759e-06, | |
| "loss": 1.0916, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 2.8, | |
| "grad_norm": 2.1915979385375977, | |
| "learning_rate": 3.4655172413793105e-06, | |
| "loss": 0.9632, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 2.81, | |
| "grad_norm": 1.7106447219848633, | |
| "learning_rate": 3.2931034482758626e-06, | |
| "loss": 1.1671, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 2.82, | |
| "grad_norm": 1.840908169746399, | |
| "learning_rate": 3.120689655172414e-06, | |
| "loss": 1.1216, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 2.83, | |
| "grad_norm": 1.71086585521698, | |
| "learning_rate": 2.9482758620689655e-06, | |
| "loss": 1.1521, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 2.84, | |
| "grad_norm": 1.500496745109558, | |
| "learning_rate": 2.775862068965517e-06, | |
| "loss": 1.1021, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 2.85, | |
| "grad_norm": 2.563215494155884, | |
| "learning_rate": 2.603448275862069e-06, | |
| "loss": 1.0535, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 2.86, | |
| "grad_norm": 1.2782319784164429, | |
| "learning_rate": 2.4310344827586205e-06, | |
| "loss": 1.1942, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 2.87, | |
| "grad_norm": 1.2960996627807617, | |
| "learning_rate": 2.258620689655172e-06, | |
| "loss": 1.1178, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 2.88, | |
| "grad_norm": 1.6457912921905518, | |
| "learning_rate": 2.0862068965517243e-06, | |
| "loss": 1.0038, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 2.89, | |
| "grad_norm": 1.8096386194229126, | |
| "learning_rate": 1.913793103448276e-06, | |
| "loss": 1.142, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 2.9, | |
| "grad_norm": 1.7525629997253418, | |
| "learning_rate": 1.7413793103448276e-06, | |
| "loss": 1.254, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 2.91, | |
| "grad_norm": 2.7859292030334473, | |
| "learning_rate": 1.5689655172413793e-06, | |
| "loss": 1.4235, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 2.92, | |
| "grad_norm": 1.4601234197616577, | |
| "learning_rate": 1.3965517241379312e-06, | |
| "loss": 1.0771, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 2.93, | |
| "grad_norm": 1.9630273580551147, | |
| "learning_rate": 1.2241379310344828e-06, | |
| "loss": 1.2312, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 2.94, | |
| "grad_norm": 1.5725574493408203, | |
| "learning_rate": 1.0517241379310345e-06, | |
| "loss": 1.026, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 2.95, | |
| "grad_norm": 1.9632741212844849, | |
| "learning_rate": 8.793103448275863e-07, | |
| "loss": 1.1516, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 2.96, | |
| "grad_norm": 2.4151031970977783, | |
| "learning_rate": 7.068965517241379e-07, | |
| "loss": 1.2166, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 2.9699999999999998, | |
| "grad_norm": 1.4842921495437622, | |
| "learning_rate": 5.344827586206897e-07, | |
| "loss": 1.1279, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 2.98, | |
| "grad_norm": 2.3842453956604004, | |
| "learning_rate": 3.6206896551724143e-07, | |
| "loss": 1.0998, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 2.99, | |
| "grad_norm": 1.8654992580413818, | |
| "learning_rate": 1.8965517241379312e-07, | |
| "loss": 1.2559, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 1.5230703353881836, | |
| "learning_rate": 1.7241379310344825e-08, | |
| "loss": 1.1137, | |
| "step": 3000 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 3000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 3, | |
| "save_steps": 100, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 2786102083584000.0, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |