| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.048760499897562, |
| "eval_steps": 500, |
| "global_step": 20000, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.001024380249948781, |
| "grad_norm": 8.322718620300293, |
| "learning_rate": 1.8e-07, |
| "loss": 1.6653, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.002048760499897562, |
| "grad_norm": 9.392577171325684, |
| "learning_rate": 3.8e-07, |
| "loss": 1.7655, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.003073140749846343, |
| "grad_norm": 12.280142784118652, |
| "learning_rate": 5.800000000000001e-07, |
| "loss": 1.7285, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.004097520999795124, |
| "grad_norm": 8.562142372131348, |
| "learning_rate": 7.799999999999999e-07, |
| "loss": 1.6348, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.005121901249743905, |
| "grad_norm": 9.944829940795898, |
| "learning_rate": 9.8e-07, |
| "loss": 1.5127, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.006146281499692686, |
| "grad_norm": 5.845076084136963, |
| "learning_rate": 1.18e-06, |
| "loss": 1.5735, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.007170661749641467, |
| "grad_norm": 8.555084228515625, |
| "learning_rate": 1.38e-06, |
| "loss": 1.3493, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.008195041999590248, |
| "grad_norm": 6.1190056800842285, |
| "learning_rate": 1.5800000000000001e-06, |
| "loss": 1.1216, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.00921942224953903, |
| "grad_norm": 3.0822441577911377, |
| "learning_rate": 1.7800000000000001e-06, |
| "loss": 0.8451, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.01024380249948781, |
| "grad_norm": 3.0187950134277344, |
| "learning_rate": 1.98e-06, |
| "loss": 0.6822, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.01126818274943659, |
| "grad_norm": 2.756537437438965, |
| "learning_rate": 2.1800000000000003e-06, |
| "loss": 0.5474, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.012292562999385371, |
| "grad_norm": 2.0504415035247803, |
| "learning_rate": 2.38e-06, |
| "loss": 0.4562, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.013316943249334154, |
| "grad_norm": 1.436387538909912, |
| "learning_rate": 2.58e-06, |
| "loss": 0.3522, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.014341323499282934, |
| "grad_norm": 1.416121006011963, |
| "learning_rate": 2.78e-06, |
| "loss": 0.2875, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.015365703749231715, |
| "grad_norm": 1.2155678272247314, |
| "learning_rate": 2.98e-06, |
| "loss": 0.2511, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.016390083999180495, |
| "grad_norm": 1.4057164192199707, |
| "learning_rate": 3.18e-06, |
| "loss": 0.2083, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.017414464249129277, |
| "grad_norm": 0.6794387102127075, |
| "learning_rate": 3.38e-06, |
| "loss": 0.166, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.01843884449907806, |
| "grad_norm": 0.8129542469978333, |
| "learning_rate": 3.58e-06, |
| "loss": 0.1392, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.01946322474902684, |
| "grad_norm": 0.5477985143661499, |
| "learning_rate": 3.7800000000000002e-06, |
| "loss": 0.1264, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.02048760499897562, |
| "grad_norm": 0.621268093585968, |
| "learning_rate": 3.98e-06, |
| "loss": 0.1131, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.0215119852489244, |
| "grad_norm": 0.6767550706863403, |
| "learning_rate": 4.18e-06, |
| "loss": 0.1021, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.02253636549887318, |
| "grad_norm": 0.48897942900657654, |
| "learning_rate": 4.3799999999999996e-06, |
| "loss": 0.0975, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.023560745748821964, |
| "grad_norm": 0.670199990272522, |
| "learning_rate": 4.58e-06, |
| "loss": 0.0963, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.024585125998770743, |
| "grad_norm": 0.845915675163269, |
| "learning_rate": 4.78e-06, |
| "loss": 0.0893, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.025609506248719525, |
| "grad_norm": 0.5154263973236084, |
| "learning_rate": 4.980000000000001e-06, |
| "loss": 0.0828, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.026633886498668307, |
| "grad_norm": 0.446519136428833, |
| "learning_rate": 5.18e-06, |
| "loss": 0.0766, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.027658266748617086, |
| "grad_norm": 0.4152374267578125, |
| "learning_rate": 5.38e-06, |
| "loss": 0.0821, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.028682646998565868, |
| "grad_norm": 0.33393970131874084, |
| "learning_rate": 5.58e-06, |
| "loss": 0.0686, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.02970702724851465, |
| "grad_norm": 0.43520069122314453, |
| "learning_rate": 5.7800000000000006e-06, |
| "loss": 0.0653, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.03073140749846343, |
| "grad_norm": 0.5556287169456482, |
| "learning_rate": 5.98e-06, |
| "loss": 0.0628, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.03175578774841221, |
| "grad_norm": 0.599671483039856, |
| "learning_rate": 6.18e-06, |
| "loss": 0.0571, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.03278016799836099, |
| "grad_norm": 0.4356689155101776, |
| "learning_rate": 6.38e-06, |
| "loss": 0.0598, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.03380454824830977, |
| "grad_norm": 0.25835078954696655, |
| "learning_rate": 6.58e-06, |
| "loss": 0.0604, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.034828928498258555, |
| "grad_norm": 0.6627889275550842, |
| "learning_rate": 6.78e-06, |
| "loss": 0.0513, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.03585330874820734, |
| "grad_norm": 0.3467154800891876, |
| "learning_rate": 6.98e-06, |
| "loss": 0.053, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.03687768899815612, |
| "grad_norm": 0.8043546080589294, |
| "learning_rate": 7.180000000000001e-06, |
| "loss": 0.0636, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.037902069248104894, |
| "grad_norm": 0.38153547048568726, |
| "learning_rate": 7.3800000000000005e-06, |
| "loss": 0.0526, |
| "step": 370 |
| }, |
| { |
| "epoch": 0.03892644949805368, |
| "grad_norm": 0.35963737964630127, |
| "learning_rate": 7.5799999999999994e-06, |
| "loss": 0.0599, |
| "step": 380 |
| }, |
| { |
| "epoch": 0.03995082974800246, |
| "grad_norm": 0.26123321056365967, |
| "learning_rate": 7.780000000000002e-06, |
| "loss": 0.0511, |
| "step": 390 |
| }, |
| { |
| "epoch": 0.04097520999795124, |
| "grad_norm": 0.29133498668670654, |
| "learning_rate": 7.98e-06, |
| "loss": 0.0476, |
| "step": 400 |
| }, |
| { |
| "epoch": 0.04199959024790002, |
| "grad_norm": 0.41868746280670166, |
| "learning_rate": 8.18e-06, |
| "loss": 0.0402, |
| "step": 410 |
| }, |
| { |
| "epoch": 0.0430239704978488, |
| "grad_norm": 0.6088660955429077, |
| "learning_rate": 8.38e-06, |
| "loss": 0.0466, |
| "step": 420 |
| }, |
| { |
| "epoch": 0.04404835074779758, |
| "grad_norm": 0.5830772519111633, |
| "learning_rate": 8.58e-06, |
| "loss": 0.0499, |
| "step": 430 |
| }, |
| { |
| "epoch": 0.04507273099774636, |
| "grad_norm": 0.6772302389144897, |
| "learning_rate": 8.78e-06, |
| "loss": 0.0456, |
| "step": 440 |
| }, |
| { |
| "epoch": 0.046097111247695145, |
| "grad_norm": 0.69731605052948, |
| "learning_rate": 8.98e-06, |
| "loss": 0.0502, |
| "step": 450 |
| }, |
| { |
| "epoch": 0.04712149149764393, |
| "grad_norm": 0.590331494808197, |
| "learning_rate": 9.18e-06, |
| "loss": 0.0499, |
| "step": 460 |
| }, |
| { |
| "epoch": 0.04814587174759271, |
| "grad_norm": 0.3974471092224121, |
| "learning_rate": 9.38e-06, |
| "loss": 0.0449, |
| "step": 470 |
| }, |
| { |
| "epoch": 0.049170251997541485, |
| "grad_norm": 0.49228665232658386, |
| "learning_rate": 9.580000000000001e-06, |
| "loss": 0.0475, |
| "step": 480 |
| }, |
| { |
| "epoch": 0.05019463224749027, |
| "grad_norm": 0.503321647644043, |
| "learning_rate": 9.780000000000001e-06, |
| "loss": 0.0401, |
| "step": 490 |
| }, |
| { |
| "epoch": 0.05121901249743905, |
| "grad_norm": 0.6616262793540955, |
| "learning_rate": 9.980000000000001e-06, |
| "loss": 0.0395, |
| "step": 500 |
| }, |
| { |
| "epoch": 0.05224339274738783, |
| "grad_norm": 0.6191039681434631, |
| "learning_rate": 1.0179999999999999e-05, |
| "loss": 0.0428, |
| "step": 510 |
| }, |
| { |
| "epoch": 0.053267772997336614, |
| "grad_norm": 0.33810850977897644, |
| "learning_rate": 1.0379999999999999e-05, |
| "loss": 0.0378, |
| "step": 520 |
| }, |
| { |
| "epoch": 0.05429215324728539, |
| "grad_norm": 0.5174738764762878, |
| "learning_rate": 1.058e-05, |
| "loss": 0.0422, |
| "step": 530 |
| }, |
| { |
| "epoch": 0.05531653349723417, |
| "grad_norm": 0.2611842751502991, |
| "learning_rate": 1.078e-05, |
| "loss": 0.0375, |
| "step": 540 |
| }, |
| { |
| "epoch": 0.056340913747182954, |
| "grad_norm": 0.28647738695144653, |
| "learning_rate": 1.098e-05, |
| "loss": 0.0371, |
| "step": 550 |
| }, |
| { |
| "epoch": 0.057365293997131736, |
| "grad_norm": 0.40737223625183105, |
| "learning_rate": 1.118e-05, |
| "loss": 0.0353, |
| "step": 560 |
| }, |
| { |
| "epoch": 0.05838967424708052, |
| "grad_norm": 0.44597023725509644, |
| "learning_rate": 1.1380000000000001e-05, |
| "loss": 0.043, |
| "step": 570 |
| }, |
| { |
| "epoch": 0.0594140544970293, |
| "grad_norm": 0.5923913717269897, |
| "learning_rate": 1.1580000000000001e-05, |
| "loss": 0.0349, |
| "step": 580 |
| }, |
| { |
| "epoch": 0.060438434746978076, |
| "grad_norm": 0.4841652512550354, |
| "learning_rate": 1.178e-05, |
| "loss": 0.0358, |
| "step": 590 |
| }, |
| { |
| "epoch": 0.06146281499692686, |
| "grad_norm": 0.2946516275405884, |
| "learning_rate": 1.198e-05, |
| "loss": 0.0347, |
| "step": 600 |
| }, |
| { |
| "epoch": 0.06248719524687564, |
| "grad_norm": 0.8168693780899048, |
| "learning_rate": 1.2180000000000002e-05, |
| "loss": 0.0423, |
| "step": 610 |
| }, |
| { |
| "epoch": 0.06351157549682442, |
| "grad_norm": 0.2948760390281677, |
| "learning_rate": 1.238e-05, |
| "loss": 0.0353, |
| "step": 620 |
| }, |
| { |
| "epoch": 0.0645359557467732, |
| "grad_norm": 0.4611808955669403, |
| "learning_rate": 1.258e-05, |
| "loss": 0.036, |
| "step": 630 |
| }, |
| { |
| "epoch": 0.06556033599672198, |
| "grad_norm": 0.32586854696273804, |
| "learning_rate": 1.278e-05, |
| "loss": 0.037, |
| "step": 640 |
| }, |
| { |
| "epoch": 0.06658471624667077, |
| "grad_norm": 0.4598027765750885, |
| "learning_rate": 1.298e-05, |
| "loss": 0.0353, |
| "step": 650 |
| }, |
| { |
| "epoch": 0.06760909649661954, |
| "grad_norm": 0.5787739157676697, |
| "learning_rate": 1.3180000000000001e-05, |
| "loss": 0.0346, |
| "step": 660 |
| }, |
| { |
| "epoch": 0.06863347674656832, |
| "grad_norm": 0.3495701849460602, |
| "learning_rate": 1.338e-05, |
| "loss": 0.0356, |
| "step": 670 |
| }, |
| { |
| "epoch": 0.06965785699651711, |
| "grad_norm": 0.43106862902641296, |
| "learning_rate": 1.358e-05, |
| "loss": 0.0367, |
| "step": 680 |
| }, |
| { |
| "epoch": 0.07068223724646588, |
| "grad_norm": 0.8802304267883301, |
| "learning_rate": 1.378e-05, |
| "loss": 0.0395, |
| "step": 690 |
| }, |
| { |
| "epoch": 0.07170661749641467, |
| "grad_norm": 0.3897860646247864, |
| "learning_rate": 1.3980000000000002e-05, |
| "loss": 0.0352, |
| "step": 700 |
| }, |
| { |
| "epoch": 0.07273099774636345, |
| "grad_norm": 0.4143000543117523, |
| "learning_rate": 1.4180000000000001e-05, |
| "loss": 0.0334, |
| "step": 710 |
| }, |
| { |
| "epoch": 0.07375537799631224, |
| "grad_norm": 0.7109991312026978, |
| "learning_rate": 1.438e-05, |
| "loss": 0.0332, |
| "step": 720 |
| }, |
| { |
| "epoch": 0.07477975824626101, |
| "grad_norm": 0.30344876646995544, |
| "learning_rate": 1.458e-05, |
| "loss": 0.0294, |
| "step": 730 |
| }, |
| { |
| "epoch": 0.07580413849620979, |
| "grad_norm": 0.3466661870479584, |
| "learning_rate": 1.4779999999999999e-05, |
| "loss": 0.0285, |
| "step": 740 |
| }, |
| { |
| "epoch": 0.07682851874615858, |
| "grad_norm": 0.22630204260349274, |
| "learning_rate": 1.498e-05, |
| "loss": 0.0283, |
| "step": 750 |
| }, |
| { |
| "epoch": 0.07785289899610735, |
| "grad_norm": 0.2611643671989441, |
| "learning_rate": 1.518e-05, |
| "loss": 0.0307, |
| "step": 760 |
| }, |
| { |
| "epoch": 0.07887727924605614, |
| "grad_norm": 0.6717830300331116, |
| "learning_rate": 1.5380000000000002e-05, |
| "loss": 0.033, |
| "step": 770 |
| }, |
| { |
| "epoch": 0.07990165949600492, |
| "grad_norm": 0.385221391916275, |
| "learning_rate": 1.558e-05, |
| "loss": 0.0334, |
| "step": 780 |
| }, |
| { |
| "epoch": 0.08092603974595369, |
| "grad_norm": 1.4035415649414062, |
| "learning_rate": 1.578e-05, |
| "loss": 0.0354, |
| "step": 790 |
| }, |
| { |
| "epoch": 0.08195041999590248, |
| "grad_norm": 0.33571937680244446, |
| "learning_rate": 1.598e-05, |
| "loss": 0.0352, |
| "step": 800 |
| }, |
| { |
| "epoch": 0.08297480024585126, |
| "grad_norm": 0.37995073199272156, |
| "learning_rate": 1.618e-05, |
| "loss": 0.0316, |
| "step": 810 |
| }, |
| { |
| "epoch": 0.08399918049580005, |
| "grad_norm": 0.35028374195098877, |
| "learning_rate": 1.6380000000000002e-05, |
| "loss": 0.0328, |
| "step": 820 |
| }, |
| { |
| "epoch": 0.08502356074574882, |
| "grad_norm": 0.257723867893219, |
| "learning_rate": 1.658e-05, |
| "loss": 0.0346, |
| "step": 830 |
| }, |
| { |
| "epoch": 0.0860479409956976, |
| "grad_norm": 0.2460213601589203, |
| "learning_rate": 1.6780000000000002e-05, |
| "loss": 0.0294, |
| "step": 840 |
| }, |
| { |
| "epoch": 0.08707232124564639, |
| "grad_norm": 0.36394014954566956, |
| "learning_rate": 1.698e-05, |
| "loss": 0.03, |
| "step": 850 |
| }, |
| { |
| "epoch": 0.08809670149559516, |
| "grad_norm": 0.3701348602771759, |
| "learning_rate": 1.718e-05, |
| "loss": 0.0289, |
| "step": 860 |
| }, |
| { |
| "epoch": 0.08912108174554395, |
| "grad_norm": 0.28825581073760986, |
| "learning_rate": 1.7380000000000003e-05, |
| "loss": 0.0262, |
| "step": 870 |
| }, |
| { |
| "epoch": 0.09014546199549273, |
| "grad_norm": 0.5280503034591675, |
| "learning_rate": 1.758e-05, |
| "loss": 0.0328, |
| "step": 880 |
| }, |
| { |
| "epoch": 0.0911698422454415, |
| "grad_norm": 0.6406586766242981, |
| "learning_rate": 1.778e-05, |
| "loss": 0.0279, |
| "step": 890 |
| }, |
| { |
| "epoch": 0.09219422249539029, |
| "grad_norm": 0.4262707829475403, |
| "learning_rate": 1.798e-05, |
| "loss": 0.0349, |
| "step": 900 |
| }, |
| { |
| "epoch": 0.09321860274533907, |
| "grad_norm": 0.44939371943473816, |
| "learning_rate": 1.818e-05, |
| "loss": 0.0297, |
| "step": 910 |
| }, |
| { |
| "epoch": 0.09424298299528786, |
| "grad_norm": 0.3454805314540863, |
| "learning_rate": 1.838e-05, |
| "loss": 0.0341, |
| "step": 920 |
| }, |
| { |
| "epoch": 0.09526736324523663, |
| "grad_norm": 0.33688879013061523, |
| "learning_rate": 1.8579999999999998e-05, |
| "loss": 0.0267, |
| "step": 930 |
| }, |
| { |
| "epoch": 0.09629174349518542, |
| "grad_norm": 0.3307080566883087, |
| "learning_rate": 1.878e-05, |
| "loss": 0.0317, |
| "step": 940 |
| }, |
| { |
| "epoch": 0.0973161237451342, |
| "grad_norm": 0.4247128963470459, |
| "learning_rate": 1.898e-05, |
| "loss": 0.028, |
| "step": 950 |
| }, |
| { |
| "epoch": 0.09834050399508297, |
| "grad_norm": 0.4177204370498657, |
| "learning_rate": 1.918e-05, |
| "loss": 0.0328, |
| "step": 960 |
| }, |
| { |
| "epoch": 0.09936488424503176, |
| "grad_norm": 0.30320802330970764, |
| "learning_rate": 1.938e-05, |
| "loss": 0.0284, |
| "step": 970 |
| }, |
| { |
| "epoch": 0.10038926449498053, |
| "grad_norm": 0.30202317237854004, |
| "learning_rate": 1.958e-05, |
| "loss": 0.0259, |
| "step": 980 |
| }, |
| { |
| "epoch": 0.10141364474492932, |
| "grad_norm": 0.5330867767333984, |
| "learning_rate": 1.978e-05, |
| "loss": 0.0286, |
| "step": 990 |
| }, |
| { |
| "epoch": 0.1024380249948781, |
| "grad_norm": 0.5143296122550964, |
| "learning_rate": 1.9980000000000002e-05, |
| "loss": 0.03, |
| "step": 1000 |
| }, |
| { |
| "epoch": 0.10346240524482687, |
| "grad_norm": 0.3872562348842621, |
| "learning_rate": 2.018e-05, |
| "loss": 0.0242, |
| "step": 1010 |
| }, |
| { |
| "epoch": 0.10448678549477566, |
| "grad_norm": 0.23155750334262848, |
| "learning_rate": 2.038e-05, |
| "loss": 0.0311, |
| "step": 1020 |
| }, |
| { |
| "epoch": 0.10551116574472444, |
| "grad_norm": 0.2522430121898651, |
| "learning_rate": 2.0580000000000003e-05, |
| "loss": 0.0271, |
| "step": 1030 |
| }, |
| { |
| "epoch": 0.10653554599467323, |
| "grad_norm": 0.3459431231021881, |
| "learning_rate": 2.078e-05, |
| "loss": 0.0269, |
| "step": 1040 |
| }, |
| { |
| "epoch": 0.107559926244622, |
| "grad_norm": 0.5609778165817261, |
| "learning_rate": 2.0980000000000002e-05, |
| "loss": 0.0273, |
| "step": 1050 |
| }, |
| { |
| "epoch": 0.10858430649457078, |
| "grad_norm": 0.38768628239631653, |
| "learning_rate": 2.118e-05, |
| "loss": 0.0324, |
| "step": 1060 |
| }, |
| { |
| "epoch": 0.10960868674451957, |
| "grad_norm": 0.4217023253440857, |
| "learning_rate": 2.1380000000000002e-05, |
| "loss": 0.0231, |
| "step": 1070 |
| }, |
| { |
| "epoch": 0.11063306699446834, |
| "grad_norm": 0.25070101022720337, |
| "learning_rate": 2.1580000000000003e-05, |
| "loss": 0.0257, |
| "step": 1080 |
| }, |
| { |
| "epoch": 0.11165744724441713, |
| "grad_norm": 0.7360861897468567, |
| "learning_rate": 2.178e-05, |
| "loss": 0.028, |
| "step": 1090 |
| }, |
| { |
| "epoch": 0.11268182749436591, |
| "grad_norm": 0.319614440202713, |
| "learning_rate": 2.198e-05, |
| "loss": 0.0233, |
| "step": 1100 |
| }, |
| { |
| "epoch": 0.11370620774431468, |
| "grad_norm": 0.347334086894989, |
| "learning_rate": 2.2179999999999998e-05, |
| "loss": 0.0238, |
| "step": 1110 |
| }, |
| { |
| "epoch": 0.11473058799426347, |
| "grad_norm": 0.24495229125022888, |
| "learning_rate": 2.238e-05, |
| "loss": 0.0228, |
| "step": 1120 |
| }, |
| { |
| "epoch": 0.11575496824421225, |
| "grad_norm": 0.2624426782131195, |
| "learning_rate": 2.258e-05, |
| "loss": 0.0239, |
| "step": 1130 |
| }, |
| { |
| "epoch": 0.11677934849416104, |
| "grad_norm": 0.2570275366306305, |
| "learning_rate": 2.278e-05, |
| "loss": 0.023, |
| "step": 1140 |
| }, |
| { |
| "epoch": 0.11780372874410981, |
| "grad_norm": 0.6030268669128418, |
| "learning_rate": 2.298e-05, |
| "loss": 0.0282, |
| "step": 1150 |
| }, |
| { |
| "epoch": 0.1188281089940586, |
| "grad_norm": 0.3113791048526764, |
| "learning_rate": 2.3179999999999998e-05, |
| "loss": 0.0224, |
| "step": 1160 |
| }, |
| { |
| "epoch": 0.11985248924400738, |
| "grad_norm": 0.39047712087631226, |
| "learning_rate": 2.338e-05, |
| "loss": 0.0261, |
| "step": 1170 |
| }, |
| { |
| "epoch": 0.12087686949395615, |
| "grad_norm": 0.317891389131546, |
| "learning_rate": 2.358e-05, |
| "loss": 0.0216, |
| "step": 1180 |
| }, |
| { |
| "epoch": 0.12190124974390494, |
| "grad_norm": 0.4647437036037445, |
| "learning_rate": 2.378e-05, |
| "loss": 0.0271, |
| "step": 1190 |
| }, |
| { |
| "epoch": 0.12292562999385372, |
| "grad_norm": 0.2541736960411072, |
| "learning_rate": 2.398e-05, |
| "loss": 0.0252, |
| "step": 1200 |
| }, |
| { |
| "epoch": 0.1239500102438025, |
| "grad_norm": 0.2833355665206909, |
| "learning_rate": 2.4180000000000002e-05, |
| "loss": 0.0239, |
| "step": 1210 |
| }, |
| { |
| "epoch": 0.12497439049375128, |
| "grad_norm": 0.3637617230415344, |
| "learning_rate": 2.438e-05, |
| "loss": 0.0279, |
| "step": 1220 |
| }, |
| { |
| "epoch": 0.12599877074370006, |
| "grad_norm": 0.40432628989219666, |
| "learning_rate": 2.4580000000000002e-05, |
| "loss": 0.0253, |
| "step": 1230 |
| }, |
| { |
| "epoch": 0.12702315099364883, |
| "grad_norm": 0.3784823715686798, |
| "learning_rate": 2.478e-05, |
| "loss": 0.0282, |
| "step": 1240 |
| }, |
| { |
| "epoch": 0.12804753124359763, |
| "grad_norm": 0.5457500219345093, |
| "learning_rate": 2.498e-05, |
| "loss": 0.0286, |
| "step": 1250 |
| }, |
| { |
| "epoch": 0.1290719114935464, |
| "grad_norm": 0.33996304869651794, |
| "learning_rate": 2.5180000000000003e-05, |
| "loss": 0.0246, |
| "step": 1260 |
| }, |
| { |
| "epoch": 0.13009629174349518, |
| "grad_norm": 0.4703163206577301, |
| "learning_rate": 2.538e-05, |
| "loss": 0.0293, |
| "step": 1270 |
| }, |
| { |
| "epoch": 0.13112067199344396, |
| "grad_norm": 0.3938589096069336, |
| "learning_rate": 2.5580000000000002e-05, |
| "loss": 0.0278, |
| "step": 1280 |
| }, |
| { |
| "epoch": 0.13214505224339274, |
| "grad_norm": 0.21108706295490265, |
| "learning_rate": 2.578e-05, |
| "loss": 0.0248, |
| "step": 1290 |
| }, |
| { |
| "epoch": 0.13316943249334154, |
| "grad_norm": 0.5333048105239868, |
| "learning_rate": 2.5980000000000002e-05, |
| "loss": 0.0255, |
| "step": 1300 |
| }, |
| { |
| "epoch": 0.1341938127432903, |
| "grad_norm": 0.39306050539016724, |
| "learning_rate": 2.618e-05, |
| "loss": 0.0267, |
| "step": 1310 |
| }, |
| { |
| "epoch": 0.1352181929932391, |
| "grad_norm": 0.32033678889274597, |
| "learning_rate": 2.6379999999999998e-05, |
| "loss": 0.0257, |
| "step": 1320 |
| }, |
| { |
| "epoch": 0.13624257324318786, |
| "grad_norm": 0.3632168769836426, |
| "learning_rate": 2.658e-05, |
| "loss": 0.0257, |
| "step": 1330 |
| }, |
| { |
| "epoch": 0.13726695349313664, |
| "grad_norm": 0.2535551190376282, |
| "learning_rate": 2.678e-05, |
| "loss": 0.0276, |
| "step": 1340 |
| }, |
| { |
| "epoch": 0.13829133374308544, |
| "grad_norm": 0.435642808675766, |
| "learning_rate": 2.698e-05, |
| "loss": 0.0216, |
| "step": 1350 |
| }, |
| { |
| "epoch": 0.13931571399303422, |
| "grad_norm": 0.4528685212135315, |
| "learning_rate": 2.718e-05, |
| "loss": 0.03, |
| "step": 1360 |
| }, |
| { |
| "epoch": 0.140340094242983, |
| "grad_norm": 0.3468475341796875, |
| "learning_rate": 2.738e-05, |
| "loss": 0.0244, |
| "step": 1370 |
| }, |
| { |
| "epoch": 0.14136447449293177, |
| "grad_norm": 0.5749343037605286, |
| "learning_rate": 2.758e-05, |
| "loss": 0.0261, |
| "step": 1380 |
| }, |
| { |
| "epoch": 0.14238885474288054, |
| "grad_norm": 0.5055791139602661, |
| "learning_rate": 2.778e-05, |
| "loss": 0.0235, |
| "step": 1390 |
| }, |
| { |
| "epoch": 0.14341323499282935, |
| "grad_norm": 0.3830702602863312, |
| "learning_rate": 2.798e-05, |
| "loss": 0.0213, |
| "step": 1400 |
| }, |
| { |
| "epoch": 0.14443761524277812, |
| "grad_norm": 0.5464547276496887, |
| "learning_rate": 2.818e-05, |
| "loss": 0.0218, |
| "step": 1410 |
| }, |
| { |
| "epoch": 0.1454619954927269, |
| "grad_norm": 0.5732727646827698, |
| "learning_rate": 2.838e-05, |
| "loss": 0.0302, |
| "step": 1420 |
| }, |
| { |
| "epoch": 0.14648637574267567, |
| "grad_norm": 1.001684308052063, |
| "learning_rate": 2.858e-05, |
| "loss": 0.0289, |
| "step": 1430 |
| }, |
| { |
| "epoch": 0.14751075599262448, |
| "grad_norm": 0.3476460576057434, |
| "learning_rate": 2.8780000000000002e-05, |
| "loss": 0.0254, |
| "step": 1440 |
| }, |
| { |
| "epoch": 0.14853513624257325, |
| "grad_norm": 0.2921610474586487, |
| "learning_rate": 2.898e-05, |
| "loss": 0.0247, |
| "step": 1450 |
| }, |
| { |
| "epoch": 0.14955951649252203, |
| "grad_norm": 0.5782442092895508, |
| "learning_rate": 2.9180000000000002e-05, |
| "loss": 0.0251, |
| "step": 1460 |
| }, |
| { |
| "epoch": 0.1505838967424708, |
| "grad_norm": 0.2764524817466736, |
| "learning_rate": 2.938e-05, |
| "loss": 0.0198, |
| "step": 1470 |
| }, |
| { |
| "epoch": 0.15160827699241958, |
| "grad_norm": 0.5051664710044861, |
| "learning_rate": 2.958e-05, |
| "loss": 0.025, |
| "step": 1480 |
| }, |
| { |
| "epoch": 0.15263265724236838, |
| "grad_norm": 0.2496298849582672, |
| "learning_rate": 2.9780000000000003e-05, |
| "loss": 0.0218, |
| "step": 1490 |
| }, |
| { |
| "epoch": 0.15365703749231716, |
| "grad_norm": 0.24859891831874847, |
| "learning_rate": 2.998e-05, |
| "loss": 0.022, |
| "step": 1500 |
| }, |
| { |
| "epoch": 0.15468141774226593, |
| "grad_norm": 0.33643341064453125, |
| "learning_rate": 2.9999992618302022e-05, |
| "loss": 0.0217, |
| "step": 1510 |
| }, |
| { |
| "epoch": 0.1557057979922147, |
| "grad_norm": 0.3352760672569275, |
| "learning_rate": 2.999996710133069e-05, |
| "loss": 0.0244, |
| "step": 1520 |
| }, |
| { |
| "epoch": 0.15673017824216348, |
| "grad_norm": 0.33244645595550537, |
| "learning_rate": 2.9999923357984923e-05, |
| "loss": 0.0208, |
| "step": 1530 |
| }, |
| { |
| "epoch": 0.15775455849211228, |
| "grad_norm": 0.4294212758541107, |
| "learning_rate": 2.9999861388317874e-05, |
| "loss": 0.0193, |
| "step": 1540 |
| }, |
| { |
| "epoch": 0.15877893874206106, |
| "grad_norm": 0.31875789165496826, |
| "learning_rate": 2.9999781192404847e-05, |
| "loss": 0.027, |
| "step": 1550 |
| }, |
| { |
| "epoch": 0.15980331899200984, |
| "grad_norm": 0.26667237281799316, |
| "learning_rate": 2.9999682770343283e-05, |
| "loss": 0.0238, |
| "step": 1560 |
| }, |
| { |
| "epoch": 0.1608276992419586, |
| "grad_norm": 0.22858957946300507, |
| "learning_rate": 2.9999566122252777e-05, |
| "loss": 0.0224, |
| "step": 1570 |
| }, |
| { |
| "epoch": 0.16185207949190739, |
| "grad_norm": 0.2861299514770508, |
| "learning_rate": 2.9999431248275065e-05, |
| "loss": 0.0242, |
| "step": 1580 |
| }, |
| { |
| "epoch": 0.1628764597418562, |
| "grad_norm": 0.3489325940608978, |
| "learning_rate": 2.9999278148574027e-05, |
| "loss": 0.0221, |
| "step": 1590 |
| }, |
| { |
| "epoch": 0.16390083999180496, |
| "grad_norm": 0.2522561848163605, |
| "learning_rate": 2.9999106823335705e-05, |
| "loss": 0.0226, |
| "step": 1600 |
| }, |
| { |
| "epoch": 0.16492522024175374, |
| "grad_norm": 0.32370346784591675, |
| "learning_rate": 2.999891727276827e-05, |
| "loss": 0.0223, |
| "step": 1610 |
| }, |
| { |
| "epoch": 0.16594960049170251, |
| "grad_norm": 0.27097204327583313, |
| "learning_rate": 2.9998709497102037e-05, |
| "loss": 0.0203, |
| "step": 1620 |
| }, |
| { |
| "epoch": 0.1669739807416513, |
| "grad_norm": 0.25871607661247253, |
| "learning_rate": 2.9998483496589485e-05, |
| "loss": 0.0242, |
| "step": 1630 |
| }, |
| { |
| "epoch": 0.1679983609916001, |
| "grad_norm": 0.316556453704834, |
| "learning_rate": 2.9998239271505217e-05, |
| "loss": 0.0221, |
| "step": 1640 |
| }, |
| { |
| "epoch": 0.16902274124154887, |
| "grad_norm": 0.6326062083244324, |
| "learning_rate": 2.9997976822145995e-05, |
| "loss": 0.0212, |
| "step": 1650 |
| }, |
| { |
| "epoch": 0.17004712149149764, |
| "grad_norm": 0.3151809871196747, |
| "learning_rate": 2.9997696148830714e-05, |
| "loss": 0.0242, |
| "step": 1660 |
| }, |
| { |
| "epoch": 0.17107150174144642, |
| "grad_norm": 0.5152978897094727, |
| "learning_rate": 2.9997397251900423e-05, |
| "loss": 0.0247, |
| "step": 1670 |
| }, |
| { |
| "epoch": 0.1720958819913952, |
| "grad_norm": 0.36530381441116333, |
| "learning_rate": 2.999708013171831e-05, |
| "loss": 0.0214, |
| "step": 1680 |
| }, |
| { |
| "epoch": 0.173120262241344, |
| "grad_norm": 0.3593733310699463, |
| "learning_rate": 2.9996744788669705e-05, |
| "loss": 0.0213, |
| "step": 1690 |
| }, |
| { |
| "epoch": 0.17414464249129277, |
| "grad_norm": 0.20558401942253113, |
| "learning_rate": 2.999639122316208e-05, |
| "loss": 0.0198, |
| "step": 1700 |
| }, |
| { |
| "epoch": 0.17516902274124155, |
| "grad_norm": 0.29789048433303833, |
| "learning_rate": 2.999601943562505e-05, |
| "loss": 0.0263, |
| "step": 1710 |
| }, |
| { |
| "epoch": 0.17619340299119032, |
| "grad_norm": 0.24272756278514862, |
| "learning_rate": 2.9995629426510377e-05, |
| "loss": 0.0219, |
| "step": 1720 |
| }, |
| { |
| "epoch": 0.1772177832411391, |
| "grad_norm": 0.32693320512771606, |
| "learning_rate": 2.9995221196291952e-05, |
| "loss": 0.0184, |
| "step": 1730 |
| }, |
| { |
| "epoch": 0.1782421634910879, |
| "grad_norm": 0.6072489023208618, |
| "learning_rate": 2.999479474546582e-05, |
| "loss": 0.0271, |
| "step": 1740 |
| }, |
| { |
| "epoch": 0.17926654374103668, |
| "grad_norm": 0.3873225450515747, |
| "learning_rate": 2.9994350074550153e-05, |
| "loss": 0.0248, |
| "step": 1750 |
| }, |
| { |
| "epoch": 0.18029092399098545, |
| "grad_norm": 0.635831356048584, |
| "learning_rate": 2.9993887184085273e-05, |
| "loss": 0.0239, |
| "step": 1760 |
| }, |
| { |
| "epoch": 0.18131530424093423, |
| "grad_norm": 0.5479065775871277, |
| "learning_rate": 2.9993406074633628e-05, |
| "loss": 0.022, |
| "step": 1770 |
| }, |
| { |
| "epoch": 0.182339684490883, |
| "grad_norm": 0.5095294117927551, |
| "learning_rate": 2.9992906746779824e-05, |
| "loss": 0.0218, |
| "step": 1780 |
| }, |
| { |
| "epoch": 0.1833640647408318, |
| "grad_norm": 0.15505866706371307, |
| "learning_rate": 2.9992389201130576e-05, |
| "loss": 0.0268, |
| "step": 1790 |
| }, |
| { |
| "epoch": 0.18438844499078058, |
| "grad_norm": 1.085028886795044, |
| "learning_rate": 2.999185343831476e-05, |
| "loss": 0.0282, |
| "step": 1800 |
| }, |
| { |
| "epoch": 0.18541282524072936, |
| "grad_norm": 0.25199732184410095, |
| "learning_rate": 2.999129945898338e-05, |
| "loss": 0.0263, |
| "step": 1810 |
| }, |
| { |
| "epoch": 0.18643720549067813, |
| "grad_norm": 0.37278130650520325, |
| "learning_rate": 2.999072726380957e-05, |
| "loss": 0.0273, |
| "step": 1820 |
| }, |
| { |
| "epoch": 0.1874615857406269, |
| "grad_norm": 0.3573269546031952, |
| "learning_rate": 2.9990136853488603e-05, |
| "loss": 0.0202, |
| "step": 1830 |
| }, |
| { |
| "epoch": 0.1884859659905757, |
| "grad_norm": 0.4594936966896057, |
| "learning_rate": 2.9989528228737878e-05, |
| "loss": 0.0221, |
| "step": 1840 |
| }, |
| { |
| "epoch": 0.18951034624052449, |
| "grad_norm": 0.26588478684425354, |
| "learning_rate": 2.9988901390296935e-05, |
| "loss": 0.0236, |
| "step": 1850 |
| }, |
| { |
| "epoch": 0.19053472649047326, |
| "grad_norm": 0.22512689232826233, |
| "learning_rate": 2.9988256338927442e-05, |
| "loss": 0.0217, |
| "step": 1860 |
| }, |
| { |
| "epoch": 0.19155910674042204, |
| "grad_norm": 0.32946255803108215, |
| "learning_rate": 2.9987593075413198e-05, |
| "loss": 0.0259, |
| "step": 1870 |
| }, |
| { |
| "epoch": 0.19258348699037084, |
| "grad_norm": 0.31236526370048523, |
| "learning_rate": 2.9986911600560128e-05, |
| "loss": 0.0205, |
| "step": 1880 |
| }, |
| { |
| "epoch": 0.19360786724031961, |
| "grad_norm": 0.5991463661193848, |
| "learning_rate": 2.998621191519629e-05, |
| "loss": 0.0241, |
| "step": 1890 |
| }, |
| { |
| "epoch": 0.1946322474902684, |
| "grad_norm": 0.28162118792533875, |
| "learning_rate": 2.9985494020171873e-05, |
| "loss": 0.024, |
| "step": 1900 |
| }, |
| { |
| "epoch": 0.19565662774021716, |
| "grad_norm": 0.2981233596801758, |
| "learning_rate": 2.9984757916359176e-05, |
| "loss": 0.0239, |
| "step": 1910 |
| }, |
| { |
| "epoch": 0.19668100799016594, |
| "grad_norm": 0.2991834580898285, |
| "learning_rate": 2.998400360465264e-05, |
| "loss": 0.0212, |
| "step": 1920 |
| }, |
| { |
| "epoch": 0.19770538824011474, |
| "grad_norm": 0.3823823630809784, |
| "learning_rate": 2.998323108596883e-05, |
| "loss": 0.0204, |
| "step": 1930 |
| }, |
| { |
| "epoch": 0.19872976849006352, |
| "grad_norm": 0.9908068180084229, |
| "learning_rate": 2.9982440361246422e-05, |
| "loss": 0.0222, |
| "step": 1940 |
| }, |
| { |
| "epoch": 0.1997541487400123, |
| "grad_norm": 0.20518527925014496, |
| "learning_rate": 2.998163143144622e-05, |
| "loss": 0.0214, |
| "step": 1950 |
| }, |
| { |
| "epoch": 0.20077852898996107, |
| "grad_norm": 0.3472554683685303, |
| "learning_rate": 2.9980804297551155e-05, |
| "loss": 0.0225, |
| "step": 1960 |
| }, |
| { |
| "epoch": 0.20180290923990984, |
| "grad_norm": 0.4956364333629608, |
| "learning_rate": 2.997995896056627e-05, |
| "loss": 0.0229, |
| "step": 1970 |
| }, |
| { |
| "epoch": 0.20282728948985865, |
| "grad_norm": 0.2038542777299881, |
| "learning_rate": 2.997909542151873e-05, |
| "loss": 0.0234, |
| "step": 1980 |
| }, |
| { |
| "epoch": 0.20385166973980742, |
| "grad_norm": 0.21829165518283844, |
| "learning_rate": 2.9978213681457823e-05, |
| "loss": 0.0192, |
| "step": 1990 |
| }, |
| { |
| "epoch": 0.2048760499897562, |
| "grad_norm": 0.9428406953811646, |
| "learning_rate": 2.9977313741454933e-05, |
| "loss": 0.0252, |
| "step": 2000 |
| }, |
| { |
| "epoch": 0.20590043023970497, |
| "grad_norm": 0.2898987829685211, |
| "learning_rate": 2.997639560260358e-05, |
| "loss": 0.0223, |
| "step": 2010 |
| }, |
| { |
| "epoch": 0.20692481048965375, |
| "grad_norm": 0.3655512034893036, |
| "learning_rate": 2.9975459266019387e-05, |
| "loss": 0.0208, |
| "step": 2020 |
| }, |
| { |
| "epoch": 0.20794919073960255, |
| "grad_norm": 0.2684790790081024, |
| "learning_rate": 2.9974504732840092e-05, |
| "loss": 0.0219, |
| "step": 2030 |
| }, |
| { |
| "epoch": 0.20897357098955133, |
| "grad_norm": 0.3942451477050781, |
| "learning_rate": 2.997353200422555e-05, |
| "loss": 0.0238, |
| "step": 2040 |
| }, |
| { |
| "epoch": 0.2099979512395001, |
| "grad_norm": 0.2401776909828186, |
| "learning_rate": 2.99725410813577e-05, |
| "loss": 0.0206, |
| "step": 2050 |
| }, |
| { |
| "epoch": 0.21102233148944888, |
| "grad_norm": 0.2266601026058197, |
| "learning_rate": 2.997153196544063e-05, |
| "loss": 0.0215, |
| "step": 2060 |
| }, |
| { |
| "epoch": 0.21204671173939765, |
| "grad_norm": 1.1342970132827759, |
| "learning_rate": 2.997050465770049e-05, |
| "loss": 0.0221, |
| "step": 2070 |
| }, |
| { |
| "epoch": 0.21307109198934646, |
| "grad_norm": 0.29266059398651123, |
| "learning_rate": 2.9969459159385562e-05, |
| "loss": 0.0204, |
| "step": 2080 |
| }, |
| { |
| "epoch": 0.21409547223929523, |
| "grad_norm": 0.6652705073356628, |
| "learning_rate": 2.9968395471766237e-05, |
| "loss": 0.021, |
| "step": 2090 |
| }, |
| { |
| "epoch": 0.215119852489244, |
| "grad_norm": 0.268440306186676, |
| "learning_rate": 2.996731359613498e-05, |
| "loss": 0.0215, |
| "step": 2100 |
| }, |
| { |
| "epoch": 0.21614423273919278, |
| "grad_norm": 0.4345571994781494, |
| "learning_rate": 2.996621353380638e-05, |
| "loss": 0.0218, |
| "step": 2110 |
| }, |
| { |
| "epoch": 0.21716861298914156, |
| "grad_norm": 0.4188554883003235, |
| "learning_rate": 2.9965095286117113e-05, |
| "loss": 0.0224, |
| "step": 2120 |
| }, |
| { |
| "epoch": 0.21819299323909036, |
| "grad_norm": 0.5320345759391785, |
| "learning_rate": 2.996395885442596e-05, |
| "loss": 0.0215, |
| "step": 2130 |
| }, |
| { |
| "epoch": 0.21921737348903914, |
| "grad_norm": 0.24155373871326447, |
| "learning_rate": 2.996280424011379e-05, |
| "loss": 0.0202, |
| "step": 2140 |
| }, |
| { |
| "epoch": 0.2202417537389879, |
| "grad_norm": 0.28205323219299316, |
| "learning_rate": 2.9961631444583565e-05, |
| "loss": 0.022, |
| "step": 2150 |
| }, |
| { |
| "epoch": 0.2212661339889367, |
| "grad_norm": 0.38679301738739014, |
| "learning_rate": 2.9960440469260347e-05, |
| "loss": 0.0198, |
| "step": 2160 |
| }, |
| { |
| "epoch": 0.22229051423888546, |
| "grad_norm": 0.4031623601913452, |
| "learning_rate": 2.995923131559128e-05, |
| "loss": 0.0233, |
| "step": 2170 |
| }, |
| { |
| "epoch": 0.22331489448883426, |
| "grad_norm": 0.25150367617607117, |
| "learning_rate": 2.99580039850456e-05, |
| "loss": 0.0188, |
| "step": 2180 |
| }, |
| { |
| "epoch": 0.22433927473878304, |
| "grad_norm": 0.3358060121536255, |
| "learning_rate": 2.9956758479114635e-05, |
| "loss": 0.0219, |
| "step": 2190 |
| }, |
| { |
| "epoch": 0.22536365498873182, |
| "grad_norm": 0.22465164959430695, |
| "learning_rate": 2.9955494799311787e-05, |
| "loss": 0.0196, |
| "step": 2200 |
| }, |
| { |
| "epoch": 0.2263880352386806, |
| "grad_norm": 0.33016011118888855, |
| "learning_rate": 2.995421294717254e-05, |
| "loss": 0.0207, |
| "step": 2210 |
| }, |
| { |
| "epoch": 0.22741241548862937, |
| "grad_norm": 0.24270613491535187, |
| "learning_rate": 2.995291292425448e-05, |
| "loss": 0.0164, |
| "step": 2220 |
| }, |
| { |
| "epoch": 0.22843679573857817, |
| "grad_norm": 0.22636091709136963, |
| "learning_rate": 2.9951594732137248e-05, |
| "loss": 0.0227, |
| "step": 2230 |
| }, |
| { |
| "epoch": 0.22946117598852694, |
| "grad_norm": 0.3150896728038788, |
| "learning_rate": 2.995025837242257e-05, |
| "loss": 0.0223, |
| "step": 2240 |
| }, |
| { |
| "epoch": 0.23048555623847572, |
| "grad_norm": 0.2518464922904968, |
| "learning_rate": 2.994890384673426e-05, |
| "loss": 0.0174, |
| "step": 2250 |
| }, |
| { |
| "epoch": 0.2315099364884245, |
| "grad_norm": 0.24445340037345886, |
| "learning_rate": 2.9947531156718174e-05, |
| "loss": 0.0205, |
| "step": 2260 |
| }, |
| { |
| "epoch": 0.2325343167383733, |
| "grad_norm": 0.46379026770591736, |
| "learning_rate": 2.994614030404228e-05, |
| "loss": 0.0227, |
| "step": 2270 |
| }, |
| { |
| "epoch": 0.23355869698832207, |
| "grad_norm": 0.3491224944591522, |
| "learning_rate": 2.994473129039659e-05, |
| "loss": 0.0174, |
| "step": 2280 |
| }, |
| { |
| "epoch": 0.23458307723827085, |
| "grad_norm": 0.22474105656147003, |
| "learning_rate": 2.9943304117493183e-05, |
| "loss": 0.0185, |
| "step": 2290 |
| }, |
| { |
| "epoch": 0.23560745748821962, |
| "grad_norm": 0.44733479619026184, |
| "learning_rate": 2.9941858787066207e-05, |
| "loss": 0.0196, |
| "step": 2300 |
| }, |
| { |
| "epoch": 0.2366318377381684, |
| "grad_norm": 0.22745877504348755, |
| "learning_rate": 2.9940395300871888e-05, |
| "loss": 0.0178, |
| "step": 2310 |
| }, |
| { |
| "epoch": 0.2376562179881172, |
| "grad_norm": 0.33024725317955017, |
| "learning_rate": 2.993891366068849e-05, |
| "loss": 0.0185, |
| "step": 2320 |
| }, |
| { |
| "epoch": 0.23868059823806598, |
| "grad_norm": 0.40524354577064514, |
| "learning_rate": 2.993741386831634e-05, |
| "loss": 0.0175, |
| "step": 2330 |
| }, |
| { |
| "epoch": 0.23970497848801475, |
| "grad_norm": 0.5133205652236938, |
| "learning_rate": 2.9935895925577837e-05, |
| "loss": 0.0186, |
| "step": 2340 |
| }, |
| { |
| "epoch": 0.24072935873796353, |
| "grad_norm": 0.3881397843360901, |
| "learning_rate": 2.9934359834317425e-05, |
| "loss": 0.0246, |
| "step": 2350 |
| }, |
| { |
| "epoch": 0.2417537389879123, |
| "grad_norm": 0.3839256167411804, |
| "learning_rate": 2.993280559640159e-05, |
| "loss": 0.0219, |
| "step": 2360 |
| }, |
| { |
| "epoch": 0.2427781192378611, |
| "grad_norm": 0.3518000543117523, |
| "learning_rate": 2.9931233213718887e-05, |
| "loss": 0.023, |
| "step": 2370 |
| }, |
| { |
| "epoch": 0.24380249948780988, |
| "grad_norm": 0.195148766040802, |
| "learning_rate": 2.992964268817991e-05, |
| "loss": 0.0186, |
| "step": 2380 |
| }, |
| { |
| "epoch": 0.24482687973775866, |
| "grad_norm": 0.23285473883152008, |
| "learning_rate": 2.9928034021717286e-05, |
| "loss": 0.0208, |
| "step": 2390 |
| }, |
| { |
| "epoch": 0.24585125998770743, |
| "grad_norm": 0.66872638463974, |
| "learning_rate": 2.9926407216285707e-05, |
| "loss": 0.0217, |
| "step": 2400 |
| }, |
| { |
| "epoch": 0.2468756402376562, |
| "grad_norm": 0.2576395869255066, |
| "learning_rate": 2.9924762273861894e-05, |
| "loss": 0.0241, |
| "step": 2410 |
| }, |
| { |
| "epoch": 0.247900020487605, |
| "grad_norm": 0.5584348440170288, |
| "learning_rate": 2.992309919644461e-05, |
| "loss": 0.0191, |
| "step": 2420 |
| }, |
| { |
| "epoch": 0.24892440073755379, |
| "grad_norm": 0.15737676620483398, |
| "learning_rate": 2.9921417986054633e-05, |
| "loss": 0.0181, |
| "step": 2430 |
| }, |
| { |
| "epoch": 0.24994878098750256, |
| "grad_norm": 0.32971587777137756, |
| "learning_rate": 2.9919718644734815e-05, |
| "loss": 0.0201, |
| "step": 2440 |
| }, |
| { |
| "epoch": 0.25097316123745134, |
| "grad_norm": 0.27433663606643677, |
| "learning_rate": 2.991800117455e-05, |
| "loss": 0.0197, |
| "step": 2450 |
| }, |
| { |
| "epoch": 0.2519975414874001, |
| "grad_norm": 0.41258522868156433, |
| "learning_rate": 2.9916265577587085e-05, |
| "loss": 0.02, |
| "step": 2460 |
| }, |
| { |
| "epoch": 0.2530219217373489, |
| "grad_norm": 0.39694029092788696, |
| "learning_rate": 2.9914511855954986e-05, |
| "loss": 0.0217, |
| "step": 2470 |
| }, |
| { |
| "epoch": 0.25404630198729766, |
| "grad_norm": 0.33073779940605164, |
| "learning_rate": 2.9912740011784626e-05, |
| "loss": 0.0226, |
| "step": 2480 |
| }, |
| { |
| "epoch": 0.2550706822372465, |
| "grad_norm": 0.21533173322677612, |
| "learning_rate": 2.991095004722898e-05, |
| "loss": 0.023, |
| "step": 2490 |
| }, |
| { |
| "epoch": 0.25609506248719527, |
| "grad_norm": 0.280460387468338, |
| "learning_rate": 2.9909141964463017e-05, |
| "loss": 0.0192, |
| "step": 2500 |
| }, |
| { |
| "epoch": 0.25711944273714404, |
| "grad_norm": 0.26816385984420776, |
| "learning_rate": 2.990731576568372e-05, |
| "loss": 0.0204, |
| "step": 2510 |
| }, |
| { |
| "epoch": 0.2581438229870928, |
| "grad_norm": 0.3736421763896942, |
| "learning_rate": 2.99054714531101e-05, |
| "loss": 0.022, |
| "step": 2520 |
| }, |
| { |
| "epoch": 0.2591682032370416, |
| "grad_norm": 0.3605470359325409, |
| "learning_rate": 2.9903609028983178e-05, |
| "loss": 0.0187, |
| "step": 2530 |
| }, |
| { |
| "epoch": 0.26019258348699037, |
| "grad_norm": 0.24020755290985107, |
| "learning_rate": 2.990172849556596e-05, |
| "loss": 0.0225, |
| "step": 2540 |
| }, |
| { |
| "epoch": 0.26121696373693915, |
| "grad_norm": 0.21847979724407196, |
| "learning_rate": 2.989982985514348e-05, |
| "loss": 0.0224, |
| "step": 2550 |
| }, |
| { |
| "epoch": 0.2622413439868879, |
| "grad_norm": 0.5397780537605286, |
| "learning_rate": 2.9897913110022766e-05, |
| "loss": 0.0192, |
| "step": 2560 |
| }, |
| { |
| "epoch": 0.2632657242368367, |
| "grad_norm": 0.41989433765411377, |
| "learning_rate": 2.989597826253284e-05, |
| "loss": 0.0183, |
| "step": 2570 |
| }, |
| { |
| "epoch": 0.26429010448678547, |
| "grad_norm": 0.564731240272522, |
| "learning_rate": 2.989402531502473e-05, |
| "loss": 0.0264, |
| "step": 2580 |
| }, |
| { |
| "epoch": 0.2653144847367343, |
| "grad_norm": 0.22362561523914337, |
| "learning_rate": 2.989205426987144e-05, |
| "loss": 0.0182, |
| "step": 2590 |
| }, |
| { |
| "epoch": 0.2663388649866831, |
| "grad_norm": 0.23313681781291962, |
| "learning_rate": 2.9890065129467988e-05, |
| "loss": 0.024, |
| "step": 2600 |
| }, |
| { |
| "epoch": 0.26736324523663185, |
| "grad_norm": 0.2222335785627365, |
| "learning_rate": 2.9888057896231364e-05, |
| "loss": 0.0211, |
| "step": 2610 |
| }, |
| { |
| "epoch": 0.2683876254865806, |
| "grad_norm": 0.33038464188575745, |
| "learning_rate": 2.988603257260054e-05, |
| "loss": 0.0192, |
| "step": 2620 |
| }, |
| { |
| "epoch": 0.2694120057365294, |
| "grad_norm": 0.38392651081085205, |
| "learning_rate": 2.988398916103648e-05, |
| "loss": 0.018, |
| "step": 2630 |
| }, |
| { |
| "epoch": 0.2704363859864782, |
| "grad_norm": 0.38158372044563293, |
| "learning_rate": 2.988192766402212e-05, |
| "loss": 0.0204, |
| "step": 2640 |
| }, |
| { |
| "epoch": 0.27146076623642695, |
| "grad_norm": 0.3307957947254181, |
| "learning_rate": 2.9879848084062378e-05, |
| "loss": 0.0218, |
| "step": 2650 |
| }, |
| { |
| "epoch": 0.27248514648637573, |
| "grad_norm": 0.24566832184791565, |
| "learning_rate": 2.9877750423684132e-05, |
| "loss": 0.0211, |
| "step": 2660 |
| }, |
| { |
| "epoch": 0.2735095267363245, |
| "grad_norm": 0.232928067445755, |
| "learning_rate": 2.987563468543624e-05, |
| "loss": 0.0203, |
| "step": 2670 |
| }, |
| { |
| "epoch": 0.2745339069862733, |
| "grad_norm": 0.299601674079895, |
| "learning_rate": 2.9873500871889533e-05, |
| "loss": 0.0237, |
| "step": 2680 |
| }, |
| { |
| "epoch": 0.2755582872362221, |
| "grad_norm": 0.2526320517063141, |
| "learning_rate": 2.9871348985636778e-05, |
| "loss": 0.02, |
| "step": 2690 |
| }, |
| { |
| "epoch": 0.2765826674861709, |
| "grad_norm": 0.31302836537361145, |
| "learning_rate": 2.986917902929273e-05, |
| "loss": 0.0187, |
| "step": 2700 |
| }, |
| { |
| "epoch": 0.27760704773611966, |
| "grad_norm": 0.25056999921798706, |
| "learning_rate": 2.9866991005494095e-05, |
| "loss": 0.0185, |
| "step": 2710 |
| }, |
| { |
| "epoch": 0.27863142798606844, |
| "grad_norm": 0.2282654047012329, |
| "learning_rate": 2.986478491689952e-05, |
| "loss": 0.0206, |
| "step": 2720 |
| }, |
| { |
| "epoch": 0.2796558082360172, |
| "grad_norm": 0.31271490454673767, |
| "learning_rate": 2.986256076618961e-05, |
| "loss": 0.0229, |
| "step": 2730 |
| }, |
| { |
| "epoch": 0.280680188485966, |
| "grad_norm": 0.24251356720924377, |
| "learning_rate": 2.986031855606692e-05, |
| "loss": 0.0167, |
| "step": 2740 |
| }, |
| { |
| "epoch": 0.28170456873591476, |
| "grad_norm": 0.7054775357246399, |
| "learning_rate": 2.9858058289255954e-05, |
| "loss": 0.0181, |
| "step": 2750 |
| }, |
| { |
| "epoch": 0.28272894898586354, |
| "grad_norm": 0.30229780077934265, |
| "learning_rate": 2.985577996850314e-05, |
| "loss": 0.0157, |
| "step": 2760 |
| }, |
| { |
| "epoch": 0.2837533292358123, |
| "grad_norm": 0.2939280569553375, |
| "learning_rate": 2.9853483596576856e-05, |
| "loss": 0.0213, |
| "step": 2770 |
| }, |
| { |
| "epoch": 0.2847777094857611, |
| "grad_norm": 0.3949582874774933, |
| "learning_rate": 2.9851169176267412e-05, |
| "loss": 0.0212, |
| "step": 2780 |
| }, |
| { |
| "epoch": 0.2858020897357099, |
| "grad_norm": 0.43970954418182373, |
| "learning_rate": 2.9848836710387044e-05, |
| "loss": 0.0201, |
| "step": 2790 |
| }, |
| { |
| "epoch": 0.2868264699856587, |
| "grad_norm": 0.7327799201011658, |
| "learning_rate": 2.9846486201769917e-05, |
| "loss": 0.0185, |
| "step": 2800 |
| }, |
| { |
| "epoch": 0.28785085023560747, |
| "grad_norm": 0.2915540337562561, |
| "learning_rate": 2.9844117653272122e-05, |
| "loss": 0.0209, |
| "step": 2810 |
| }, |
| { |
| "epoch": 0.28887523048555624, |
| "grad_norm": 0.5864996314048767, |
| "learning_rate": 2.9841731067771676e-05, |
| "loss": 0.0192, |
| "step": 2820 |
| }, |
| { |
| "epoch": 0.289899610735505, |
| "grad_norm": 0.18030302226543427, |
| "learning_rate": 2.9839326448168492e-05, |
| "loss": 0.0202, |
| "step": 2830 |
| }, |
| { |
| "epoch": 0.2909239909854538, |
| "grad_norm": 0.22470015287399292, |
| "learning_rate": 2.983690379738442e-05, |
| "loss": 0.0169, |
| "step": 2840 |
| }, |
| { |
| "epoch": 0.29194837123540257, |
| "grad_norm": 0.21245791018009186, |
| "learning_rate": 2.9834463118363204e-05, |
| "loss": 0.0179, |
| "step": 2850 |
| }, |
| { |
| "epoch": 0.29297275148535135, |
| "grad_norm": 0.3034825623035431, |
| "learning_rate": 2.98320044140705e-05, |
| "loss": 0.0219, |
| "step": 2860 |
| }, |
| { |
| "epoch": 0.2939971317353001, |
| "grad_norm": 0.2279258817434311, |
| "learning_rate": 2.9829527687493868e-05, |
| "loss": 0.0184, |
| "step": 2870 |
| }, |
| { |
| "epoch": 0.29502151198524895, |
| "grad_norm": 0.22371402382850647, |
| "learning_rate": 2.9827032941642765e-05, |
| "loss": 0.0158, |
| "step": 2880 |
| }, |
| { |
| "epoch": 0.2960458922351977, |
| "grad_norm": 0.34844109416007996, |
| "learning_rate": 2.9824520179548535e-05, |
| "loss": 0.0188, |
| "step": 2890 |
| }, |
| { |
| "epoch": 0.2970702724851465, |
| "grad_norm": 0.3442723751068115, |
| "learning_rate": 2.9821989404264428e-05, |
| "loss": 0.0178, |
| "step": 2900 |
| }, |
| { |
| "epoch": 0.2980946527350953, |
| "grad_norm": 0.335965096950531, |
| "learning_rate": 2.9819440618865572e-05, |
| "loss": 0.0217, |
| "step": 2910 |
| }, |
| { |
| "epoch": 0.29911903298504405, |
| "grad_norm": 0.22416916489601135, |
| "learning_rate": 2.981687382644898e-05, |
| "loss": 0.0175, |
| "step": 2920 |
| }, |
| { |
| "epoch": 0.30014341323499283, |
| "grad_norm": 0.2343377023935318, |
| "learning_rate": 2.9814289030133556e-05, |
| "loss": 0.0175, |
| "step": 2930 |
| }, |
| { |
| "epoch": 0.3011677934849416, |
| "grad_norm": 0.3242785930633545, |
| "learning_rate": 2.9811686233060058e-05, |
| "loss": 0.0194, |
| "step": 2940 |
| }, |
| { |
| "epoch": 0.3021921737348904, |
| "grad_norm": 0.26919203996658325, |
| "learning_rate": 2.9809065438391138e-05, |
| "loss": 0.0221, |
| "step": 2950 |
| }, |
| { |
| "epoch": 0.30321655398483915, |
| "grad_norm": 0.4272896349430084, |
| "learning_rate": 2.9806426649311304e-05, |
| "loss": 0.0175, |
| "step": 2960 |
| }, |
| { |
| "epoch": 0.30424093423478793, |
| "grad_norm": 0.18241579830646515, |
| "learning_rate": 2.980376986902694e-05, |
| "loss": 0.0173, |
| "step": 2970 |
| }, |
| { |
| "epoch": 0.30526531448473676, |
| "grad_norm": 0.4052042067050934, |
| "learning_rate": 2.9801095100766276e-05, |
| "loss": 0.0213, |
| "step": 2980 |
| }, |
| { |
| "epoch": 0.30628969473468554, |
| "grad_norm": 0.275488942861557, |
| "learning_rate": 2.979840234777942e-05, |
| "loss": 0.0229, |
| "step": 2990 |
| }, |
| { |
| "epoch": 0.3073140749846343, |
| "grad_norm": 0.24694477021694183, |
| "learning_rate": 2.979569161333831e-05, |
| "loss": 0.0168, |
| "step": 3000 |
| }, |
| { |
| "epoch": 0.3083384552345831, |
| "grad_norm": 0.16428786516189575, |
| "learning_rate": 2.9792962900736744e-05, |
| "loss": 0.0199, |
| "step": 3010 |
| }, |
| { |
| "epoch": 0.30936283548453186, |
| "grad_norm": 0.21410594880580902, |
| "learning_rate": 2.9790216213290372e-05, |
| "loss": 0.017, |
| "step": 3020 |
| }, |
| { |
| "epoch": 0.31038721573448064, |
| "grad_norm": 0.16461731493473053, |
| "learning_rate": 2.9787451554336672e-05, |
| "loss": 0.017, |
| "step": 3030 |
| }, |
| { |
| "epoch": 0.3114115959844294, |
| "grad_norm": 0.3472115397453308, |
| "learning_rate": 2.9784668927234972e-05, |
| "loss": 0.0194, |
| "step": 3040 |
| }, |
| { |
| "epoch": 0.3124359762343782, |
| "grad_norm": 0.3556407392024994, |
| "learning_rate": 2.9781868335366418e-05, |
| "loss": 0.0208, |
| "step": 3050 |
| }, |
| { |
| "epoch": 0.31346035648432696, |
| "grad_norm": 0.21875253319740295, |
| "learning_rate": 2.9779049782134e-05, |
| "loss": 0.0206, |
| "step": 3060 |
| }, |
| { |
| "epoch": 0.31448473673427574, |
| "grad_norm": 0.2815505862236023, |
| "learning_rate": 2.9776213270962523e-05, |
| "loss": 0.017, |
| "step": 3070 |
| }, |
| { |
| "epoch": 0.31550911698422457, |
| "grad_norm": 0.22914394736289978, |
| "learning_rate": 2.977335880529861e-05, |
| "loss": 0.0202, |
| "step": 3080 |
| }, |
| { |
| "epoch": 0.31653349723417334, |
| "grad_norm": 0.3067605793476105, |
| "learning_rate": 2.977048638861072e-05, |
| "loss": 0.0214, |
| "step": 3090 |
| }, |
| { |
| "epoch": 0.3175578774841221, |
| "grad_norm": 0.30246803164482117, |
| "learning_rate": 2.97675960243891e-05, |
| "loss": 0.0168, |
| "step": 3100 |
| }, |
| { |
| "epoch": 0.3185822577340709, |
| "grad_norm": 0.19881637394428253, |
| "learning_rate": 2.976468771614582e-05, |
| "loss": 0.0204, |
| "step": 3110 |
| }, |
| { |
| "epoch": 0.31960663798401967, |
| "grad_norm": 0.30013415217399597, |
| "learning_rate": 2.9761761467414736e-05, |
| "loss": 0.0235, |
| "step": 3120 |
| }, |
| { |
| "epoch": 0.32063101823396845, |
| "grad_norm": 0.47801586985588074, |
| "learning_rate": 2.9758817281751538e-05, |
| "loss": 0.02, |
| "step": 3130 |
| }, |
| { |
| "epoch": 0.3216553984839172, |
| "grad_norm": 0.23129256069660187, |
| "learning_rate": 2.975585516273367e-05, |
| "loss": 0.0235, |
| "step": 3140 |
| }, |
| { |
| "epoch": 0.322679778733866, |
| "grad_norm": 0.25135642290115356, |
| "learning_rate": 2.975287511396039e-05, |
| "loss": 0.0199, |
| "step": 3150 |
| }, |
| { |
| "epoch": 0.32370415898381477, |
| "grad_norm": 0.42432141304016113, |
| "learning_rate": 2.9749877139052745e-05, |
| "loss": 0.0209, |
| "step": 3160 |
| }, |
| { |
| "epoch": 0.32472853923376355, |
| "grad_norm": 0.24515940248966217, |
| "learning_rate": 2.9746861241653555e-05, |
| "loss": 0.0197, |
| "step": 3170 |
| }, |
| { |
| "epoch": 0.3257529194837124, |
| "grad_norm": 0.30593010783195496, |
| "learning_rate": 2.9743827425427416e-05, |
| "loss": 0.0188, |
| "step": 3180 |
| }, |
| { |
| "epoch": 0.32677729973366115, |
| "grad_norm": 0.28354907035827637, |
| "learning_rate": 2.9740775694060704e-05, |
| "loss": 0.0198, |
| "step": 3190 |
| }, |
| { |
| "epoch": 0.32780167998360993, |
| "grad_norm": 0.2540890574455261, |
| "learning_rate": 2.973770605126156e-05, |
| "loss": 0.0182, |
| "step": 3200 |
| }, |
| { |
| "epoch": 0.3288260602335587, |
| "grad_norm": 0.2954034209251404, |
| "learning_rate": 2.9734618500759886e-05, |
| "loss": 0.0201, |
| "step": 3210 |
| }, |
| { |
| "epoch": 0.3298504404835075, |
| "grad_norm": 0.2634400427341461, |
| "learning_rate": 2.9731513046307356e-05, |
| "loss": 0.0193, |
| "step": 3220 |
| }, |
| { |
| "epoch": 0.33087482073345625, |
| "grad_norm": 0.5348861217498779, |
| "learning_rate": 2.972838969167738e-05, |
| "loss": 0.0172, |
| "step": 3230 |
| }, |
| { |
| "epoch": 0.33189920098340503, |
| "grad_norm": 0.2987020015716553, |
| "learning_rate": 2.972524844066513e-05, |
| "loss": 0.0192, |
| "step": 3240 |
| }, |
| { |
| "epoch": 0.3329235812333538, |
| "grad_norm": 0.288549542427063, |
| "learning_rate": 2.9722089297087524e-05, |
| "loss": 0.0189, |
| "step": 3250 |
| }, |
| { |
| "epoch": 0.3339479614833026, |
| "grad_norm": 0.32910794019699097, |
| "learning_rate": 2.971891226478322e-05, |
| "loss": 0.0185, |
| "step": 3260 |
| }, |
| { |
| "epoch": 0.33497234173325136, |
| "grad_norm": 0.23422807455062866, |
| "learning_rate": 2.9715717347612615e-05, |
| "loss": 0.022, |
| "step": 3270 |
| }, |
| { |
| "epoch": 0.3359967219832002, |
| "grad_norm": 0.27589085698127747, |
| "learning_rate": 2.971250454945783e-05, |
| "loss": 0.0167, |
| "step": 3280 |
| }, |
| { |
| "epoch": 0.33702110223314896, |
| "grad_norm": 0.1966925412416458, |
| "learning_rate": 2.970927387422272e-05, |
| "loss": 0.0199, |
| "step": 3290 |
| }, |
| { |
| "epoch": 0.33804548248309774, |
| "grad_norm": 0.3010067939758301, |
| "learning_rate": 2.9706025325832858e-05, |
| "loss": 0.0201, |
| "step": 3300 |
| }, |
| { |
| "epoch": 0.3390698627330465, |
| "grad_norm": 0.21337535977363586, |
| "learning_rate": 2.9702758908235534e-05, |
| "loss": 0.0207, |
| "step": 3310 |
| }, |
| { |
| "epoch": 0.3400942429829953, |
| "grad_norm": 0.3775956928730011, |
| "learning_rate": 2.9699474625399765e-05, |
| "loss": 0.022, |
| "step": 3320 |
| }, |
| { |
| "epoch": 0.34111862323294406, |
| "grad_norm": 0.19842292368412018, |
| "learning_rate": 2.9696172481316246e-05, |
| "loss": 0.0165, |
| "step": 3330 |
| }, |
| { |
| "epoch": 0.34214300348289284, |
| "grad_norm": 0.26227444410324097, |
| "learning_rate": 2.9692852479997407e-05, |
| "loss": 0.0213, |
| "step": 3340 |
| }, |
| { |
| "epoch": 0.3431673837328416, |
| "grad_norm": 0.5673602223396301, |
| "learning_rate": 2.9689514625477364e-05, |
| "loss": 0.0211, |
| "step": 3350 |
| }, |
| { |
| "epoch": 0.3441917639827904, |
| "grad_norm": 0.13742130994796753, |
| "learning_rate": 2.968615892181192e-05, |
| "loss": 0.0175, |
| "step": 3360 |
| }, |
| { |
| "epoch": 0.3452161442327392, |
| "grad_norm": 0.2696261703968048, |
| "learning_rate": 2.9682785373078567e-05, |
| "loss": 0.0189, |
| "step": 3370 |
| }, |
| { |
| "epoch": 0.346240524482688, |
| "grad_norm": 0.3034473657608032, |
| "learning_rate": 2.9679393983376493e-05, |
| "loss": 0.0176, |
| "step": 3380 |
| }, |
| { |
| "epoch": 0.34726490473263677, |
| "grad_norm": 0.31520402431488037, |
| "learning_rate": 2.9675984756826545e-05, |
| "loss": 0.016, |
| "step": 3390 |
| }, |
| { |
| "epoch": 0.34828928498258555, |
| "grad_norm": 0.3086974620819092, |
| "learning_rate": 2.9672557697571273e-05, |
| "loss": 0.0195, |
| "step": 3400 |
| }, |
| { |
| "epoch": 0.3493136652325343, |
| "grad_norm": 0.42995473742485046, |
| "learning_rate": 2.966911280977486e-05, |
| "loss": 0.0181, |
| "step": 3410 |
| }, |
| { |
| "epoch": 0.3503380454824831, |
| "grad_norm": 0.20370672643184662, |
| "learning_rate": 2.966565009762318e-05, |
| "loss": 0.016, |
| "step": 3420 |
| }, |
| { |
| "epoch": 0.35136242573243187, |
| "grad_norm": 0.27999642491340637, |
| "learning_rate": 2.966216956532375e-05, |
| "loss": 0.0173, |
| "step": 3430 |
| }, |
| { |
| "epoch": 0.35238680598238065, |
| "grad_norm": 1.1267510652542114, |
| "learning_rate": 2.9658671217105744e-05, |
| "loss": 0.0179, |
| "step": 3440 |
| }, |
| { |
| "epoch": 0.3534111862323294, |
| "grad_norm": 0.3172823488712311, |
| "learning_rate": 2.965515505721999e-05, |
| "loss": 0.0181, |
| "step": 3450 |
| }, |
| { |
| "epoch": 0.3544355664822782, |
| "grad_norm": 0.29089081287384033, |
| "learning_rate": 2.9651621089938948e-05, |
| "loss": 0.0155, |
| "step": 3460 |
| }, |
| { |
| "epoch": 0.35545994673222703, |
| "grad_norm": 0.6848070621490479, |
| "learning_rate": 2.9648069319556725e-05, |
| "loss": 0.0172, |
| "step": 3470 |
| }, |
| { |
| "epoch": 0.3564843269821758, |
| "grad_norm": 0.44010260701179504, |
| "learning_rate": 2.9644499750389057e-05, |
| "loss": 0.0229, |
| "step": 3480 |
| }, |
| { |
| "epoch": 0.3575087072321246, |
| "grad_norm": 0.23993587493896484, |
| "learning_rate": 2.9640912386773305e-05, |
| "loss": 0.0172, |
| "step": 3490 |
| }, |
| { |
| "epoch": 0.35853308748207335, |
| "grad_norm": 0.22536428272724152, |
| "learning_rate": 2.9637307233068455e-05, |
| "loss": 0.0188, |
| "step": 3500 |
| }, |
| { |
| "epoch": 0.35955746773202213, |
| "grad_norm": 0.3249106705188751, |
| "learning_rate": 2.9633684293655108e-05, |
| "loss": 0.0257, |
| "step": 3510 |
| }, |
| { |
| "epoch": 0.3605818479819709, |
| "grad_norm": 0.2145848423242569, |
| "learning_rate": 2.963004357293548e-05, |
| "loss": 0.0172, |
| "step": 3520 |
| }, |
| { |
| "epoch": 0.3616062282319197, |
| "grad_norm": 0.18960341811180115, |
| "learning_rate": 2.962638507533339e-05, |
| "loss": 0.0182, |
| "step": 3530 |
| }, |
| { |
| "epoch": 0.36263060848186845, |
| "grad_norm": 0.30007797479629517, |
| "learning_rate": 2.962270880529425e-05, |
| "loss": 0.0197, |
| "step": 3540 |
| }, |
| { |
| "epoch": 0.36365498873181723, |
| "grad_norm": 0.199548602104187, |
| "learning_rate": 2.9619014767285084e-05, |
| "loss": 0.0173, |
| "step": 3550 |
| }, |
| { |
| "epoch": 0.364679368981766, |
| "grad_norm": 0.2607710659503937, |
| "learning_rate": 2.9615302965794488e-05, |
| "loss": 0.017, |
| "step": 3560 |
| }, |
| { |
| "epoch": 0.36570374923171484, |
| "grad_norm": 0.38235655426979065, |
| "learning_rate": 2.961157340533266e-05, |
| "loss": 0.0171, |
| "step": 3570 |
| }, |
| { |
| "epoch": 0.3667281294816636, |
| "grad_norm": 0.2639552354812622, |
| "learning_rate": 2.9607826090431364e-05, |
| "loss": 0.017, |
| "step": 3580 |
| }, |
| { |
| "epoch": 0.3677525097316124, |
| "grad_norm": 0.3174300193786621, |
| "learning_rate": 2.9604061025643936e-05, |
| "loss": 0.0172, |
| "step": 3590 |
| }, |
| { |
| "epoch": 0.36877688998156116, |
| "grad_norm": 0.21632853150367737, |
| "learning_rate": 2.9600278215545292e-05, |
| "loss": 0.0189, |
| "step": 3600 |
| }, |
| { |
| "epoch": 0.36980127023150994, |
| "grad_norm": 0.3494323194026947, |
| "learning_rate": 2.95964776647319e-05, |
| "loss": 0.0145, |
| "step": 3610 |
| }, |
| { |
| "epoch": 0.3708256504814587, |
| "grad_norm": 0.29196152091026306, |
| "learning_rate": 2.9592659377821787e-05, |
| "loss": 0.0187, |
| "step": 3620 |
| }, |
| { |
| "epoch": 0.3718500307314075, |
| "grad_norm": 0.47920531034469604, |
| "learning_rate": 2.9588823359454528e-05, |
| "loss": 0.0198, |
| "step": 3630 |
| }, |
| { |
| "epoch": 0.37287441098135626, |
| "grad_norm": 0.30677318572998047, |
| "learning_rate": 2.958496961429126e-05, |
| "loss": 0.0154, |
| "step": 3640 |
| }, |
| { |
| "epoch": 0.37389879123130504, |
| "grad_norm": 0.7479692101478577, |
| "learning_rate": 2.9581098147014628e-05, |
| "loss": 0.0169, |
| "step": 3650 |
| }, |
| { |
| "epoch": 0.3749231714812538, |
| "grad_norm": 0.307331919670105, |
| "learning_rate": 2.9577208962328843e-05, |
| "loss": 0.0187, |
| "step": 3660 |
| }, |
| { |
| "epoch": 0.37594755173120264, |
| "grad_norm": 0.23414087295532227, |
| "learning_rate": 2.9573302064959633e-05, |
| "loss": 0.0225, |
| "step": 3670 |
| }, |
| { |
| "epoch": 0.3769719319811514, |
| "grad_norm": 0.21522125601768494, |
| "learning_rate": 2.9569377459654235e-05, |
| "loss": 0.0173, |
| "step": 3680 |
| }, |
| { |
| "epoch": 0.3779963122311002, |
| "grad_norm": 0.6576694846153259, |
| "learning_rate": 2.9565435151181424e-05, |
| "loss": 0.0216, |
| "step": 3690 |
| }, |
| { |
| "epoch": 0.37902069248104897, |
| "grad_norm": 0.3197406530380249, |
| "learning_rate": 2.9561475144331473e-05, |
| "loss": 0.0176, |
| "step": 3700 |
| }, |
| { |
| "epoch": 0.38004507273099775, |
| "grad_norm": 0.4311569333076477, |
| "learning_rate": 2.9557497443916162e-05, |
| "loss": 0.0205, |
| "step": 3710 |
| }, |
| { |
| "epoch": 0.3810694529809465, |
| "grad_norm": 0.17287030816078186, |
| "learning_rate": 2.9553502054768774e-05, |
| "loss": 0.0175, |
| "step": 3720 |
| }, |
| { |
| "epoch": 0.3820938332308953, |
| "grad_norm": 0.16772933304309845, |
| "learning_rate": 2.954948898174409e-05, |
| "loss": 0.0176, |
| "step": 3730 |
| }, |
| { |
| "epoch": 0.38311821348084407, |
| "grad_norm": 0.17275920510292053, |
| "learning_rate": 2.954545822971836e-05, |
| "loss": 0.0181, |
| "step": 3740 |
| }, |
| { |
| "epoch": 0.38414259373079285, |
| "grad_norm": 0.32417425513267517, |
| "learning_rate": 2.9541409803589335e-05, |
| "loss": 0.0162, |
| "step": 3750 |
| }, |
| { |
| "epoch": 0.3851669739807417, |
| "grad_norm": 0.2602884769439697, |
| "learning_rate": 2.9537343708276235e-05, |
| "loss": 0.0223, |
| "step": 3760 |
| }, |
| { |
| "epoch": 0.38619135423069045, |
| "grad_norm": 0.35604560375213623, |
| "learning_rate": 2.953325994871975e-05, |
| "loss": 0.0176, |
| "step": 3770 |
| }, |
| { |
| "epoch": 0.38721573448063923, |
| "grad_norm": 0.19554060697555542, |
| "learning_rate": 2.9529158529882023e-05, |
| "loss": 0.0148, |
| "step": 3780 |
| }, |
| { |
| "epoch": 0.388240114730588, |
| "grad_norm": 0.28613370656967163, |
| "learning_rate": 2.9525039456746677e-05, |
| "loss": 0.0175, |
| "step": 3790 |
| }, |
| { |
| "epoch": 0.3892644949805368, |
| "grad_norm": 0.18975043296813965, |
| "learning_rate": 2.952090273431877e-05, |
| "loss": 0.0183, |
| "step": 3800 |
| }, |
| { |
| "epoch": 0.39028887523048555, |
| "grad_norm": 0.3048374354839325, |
| "learning_rate": 2.9516748367624808e-05, |
| "loss": 0.0218, |
| "step": 3810 |
| }, |
| { |
| "epoch": 0.39131325548043433, |
| "grad_norm": 0.219102680683136, |
| "learning_rate": 2.951257636171274e-05, |
| "loss": 0.018, |
| "step": 3820 |
| }, |
| { |
| "epoch": 0.3923376357303831, |
| "grad_norm": 0.27185437083244324, |
| "learning_rate": 2.9508386721651944e-05, |
| "loss": 0.0171, |
| "step": 3830 |
| }, |
| { |
| "epoch": 0.3933620159803319, |
| "grad_norm": 0.4174872040748596, |
| "learning_rate": 2.9504179452533228e-05, |
| "loss": 0.0193, |
| "step": 3840 |
| }, |
| { |
| "epoch": 0.39438639623028066, |
| "grad_norm": 0.22567470371723175, |
| "learning_rate": 2.9499954559468826e-05, |
| "loss": 0.0153, |
| "step": 3850 |
| }, |
| { |
| "epoch": 0.3954107764802295, |
| "grad_norm": 0.4505351185798645, |
| "learning_rate": 2.9495712047592378e-05, |
| "loss": 0.0163, |
| "step": 3860 |
| }, |
| { |
| "epoch": 0.39643515673017826, |
| "grad_norm": 0.30136609077453613, |
| "learning_rate": 2.949145192205893e-05, |
| "loss": 0.0165, |
| "step": 3870 |
| }, |
| { |
| "epoch": 0.39745953698012704, |
| "grad_norm": 0.2322685569524765, |
| "learning_rate": 2.9487174188044953e-05, |
| "loss": 0.0164, |
| "step": 3880 |
| }, |
| { |
| "epoch": 0.3984839172300758, |
| "grad_norm": 0.24659836292266846, |
| "learning_rate": 2.9482878850748276e-05, |
| "loss": 0.0185, |
| "step": 3890 |
| }, |
| { |
| "epoch": 0.3995082974800246, |
| "grad_norm": 0.28548499941825867, |
| "learning_rate": 2.9478565915388157e-05, |
| "loss": 0.0177, |
| "step": 3900 |
| }, |
| { |
| "epoch": 0.40053267772997336, |
| "grad_norm": 0.30812445282936096, |
| "learning_rate": 2.9474235387205207e-05, |
| "loss": 0.019, |
| "step": 3910 |
| }, |
| { |
| "epoch": 0.40155705797992214, |
| "grad_norm": 0.29529473185539246, |
| "learning_rate": 2.946988727146143e-05, |
| "loss": 0.0179, |
| "step": 3920 |
| }, |
| { |
| "epoch": 0.4025814382298709, |
| "grad_norm": 0.20870371162891388, |
| "learning_rate": 2.9465521573440196e-05, |
| "loss": 0.0159, |
| "step": 3930 |
| }, |
| { |
| "epoch": 0.4036058184798197, |
| "grad_norm": 0.322835773229599, |
| "learning_rate": 2.946113829844624e-05, |
| "loss": 0.0174, |
| "step": 3940 |
| }, |
| { |
| "epoch": 0.40463019872976846, |
| "grad_norm": 0.1954292207956314, |
| "learning_rate": 2.945673745180566e-05, |
| "loss": 0.015, |
| "step": 3950 |
| }, |
| { |
| "epoch": 0.4056545789797173, |
| "grad_norm": 0.3592565655708313, |
| "learning_rate": 2.945231903886589e-05, |
| "loss": 0.0157, |
| "step": 3960 |
| }, |
| { |
| "epoch": 0.40667895922966607, |
| "grad_norm": 0.295998215675354, |
| "learning_rate": 2.9447883064995728e-05, |
| "loss": 0.0175, |
| "step": 3970 |
| }, |
| { |
| "epoch": 0.40770333947961485, |
| "grad_norm": 0.20739391446113586, |
| "learning_rate": 2.9443429535585288e-05, |
| "loss": 0.0159, |
| "step": 3980 |
| }, |
| { |
| "epoch": 0.4087277197295636, |
| "grad_norm": 0.32952404022216797, |
| "learning_rate": 2.9438958456046044e-05, |
| "loss": 0.0159, |
| "step": 3990 |
| }, |
| { |
| "epoch": 0.4097520999795124, |
| "grad_norm": 0.19460247457027435, |
| "learning_rate": 2.9434469831810765e-05, |
| "loss": 0.0192, |
| "step": 4000 |
| }, |
| { |
| "epoch": 0.41077648022946117, |
| "grad_norm": 0.6853150725364685, |
| "learning_rate": 2.9429963668333566e-05, |
| "loss": 0.0185, |
| "step": 4010 |
| }, |
| { |
| "epoch": 0.41180086047940995, |
| "grad_norm": 0.3533918857574463, |
| "learning_rate": 2.9425439971089845e-05, |
| "loss": 0.0172, |
| "step": 4020 |
| }, |
| { |
| "epoch": 0.4128252407293587, |
| "grad_norm": 0.2702498137950897, |
| "learning_rate": 2.9420898745576336e-05, |
| "loss": 0.0145, |
| "step": 4030 |
| }, |
| { |
| "epoch": 0.4138496209793075, |
| "grad_norm": 0.23313632607460022, |
| "learning_rate": 2.9416339997311044e-05, |
| "loss": 0.0177, |
| "step": 4040 |
| }, |
| { |
| "epoch": 0.4148740012292563, |
| "grad_norm": 0.2649673521518707, |
| "learning_rate": 2.941176373183329e-05, |
| "loss": 0.0201, |
| "step": 4050 |
| }, |
| { |
| "epoch": 0.4158983814792051, |
| "grad_norm": 0.3330475389957428, |
| "learning_rate": 2.9407169954703653e-05, |
| "loss": 0.0172, |
| "step": 4060 |
| }, |
| { |
| "epoch": 0.4169227617291539, |
| "grad_norm": 0.39190003275871277, |
| "learning_rate": 2.9402558671504018e-05, |
| "loss": 0.0183, |
| "step": 4070 |
| }, |
| { |
| "epoch": 0.41794714197910265, |
| "grad_norm": 0.4329347312450409, |
| "learning_rate": 2.9397929887837524e-05, |
| "loss": 0.0192, |
| "step": 4080 |
| }, |
| { |
| "epoch": 0.41897152222905143, |
| "grad_norm": 0.36494845151901245, |
| "learning_rate": 2.939328360932858e-05, |
| "loss": 0.0196, |
| "step": 4090 |
| }, |
| { |
| "epoch": 0.4199959024790002, |
| "grad_norm": 0.5593007802963257, |
| "learning_rate": 2.938861984162285e-05, |
| "loss": 0.0176, |
| "step": 4100 |
| }, |
| { |
| "epoch": 0.421020282728949, |
| "grad_norm": 0.46095019578933716, |
| "learning_rate": 2.938393859038726e-05, |
| "loss": 0.0181, |
| "step": 4110 |
| }, |
| { |
| "epoch": 0.42204466297889776, |
| "grad_norm": 0.17580917477607727, |
| "learning_rate": 2.9379239861309964e-05, |
| "loss": 0.0194, |
| "step": 4120 |
| }, |
| { |
| "epoch": 0.42306904322884653, |
| "grad_norm": 0.25666120648384094, |
| "learning_rate": 2.937452366010036e-05, |
| "loss": 0.0173, |
| "step": 4130 |
| }, |
| { |
| "epoch": 0.4240934234787953, |
| "grad_norm": 0.36593982577323914, |
| "learning_rate": 2.936978999248908e-05, |
| "loss": 0.0167, |
| "step": 4140 |
| }, |
| { |
| "epoch": 0.42511780372874414, |
| "grad_norm": 0.2073334902524948, |
| "learning_rate": 2.936503886422798e-05, |
| "loss": 0.0179, |
| "step": 4150 |
| }, |
| { |
| "epoch": 0.4261421839786929, |
| "grad_norm": 0.20939789712429047, |
| "learning_rate": 2.9360270281090122e-05, |
| "loss": 0.0163, |
| "step": 4160 |
| }, |
| { |
| "epoch": 0.4271665642286417, |
| "grad_norm": 0.16788826882839203, |
| "learning_rate": 2.935548424886979e-05, |
| "loss": 0.0175, |
| "step": 4170 |
| }, |
| { |
| "epoch": 0.42819094447859046, |
| "grad_norm": 0.637956976890564, |
| "learning_rate": 2.935068077338246e-05, |
| "loss": 0.0217, |
| "step": 4180 |
| }, |
| { |
| "epoch": 0.42921532472853924, |
| "grad_norm": 0.19704151153564453, |
| "learning_rate": 2.934585986046481e-05, |
| "loss": 0.0159, |
| "step": 4190 |
| }, |
| { |
| "epoch": 0.430239704978488, |
| "grad_norm": 0.28079935908317566, |
| "learning_rate": 2.93410215159747e-05, |
| "loss": 0.0164, |
| "step": 4200 |
| }, |
| { |
| "epoch": 0.4312640852284368, |
| "grad_norm": 0.15132665634155273, |
| "learning_rate": 2.933616574579118e-05, |
| "loss": 0.0173, |
| "step": 4210 |
| }, |
| { |
| "epoch": 0.43228846547838556, |
| "grad_norm": 0.2978609502315521, |
| "learning_rate": 2.933129255581447e-05, |
| "loss": 0.0202, |
| "step": 4220 |
| }, |
| { |
| "epoch": 0.43331284572833434, |
| "grad_norm": 0.23671455681324005, |
| "learning_rate": 2.9326401951965954e-05, |
| "loss": 0.0194, |
| "step": 4230 |
| }, |
| { |
| "epoch": 0.4343372259782831, |
| "grad_norm": 0.16392715275287628, |
| "learning_rate": 2.9321493940188165e-05, |
| "loss": 0.0209, |
| "step": 4240 |
| }, |
| { |
| "epoch": 0.43536160622823195, |
| "grad_norm": 0.2119358479976654, |
| "learning_rate": 2.931656852644482e-05, |
| "loss": 0.0227, |
| "step": 4250 |
| }, |
| { |
| "epoch": 0.4363859864781807, |
| "grad_norm": 0.29245108366012573, |
| "learning_rate": 2.9311625716720748e-05, |
| "loss": 0.0201, |
| "step": 4260 |
| }, |
| { |
| "epoch": 0.4374103667281295, |
| "grad_norm": 0.34422460198402405, |
| "learning_rate": 2.9306665517021927e-05, |
| "loss": 0.019, |
| "step": 4270 |
| }, |
| { |
| "epoch": 0.43843474697807827, |
| "grad_norm": 0.35365840792655945, |
| "learning_rate": 2.9301687933375483e-05, |
| "loss": 0.0191, |
| "step": 4280 |
| }, |
| { |
| "epoch": 0.43945912722802705, |
| "grad_norm": 0.3687003254890442, |
| "learning_rate": 2.929669297182963e-05, |
| "loss": 0.0164, |
| "step": 4290 |
| }, |
| { |
| "epoch": 0.4404835074779758, |
| "grad_norm": 0.43373772501945496, |
| "learning_rate": 2.929168063845373e-05, |
| "loss": 0.0195, |
| "step": 4300 |
| }, |
| { |
| "epoch": 0.4415078877279246, |
| "grad_norm": 0.5312246084213257, |
| "learning_rate": 2.9286650939338236e-05, |
| "loss": 0.0157, |
| "step": 4310 |
| }, |
| { |
| "epoch": 0.4425322679778734, |
| "grad_norm": 0.4340318441390991, |
| "learning_rate": 2.9281603880594706e-05, |
| "loss": 0.0168, |
| "step": 4320 |
| }, |
| { |
| "epoch": 0.44355664822782215, |
| "grad_norm": 0.3790690302848816, |
| "learning_rate": 2.9276539468355796e-05, |
| "loss": 0.0171, |
| "step": 4330 |
| }, |
| { |
| "epoch": 0.4445810284777709, |
| "grad_norm": 0.354899138212204, |
| "learning_rate": 2.9271457708775244e-05, |
| "loss": 0.0171, |
| "step": 4340 |
| }, |
| { |
| "epoch": 0.44560540872771975, |
| "grad_norm": 0.22029584646224976, |
| "learning_rate": 2.9266358608027863e-05, |
| "loss": 0.0185, |
| "step": 4350 |
| }, |
| { |
| "epoch": 0.44662978897766853, |
| "grad_norm": 0.2535483241081238, |
| "learning_rate": 2.9261242172309543e-05, |
| "loss": 0.0185, |
| "step": 4360 |
| }, |
| { |
| "epoch": 0.4476541692276173, |
| "grad_norm": 0.23264795541763306, |
| "learning_rate": 2.925610840783723e-05, |
| "loss": 0.0175, |
| "step": 4370 |
| }, |
| { |
| "epoch": 0.4486785494775661, |
| "grad_norm": 0.24278679490089417, |
| "learning_rate": 2.9250957320848946e-05, |
| "loss": 0.0158, |
| "step": 4380 |
| }, |
| { |
| "epoch": 0.44970292972751486, |
| "grad_norm": 0.2382252812385559, |
| "learning_rate": 2.924578891760374e-05, |
| "loss": 0.0183, |
| "step": 4390 |
| }, |
| { |
| "epoch": 0.45072730997746363, |
| "grad_norm": 0.18646247684955597, |
| "learning_rate": 2.924060320438171e-05, |
| "loss": 0.0151, |
| "step": 4400 |
| }, |
| { |
| "epoch": 0.4517516902274124, |
| "grad_norm": 0.26589062809944153, |
| "learning_rate": 2.9235400187483978e-05, |
| "loss": 0.0191, |
| "step": 4410 |
| }, |
| { |
| "epoch": 0.4527760704773612, |
| "grad_norm": 0.4926900863647461, |
| "learning_rate": 2.9230179873232716e-05, |
| "loss": 0.0199, |
| "step": 4420 |
| }, |
| { |
| "epoch": 0.45380045072730996, |
| "grad_norm": 0.41853833198547363, |
| "learning_rate": 2.922494226797108e-05, |
| "loss": 0.0171, |
| "step": 4430 |
| }, |
| { |
| "epoch": 0.45482483097725873, |
| "grad_norm": 0.4443478286266327, |
| "learning_rate": 2.9219687378063275e-05, |
| "loss": 0.0166, |
| "step": 4440 |
| }, |
| { |
| "epoch": 0.45584921122720756, |
| "grad_norm": 0.21750293672084808, |
| "learning_rate": 2.921441520989447e-05, |
| "loss": 0.0182, |
| "step": 4450 |
| }, |
| { |
| "epoch": 0.45687359147715634, |
| "grad_norm": 0.24466253817081451, |
| "learning_rate": 2.9209125769870854e-05, |
| "loss": 0.0148, |
| "step": 4460 |
| }, |
| { |
| "epoch": 0.4578979717271051, |
| "grad_norm": 0.16128693521022797, |
| "learning_rate": 2.92038190644196e-05, |
| "loss": 0.0156, |
| "step": 4470 |
| }, |
| { |
| "epoch": 0.4589223519770539, |
| "grad_norm": 0.3599347174167633, |
| "learning_rate": 2.9198495099988843e-05, |
| "loss": 0.0175, |
| "step": 4480 |
| }, |
| { |
| "epoch": 0.45994673222700266, |
| "grad_norm": 0.3112831711769104, |
| "learning_rate": 2.9193153883047712e-05, |
| "loss": 0.0167, |
| "step": 4490 |
| }, |
| { |
| "epoch": 0.46097111247695144, |
| "grad_norm": 0.10896246880292892, |
| "learning_rate": 2.9187795420086287e-05, |
| "loss": 0.0161, |
| "step": 4500 |
| }, |
| { |
| "epoch": 0.4619954927269002, |
| "grad_norm": 0.3054218590259552, |
| "learning_rate": 2.9182419717615602e-05, |
| "loss": 0.0167, |
| "step": 4510 |
| }, |
| { |
| "epoch": 0.463019872976849, |
| "grad_norm": 0.3468566834926605, |
| "learning_rate": 2.917702678216765e-05, |
| "loss": 0.0152, |
| "step": 4520 |
| }, |
| { |
| "epoch": 0.46404425322679776, |
| "grad_norm": 0.5659927725791931, |
| "learning_rate": 2.9171616620295347e-05, |
| "loss": 0.0209, |
| "step": 4530 |
| }, |
| { |
| "epoch": 0.4650686334767466, |
| "grad_norm": 0.1763537973165512, |
| "learning_rate": 2.9166189238572554e-05, |
| "loss": 0.0171, |
| "step": 4540 |
| }, |
| { |
| "epoch": 0.46609301372669537, |
| "grad_norm": 0.13917961716651917, |
| "learning_rate": 2.9160744643594052e-05, |
| "loss": 0.0146, |
| "step": 4550 |
| }, |
| { |
| "epoch": 0.46711739397664415, |
| "grad_norm": 0.19241924583911896, |
| "learning_rate": 2.915528284197554e-05, |
| "loss": 0.0159, |
| "step": 4560 |
| }, |
| { |
| "epoch": 0.4681417742265929, |
| "grad_norm": 0.3201506435871124, |
| "learning_rate": 2.914980384035362e-05, |
| "loss": 0.0169, |
| "step": 4570 |
| }, |
| { |
| "epoch": 0.4691661544765417, |
| "grad_norm": 0.4869763255119324, |
| "learning_rate": 2.9144307645385787e-05, |
| "loss": 0.015, |
| "step": 4580 |
| }, |
| { |
| "epoch": 0.47019053472649047, |
| "grad_norm": 0.16011469066143036, |
| "learning_rate": 2.9138794263750454e-05, |
| "loss": 0.0167, |
| "step": 4590 |
| }, |
| { |
| "epoch": 0.47121491497643925, |
| "grad_norm": 0.3395856022834778, |
| "learning_rate": 2.913326370214688e-05, |
| "loss": 0.0192, |
| "step": 4600 |
| }, |
| { |
| "epoch": 0.472239295226388, |
| "grad_norm": 0.25072088837623596, |
| "learning_rate": 2.9127715967295234e-05, |
| "loss": 0.0166, |
| "step": 4610 |
| }, |
| { |
| "epoch": 0.4732636754763368, |
| "grad_norm": 0.1807279884815216, |
| "learning_rate": 2.912215106593653e-05, |
| "loss": 0.0171, |
| "step": 4620 |
| }, |
| { |
| "epoch": 0.4742880557262856, |
| "grad_norm": 0.2525402307510376, |
| "learning_rate": 2.911656900483265e-05, |
| "loss": 0.0183, |
| "step": 4630 |
| }, |
| { |
| "epoch": 0.4753124359762344, |
| "grad_norm": 0.25578734278678894, |
| "learning_rate": 2.9110969790766322e-05, |
| "loss": 0.0151, |
| "step": 4640 |
| }, |
| { |
| "epoch": 0.4763368162261832, |
| "grad_norm": 0.4099701941013336, |
| "learning_rate": 2.910535343054112e-05, |
| "loss": 0.0155, |
| "step": 4650 |
| }, |
| { |
| "epoch": 0.47736119647613195, |
| "grad_norm": 0.5885595679283142, |
| "learning_rate": 2.909971993098145e-05, |
| "loss": 0.0157, |
| "step": 4660 |
| }, |
| { |
| "epoch": 0.47838557672608073, |
| "grad_norm": 0.44972121715545654, |
| "learning_rate": 2.9094069298932552e-05, |
| "loss": 0.0195, |
| "step": 4670 |
| }, |
| { |
| "epoch": 0.4794099569760295, |
| "grad_norm": 0.3503476679325104, |
| "learning_rate": 2.908840154126047e-05, |
| "loss": 0.0177, |
| "step": 4680 |
| }, |
| { |
| "epoch": 0.4804343372259783, |
| "grad_norm": 0.6346178650856018, |
| "learning_rate": 2.9082716664852063e-05, |
| "loss": 0.0171, |
| "step": 4690 |
| }, |
| { |
| "epoch": 0.48145871747592706, |
| "grad_norm": 0.48449280858039856, |
| "learning_rate": 2.9077014676615e-05, |
| "loss": 0.0185, |
| "step": 4700 |
| }, |
| { |
| "epoch": 0.48248309772587583, |
| "grad_norm": 0.29100608825683594, |
| "learning_rate": 2.9071295583477724e-05, |
| "loss": 0.0178, |
| "step": 4710 |
| }, |
| { |
| "epoch": 0.4835074779758246, |
| "grad_norm": 0.35782453417778015, |
| "learning_rate": 2.906555939238948e-05, |
| "loss": 0.0172, |
| "step": 4720 |
| }, |
| { |
| "epoch": 0.4845318582257734, |
| "grad_norm": 0.5882178544998169, |
| "learning_rate": 2.905980611032028e-05, |
| "loss": 0.0179, |
| "step": 4730 |
| }, |
| { |
| "epoch": 0.4855562384757222, |
| "grad_norm": 0.4882688522338867, |
| "learning_rate": 2.9054035744260904e-05, |
| "loss": 0.0165, |
| "step": 4740 |
| }, |
| { |
| "epoch": 0.486580618725671, |
| "grad_norm": 0.15563088655471802, |
| "learning_rate": 2.9048248301222885e-05, |
| "loss": 0.0172, |
| "step": 4750 |
| }, |
| { |
| "epoch": 0.48760499897561976, |
| "grad_norm": 0.25728586316108704, |
| "learning_rate": 2.904244378823852e-05, |
| "loss": 0.0151, |
| "step": 4760 |
| }, |
| { |
| "epoch": 0.48862937922556854, |
| "grad_norm": 0.36741533875465393, |
| "learning_rate": 2.9036622212360836e-05, |
| "loss": 0.0149, |
| "step": 4770 |
| }, |
| { |
| "epoch": 0.4896537594755173, |
| "grad_norm": 0.6343735456466675, |
| "learning_rate": 2.9030783580663596e-05, |
| "loss": 0.0174, |
| "step": 4780 |
| }, |
| { |
| "epoch": 0.4906781397254661, |
| "grad_norm": 0.28174135088920593, |
| "learning_rate": 2.902492790024129e-05, |
| "loss": 0.0172, |
| "step": 4790 |
| }, |
| { |
| "epoch": 0.49170251997541486, |
| "grad_norm": 0.2284751832485199, |
| "learning_rate": 2.901905517820912e-05, |
| "loss": 0.0199, |
| "step": 4800 |
| }, |
| { |
| "epoch": 0.49272690022536364, |
| "grad_norm": 0.16254670917987823, |
| "learning_rate": 2.9013165421703e-05, |
| "loss": 0.0134, |
| "step": 4810 |
| }, |
| { |
| "epoch": 0.4937512804753124, |
| "grad_norm": 0.1736937165260315, |
| "learning_rate": 2.9007258637879534e-05, |
| "loss": 0.0149, |
| "step": 4820 |
| }, |
| { |
| "epoch": 0.4947756607252612, |
| "grad_norm": 0.15401211380958557, |
| "learning_rate": 2.9001334833916025e-05, |
| "loss": 0.0177, |
| "step": 4830 |
| }, |
| { |
| "epoch": 0.49580004097521, |
| "grad_norm": 0.29670220613479614, |
| "learning_rate": 2.899539401701046e-05, |
| "loss": 0.0163, |
| "step": 4840 |
| }, |
| { |
| "epoch": 0.4968244212251588, |
| "grad_norm": 0.7458762526512146, |
| "learning_rate": 2.8989436194381474e-05, |
| "loss": 0.0143, |
| "step": 4850 |
| }, |
| { |
| "epoch": 0.49784880147510757, |
| "grad_norm": 0.23505784571170807, |
| "learning_rate": 2.8983461373268404e-05, |
| "loss": 0.0188, |
| "step": 4860 |
| }, |
| { |
| "epoch": 0.49887318172505635, |
| "grad_norm": 0.20350441336631775, |
| "learning_rate": 2.8977469560931205e-05, |
| "loss": 0.0139, |
| "step": 4870 |
| }, |
| { |
| "epoch": 0.4998975619750051, |
| "grad_norm": 0.1895584911108017, |
| "learning_rate": 2.8971460764650502e-05, |
| "loss": 0.0178, |
| "step": 4880 |
| }, |
| { |
| "epoch": 0.5009219422249539, |
| "grad_norm": 0.16561292111873627, |
| "learning_rate": 2.8965434991727545e-05, |
| "loss": 0.0168, |
| "step": 4890 |
| }, |
| { |
| "epoch": 0.5019463224749027, |
| "grad_norm": 0.1712399572134018, |
| "learning_rate": 2.895939224948422e-05, |
| "loss": 0.0172, |
| "step": 4900 |
| }, |
| { |
| "epoch": 0.5029707027248514, |
| "grad_norm": 0.44496676325798035, |
| "learning_rate": 2.895333254526303e-05, |
| "loss": 0.0182, |
| "step": 4910 |
| }, |
| { |
| "epoch": 0.5039950829748002, |
| "grad_norm": 0.2654223144054413, |
| "learning_rate": 2.8947255886427084e-05, |
| "loss": 0.0183, |
| "step": 4920 |
| }, |
| { |
| "epoch": 0.505019463224749, |
| "grad_norm": 0.2276906669139862, |
| "learning_rate": 2.8941162280360095e-05, |
| "loss": 0.0177, |
| "step": 4930 |
| }, |
| { |
| "epoch": 0.5060438434746978, |
| "grad_norm": 0.17274074256420135, |
| "learning_rate": 2.8935051734466368e-05, |
| "loss": 0.018, |
| "step": 4940 |
| }, |
| { |
| "epoch": 0.5070682237246465, |
| "grad_norm": 0.1987011432647705, |
| "learning_rate": 2.8928924256170792e-05, |
| "loss": 0.0148, |
| "step": 4950 |
| }, |
| { |
| "epoch": 0.5080926039745953, |
| "grad_norm": 0.6934794783592224, |
| "learning_rate": 2.8922779852918836e-05, |
| "loss": 0.0168, |
| "step": 4960 |
| }, |
| { |
| "epoch": 0.5091169842245441, |
| "grad_norm": 0.17846709489822388, |
| "learning_rate": 2.8916618532176526e-05, |
| "loss": 0.016, |
| "step": 4970 |
| }, |
| { |
| "epoch": 0.510141364474493, |
| "grad_norm": 0.15801241993904114, |
| "learning_rate": 2.891044030143045e-05, |
| "loss": 0.0179, |
| "step": 4980 |
| }, |
| { |
| "epoch": 0.5111657447244418, |
| "grad_norm": 0.36965882778167725, |
| "learning_rate": 2.890424516818774e-05, |
| "loss": 0.0163, |
| "step": 4990 |
| }, |
| { |
| "epoch": 0.5121901249743905, |
| "grad_norm": 0.43958860635757446, |
| "learning_rate": 2.8898033139976057e-05, |
| "loss": 0.0204, |
| "step": 5000 |
| }, |
| { |
| "epoch": 0.5132145052243393, |
| "grad_norm": 0.2637784481048584, |
| "learning_rate": 2.889180422434362e-05, |
| "loss": 0.0194, |
| "step": 5010 |
| }, |
| { |
| "epoch": 0.5142388854742881, |
| "grad_norm": 0.4712976813316345, |
| "learning_rate": 2.8885558428859136e-05, |
| "loss": 0.0151, |
| "step": 5020 |
| }, |
| { |
| "epoch": 0.5152632657242369, |
| "grad_norm": 0.23316746950149536, |
| "learning_rate": 2.8879295761111835e-05, |
| "loss": 0.0166, |
| "step": 5030 |
| }, |
| { |
| "epoch": 0.5162876459741856, |
| "grad_norm": 0.22744247317314148, |
| "learning_rate": 2.8873016228711452e-05, |
| "loss": 0.0185, |
| "step": 5040 |
| }, |
| { |
| "epoch": 0.5173120262241344, |
| "grad_norm": 0.3363291621208191, |
| "learning_rate": 2.8866719839288215e-05, |
| "loss": 0.0193, |
| "step": 5050 |
| }, |
| { |
| "epoch": 0.5183364064740832, |
| "grad_norm": 0.3926180601119995, |
| "learning_rate": 2.886040660049283e-05, |
| "loss": 0.0176, |
| "step": 5060 |
| }, |
| { |
| "epoch": 0.519360786724032, |
| "grad_norm": 0.26203733682632446, |
| "learning_rate": 2.8854076519996467e-05, |
| "loss": 0.0177, |
| "step": 5070 |
| }, |
| { |
| "epoch": 0.5203851669739807, |
| "grad_norm": 0.2122904509305954, |
| "learning_rate": 2.884772960549078e-05, |
| "loss": 0.0183, |
| "step": 5080 |
| }, |
| { |
| "epoch": 0.5214095472239295, |
| "grad_norm": 0.37301889061927795, |
| "learning_rate": 2.8841365864687867e-05, |
| "loss": 0.0194, |
| "step": 5090 |
| }, |
| { |
| "epoch": 0.5224339274738783, |
| "grad_norm": 0.27067720890045166, |
| "learning_rate": 2.8834985305320276e-05, |
| "loss": 0.0179, |
| "step": 5100 |
| }, |
| { |
| "epoch": 0.5234583077238271, |
| "grad_norm": 0.17240874469280243, |
| "learning_rate": 2.882858793514098e-05, |
| "loss": 0.0179, |
| "step": 5110 |
| }, |
| { |
| "epoch": 0.5244826879737758, |
| "grad_norm": 0.5584284067153931, |
| "learning_rate": 2.8822173761923398e-05, |
| "loss": 0.0174, |
| "step": 5120 |
| }, |
| { |
| "epoch": 0.5255070682237246, |
| "grad_norm": 0.19335246086120605, |
| "learning_rate": 2.8815742793461343e-05, |
| "loss": 0.0194, |
| "step": 5130 |
| }, |
| { |
| "epoch": 0.5265314484736734, |
| "grad_norm": 0.3710666298866272, |
| "learning_rate": 2.880929503756906e-05, |
| "loss": 0.0176, |
| "step": 5140 |
| }, |
| { |
| "epoch": 0.5275558287236222, |
| "grad_norm": 0.19493483006954193, |
| "learning_rate": 2.880283050208117e-05, |
| "loss": 0.0185, |
| "step": 5150 |
| }, |
| { |
| "epoch": 0.5285802089735709, |
| "grad_norm": 0.38253965973854065, |
| "learning_rate": 2.8796349194852704e-05, |
| "loss": 0.0189, |
| "step": 5160 |
| }, |
| { |
| "epoch": 0.5296045892235197, |
| "grad_norm": 0.22822563350200653, |
| "learning_rate": 2.8789851123759052e-05, |
| "loss": 0.018, |
| "step": 5170 |
| }, |
| { |
| "epoch": 0.5306289694734686, |
| "grad_norm": 0.24757550656795502, |
| "learning_rate": 2.8783336296695985e-05, |
| "loss": 0.0195, |
| "step": 5180 |
| }, |
| { |
| "epoch": 0.5316533497234174, |
| "grad_norm": 0.17331869900226593, |
| "learning_rate": 2.877680472157964e-05, |
| "loss": 0.0197, |
| "step": 5190 |
| }, |
| { |
| "epoch": 0.5326777299733662, |
| "grad_norm": 0.1980770230293274, |
| "learning_rate": 2.8770256406346483e-05, |
| "loss": 0.0177, |
| "step": 5200 |
| }, |
| { |
| "epoch": 0.5337021102233149, |
| "grad_norm": 0.22846157848834991, |
| "learning_rate": 2.8763691358953348e-05, |
| "loss": 0.0156, |
| "step": 5210 |
| }, |
| { |
| "epoch": 0.5347264904732637, |
| "grad_norm": 0.29557934403419495, |
| "learning_rate": 2.8757109587377377e-05, |
| "loss": 0.0178, |
| "step": 5220 |
| }, |
| { |
| "epoch": 0.5357508707232125, |
| "grad_norm": 0.24691875278949738, |
| "learning_rate": 2.8750511099616047e-05, |
| "loss": 0.0154, |
| "step": 5230 |
| }, |
| { |
| "epoch": 0.5367752509731613, |
| "grad_norm": 0.19662101566791534, |
| "learning_rate": 2.8743895903687143e-05, |
| "loss": 0.018, |
| "step": 5240 |
| }, |
| { |
| "epoch": 0.53779963122311, |
| "grad_norm": 0.17772042751312256, |
| "learning_rate": 2.8737264007628753e-05, |
| "loss": 0.0184, |
| "step": 5250 |
| }, |
| { |
| "epoch": 0.5388240114730588, |
| "grad_norm": 0.2527352273464203, |
| "learning_rate": 2.8730615419499247e-05, |
| "loss": 0.0182, |
| "step": 5260 |
| }, |
| { |
| "epoch": 0.5398483917230076, |
| "grad_norm": 0.49815821647644043, |
| "learning_rate": 2.8723950147377303e-05, |
| "loss": 0.0175, |
| "step": 5270 |
| }, |
| { |
| "epoch": 0.5408727719729564, |
| "grad_norm": 0.18781927227973938, |
| "learning_rate": 2.871726819936184e-05, |
| "loss": 0.0187, |
| "step": 5280 |
| }, |
| { |
| "epoch": 0.5418971522229051, |
| "grad_norm": 0.26046690344810486, |
| "learning_rate": 2.8710569583572068e-05, |
| "loss": 0.0184, |
| "step": 5290 |
| }, |
| { |
| "epoch": 0.5429215324728539, |
| "grad_norm": 0.19386866688728333, |
| "learning_rate": 2.870385430814742e-05, |
| "loss": 0.0173, |
| "step": 5300 |
| }, |
| { |
| "epoch": 0.5439459127228027, |
| "grad_norm": 0.3855115473270416, |
| "learning_rate": 2.8697122381247605e-05, |
| "loss": 0.0189, |
| "step": 5310 |
| }, |
| { |
| "epoch": 0.5449702929727515, |
| "grad_norm": 0.2893480062484741, |
| "learning_rate": 2.8690373811052545e-05, |
| "loss": 0.015, |
| "step": 5320 |
| }, |
| { |
| "epoch": 0.5459946732227002, |
| "grad_norm": 0.2906562089920044, |
| "learning_rate": 2.8683608605762378e-05, |
| "loss": 0.02, |
| "step": 5330 |
| }, |
| { |
| "epoch": 0.547019053472649, |
| "grad_norm": 0.5615838170051575, |
| "learning_rate": 2.8676826773597485e-05, |
| "loss": 0.0183, |
| "step": 5340 |
| }, |
| { |
| "epoch": 0.5480434337225978, |
| "grad_norm": 0.2854063808917999, |
| "learning_rate": 2.8670028322798413e-05, |
| "loss": 0.0174, |
| "step": 5350 |
| }, |
| { |
| "epoch": 0.5490678139725466, |
| "grad_norm": 0.2982889413833618, |
| "learning_rate": 2.8663213261625932e-05, |
| "loss": 0.0196, |
| "step": 5360 |
| }, |
| { |
| "epoch": 0.5500921942224954, |
| "grad_norm": 0.2706809937953949, |
| "learning_rate": 2.865638159836098e-05, |
| "loss": 0.0184, |
| "step": 5370 |
| }, |
| { |
| "epoch": 0.5511165744724442, |
| "grad_norm": 0.23803141713142395, |
| "learning_rate": 2.8649533341304677e-05, |
| "loss": 0.0141, |
| "step": 5380 |
| }, |
| { |
| "epoch": 0.552140954722393, |
| "grad_norm": 0.3107606768608093, |
| "learning_rate": 2.8642668498778297e-05, |
| "loss": 0.0158, |
| "step": 5390 |
| }, |
| { |
| "epoch": 0.5531653349723418, |
| "grad_norm": 0.16552206873893738, |
| "learning_rate": 2.8635787079123268e-05, |
| "loss": 0.016, |
| "step": 5400 |
| }, |
| { |
| "epoch": 0.5541897152222905, |
| "grad_norm": 0.2437811940908432, |
| "learning_rate": 2.8628889090701175e-05, |
| "loss": 0.0149, |
| "step": 5410 |
| }, |
| { |
| "epoch": 0.5552140954722393, |
| "grad_norm": 0.301443874835968, |
| "learning_rate": 2.862197454189371e-05, |
| "loss": 0.0168, |
| "step": 5420 |
| }, |
| { |
| "epoch": 0.5562384757221881, |
| "grad_norm": 0.32576826214790344, |
| "learning_rate": 2.8615043441102708e-05, |
| "loss": 0.0193, |
| "step": 5430 |
| }, |
| { |
| "epoch": 0.5572628559721369, |
| "grad_norm": 0.1850123554468155, |
| "learning_rate": 2.860809579675011e-05, |
| "loss": 0.0155, |
| "step": 5440 |
| }, |
| { |
| "epoch": 0.5582872362220856, |
| "grad_norm": 0.3755355775356293, |
| "learning_rate": 2.860113161727796e-05, |
| "loss": 0.0163, |
| "step": 5450 |
| }, |
| { |
| "epoch": 0.5593116164720344, |
| "grad_norm": 0.5295906066894531, |
| "learning_rate": 2.8594150911148393e-05, |
| "loss": 0.0161, |
| "step": 5460 |
| }, |
| { |
| "epoch": 0.5603359967219832, |
| "grad_norm": 0.8975720405578613, |
| "learning_rate": 2.8587153686843626e-05, |
| "loss": 0.0148, |
| "step": 5470 |
| }, |
| { |
| "epoch": 0.561360376971932, |
| "grad_norm": 0.33269113302230835, |
| "learning_rate": 2.858013995286594e-05, |
| "loss": 0.0201, |
| "step": 5480 |
| }, |
| { |
| "epoch": 0.5623847572218807, |
| "grad_norm": 0.21005471050739288, |
| "learning_rate": 2.857310971773769e-05, |
| "loss": 0.0201, |
| "step": 5490 |
| }, |
| { |
| "epoch": 0.5634091374718295, |
| "grad_norm": 0.3052314221858978, |
| "learning_rate": 2.8566062990001274e-05, |
| "loss": 0.0146, |
| "step": 5500 |
| }, |
| { |
| "epoch": 0.5644335177217783, |
| "grad_norm": 0.18676097691059113, |
| "learning_rate": 2.8558999778219126e-05, |
| "loss": 0.0145, |
| "step": 5510 |
| }, |
| { |
| "epoch": 0.5654578979717271, |
| "grad_norm": 0.192224383354187, |
| "learning_rate": 2.8551920090973722e-05, |
| "loss": 0.015, |
| "step": 5520 |
| }, |
| { |
| "epoch": 0.5664822782216759, |
| "grad_norm": 0.26617810130119324, |
| "learning_rate": 2.854482393686754e-05, |
| "loss": 0.0133, |
| "step": 5530 |
| }, |
| { |
| "epoch": 0.5675066584716246, |
| "grad_norm": 0.3860047161579132, |
| "learning_rate": 2.8537711324523086e-05, |
| "loss": 0.0155, |
| "step": 5540 |
| }, |
| { |
| "epoch": 0.5685310387215734, |
| "grad_norm": 0.17982348799705505, |
| "learning_rate": 2.8530582262582845e-05, |
| "loss": 0.0177, |
| "step": 5550 |
| }, |
| { |
| "epoch": 0.5695554189715222, |
| "grad_norm": 0.20987384021282196, |
| "learning_rate": 2.8523436759709306e-05, |
| "loss": 0.0141, |
| "step": 5560 |
| }, |
| { |
| "epoch": 0.5705797992214711, |
| "grad_norm": 0.23272131383419037, |
| "learning_rate": 2.8516274824584926e-05, |
| "loss": 0.0187, |
| "step": 5570 |
| }, |
| { |
| "epoch": 0.5716041794714198, |
| "grad_norm": 0.7056375741958618, |
| "learning_rate": 2.8509096465912136e-05, |
| "loss": 0.0177, |
| "step": 5580 |
| }, |
| { |
| "epoch": 0.5726285597213686, |
| "grad_norm": 0.18250614404678345, |
| "learning_rate": 2.8501901692413315e-05, |
| "loss": 0.0175, |
| "step": 5590 |
| }, |
| { |
| "epoch": 0.5736529399713174, |
| "grad_norm": 0.2899186909198761, |
| "learning_rate": 2.8494690512830788e-05, |
| "loss": 0.0174, |
| "step": 5600 |
| }, |
| { |
| "epoch": 0.5746773202212662, |
| "grad_norm": 0.43760666251182556, |
| "learning_rate": 2.8487462935926823e-05, |
| "loss": 0.0212, |
| "step": 5610 |
| }, |
| { |
| "epoch": 0.5757017004712149, |
| "grad_norm": 0.322891503572464, |
| "learning_rate": 2.8480218970483606e-05, |
| "loss": 0.0205, |
| "step": 5620 |
| }, |
| { |
| "epoch": 0.5767260807211637, |
| "grad_norm": 0.2490316927433014, |
| "learning_rate": 2.847295862530324e-05, |
| "loss": 0.0188, |
| "step": 5630 |
| }, |
| { |
| "epoch": 0.5777504609711125, |
| "grad_norm": 0.29211658239364624, |
| "learning_rate": 2.846568190920773e-05, |
| "loss": 0.0193, |
| "step": 5640 |
| }, |
| { |
| "epoch": 0.5787748412210613, |
| "grad_norm": 0.32587870955467224, |
| "learning_rate": 2.8458388831038965e-05, |
| "loss": 0.0201, |
| "step": 5650 |
| }, |
| { |
| "epoch": 0.57979922147101, |
| "grad_norm": 0.31571274995803833, |
| "learning_rate": 2.8451079399658725e-05, |
| "loss": 0.0166, |
| "step": 5660 |
| }, |
| { |
| "epoch": 0.5808236017209588, |
| "grad_norm": 0.2568550109863281, |
| "learning_rate": 2.8443753623948665e-05, |
| "loss": 0.0177, |
| "step": 5670 |
| }, |
| { |
| "epoch": 0.5818479819709076, |
| "grad_norm": 0.22321592271327972, |
| "learning_rate": 2.8436411512810288e-05, |
| "loss": 0.0162, |
| "step": 5680 |
| }, |
| { |
| "epoch": 0.5828723622208564, |
| "grad_norm": 0.19599325954914093, |
| "learning_rate": 2.842905307516495e-05, |
| "loss": 0.0166, |
| "step": 5690 |
| }, |
| { |
| "epoch": 0.5838967424708051, |
| "grad_norm": 0.38471075892448425, |
| "learning_rate": 2.8421678319953854e-05, |
| "loss": 0.0164, |
| "step": 5700 |
| }, |
| { |
| "epoch": 0.5849211227207539, |
| "grad_norm": 0.3636429011821747, |
| "learning_rate": 2.841428725613801e-05, |
| "loss": 0.0148, |
| "step": 5710 |
| }, |
| { |
| "epoch": 0.5859455029707027, |
| "grad_norm": 0.36383137106895447, |
| "learning_rate": 2.8406879892698272e-05, |
| "loss": 0.0172, |
| "step": 5720 |
| }, |
| { |
| "epoch": 0.5869698832206515, |
| "grad_norm": 0.29027339816093445, |
| "learning_rate": 2.8399456238635274e-05, |
| "loss": 0.0161, |
| "step": 5730 |
| }, |
| { |
| "epoch": 0.5879942634706002, |
| "grad_norm": 0.23695923388004303, |
| "learning_rate": 2.8392016302969457e-05, |
| "loss": 0.014, |
| "step": 5740 |
| }, |
| { |
| "epoch": 0.589018643720549, |
| "grad_norm": 0.31538084149360657, |
| "learning_rate": 2.8384560094741055e-05, |
| "loss": 0.0154, |
| "step": 5750 |
| }, |
| { |
| "epoch": 0.5900430239704979, |
| "grad_norm": 0.287030965089798, |
| "learning_rate": 2.8377087623010054e-05, |
| "loss": 0.0165, |
| "step": 5760 |
| }, |
| { |
| "epoch": 0.5910674042204467, |
| "grad_norm": 0.536839485168457, |
| "learning_rate": 2.8369598896856213e-05, |
| "loss": 0.017, |
| "step": 5770 |
| }, |
| { |
| "epoch": 0.5920917844703955, |
| "grad_norm": 0.993626058101654, |
| "learning_rate": 2.836209392537904e-05, |
| "loss": 0.0167, |
| "step": 5780 |
| }, |
| { |
| "epoch": 0.5931161647203442, |
| "grad_norm": 0.174278125166893, |
| "learning_rate": 2.8354572717697787e-05, |
| "loss": 0.0155, |
| "step": 5790 |
| }, |
| { |
| "epoch": 0.594140544970293, |
| "grad_norm": 0.40083181858062744, |
| "learning_rate": 2.8347035282951428e-05, |
| "loss": 0.0156, |
| "step": 5800 |
| }, |
| { |
| "epoch": 0.5951649252202418, |
| "grad_norm": 0.25317567586898804, |
| "learning_rate": 2.8339481630298653e-05, |
| "loss": 0.017, |
| "step": 5810 |
| }, |
| { |
| "epoch": 0.5961893054701906, |
| "grad_norm": 0.7192391157150269, |
| "learning_rate": 2.833191176891787e-05, |
| "loss": 0.0162, |
| "step": 5820 |
| }, |
| { |
| "epoch": 0.5972136857201393, |
| "grad_norm": 0.33481308817863464, |
| "learning_rate": 2.832432570800717e-05, |
| "loss": 0.0189, |
| "step": 5830 |
| }, |
| { |
| "epoch": 0.5982380659700881, |
| "grad_norm": 0.47620871663093567, |
| "learning_rate": 2.8316723456784332e-05, |
| "loss": 0.0175, |
| "step": 5840 |
| }, |
| { |
| "epoch": 0.5992624462200369, |
| "grad_norm": 0.32609736919403076, |
| "learning_rate": 2.8309105024486808e-05, |
| "loss": 0.0168, |
| "step": 5850 |
| }, |
| { |
| "epoch": 0.6002868264699857, |
| "grad_norm": 0.2334146797657013, |
| "learning_rate": 2.830147042037172e-05, |
| "loss": 0.0179, |
| "step": 5860 |
| }, |
| { |
| "epoch": 0.6013112067199344, |
| "grad_norm": 0.3953694999217987, |
| "learning_rate": 2.8293819653715816e-05, |
| "loss": 0.0192, |
| "step": 5870 |
| }, |
| { |
| "epoch": 0.6023355869698832, |
| "grad_norm": 0.30415719747543335, |
| "learning_rate": 2.828615273381551e-05, |
| "loss": 0.0169, |
| "step": 5880 |
| }, |
| { |
| "epoch": 0.603359967219832, |
| "grad_norm": 0.22127294540405273, |
| "learning_rate": 2.827846966998683e-05, |
| "loss": 0.0154, |
| "step": 5890 |
| }, |
| { |
| "epoch": 0.6043843474697808, |
| "grad_norm": 0.2622010409832001, |
| "learning_rate": 2.8270770471565422e-05, |
| "loss": 0.0169, |
| "step": 5900 |
| }, |
| { |
| "epoch": 0.6054087277197295, |
| "grad_norm": 0.21714642643928528, |
| "learning_rate": 2.826305514790654e-05, |
| "loss": 0.0175, |
| "step": 5910 |
| }, |
| { |
| "epoch": 0.6064331079696783, |
| "grad_norm": 0.20916715264320374, |
| "learning_rate": 2.825532370838503e-05, |
| "loss": 0.0166, |
| "step": 5920 |
| }, |
| { |
| "epoch": 0.6074574882196271, |
| "grad_norm": 0.32636559009552, |
| "learning_rate": 2.824757616239532e-05, |
| "loss": 0.0173, |
| "step": 5930 |
| }, |
| { |
| "epoch": 0.6084818684695759, |
| "grad_norm": 0.3978821933269501, |
| "learning_rate": 2.8239812519351404e-05, |
| "loss": 0.0182, |
| "step": 5940 |
| }, |
| { |
| "epoch": 0.6095062487195246, |
| "grad_norm": 0.28152239322662354, |
| "learning_rate": 2.823203278868685e-05, |
| "loss": 0.0205, |
| "step": 5950 |
| }, |
| { |
| "epoch": 0.6105306289694735, |
| "grad_norm": 0.42363396286964417, |
| "learning_rate": 2.8224236979854752e-05, |
| "loss": 0.021, |
| "step": 5960 |
| }, |
| { |
| "epoch": 0.6115550092194223, |
| "grad_norm": 0.2511829435825348, |
| "learning_rate": 2.821642510232776e-05, |
| "loss": 0.0199, |
| "step": 5970 |
| }, |
| { |
| "epoch": 0.6125793894693711, |
| "grad_norm": 0.250808984041214, |
| "learning_rate": 2.8208597165598046e-05, |
| "loss": 0.0215, |
| "step": 5980 |
| }, |
| { |
| "epoch": 0.6136037697193198, |
| "grad_norm": 0.36321577429771423, |
| "learning_rate": 2.8200753179177285e-05, |
| "loss": 0.0193, |
| "step": 5990 |
| }, |
| { |
| "epoch": 0.6146281499692686, |
| "grad_norm": 0.28019794821739197, |
| "learning_rate": 2.819289315259666e-05, |
| "loss": 0.0185, |
| "step": 6000 |
| }, |
| { |
| "epoch": 0.6156525302192174, |
| "grad_norm": 0.32472798228263855, |
| "learning_rate": 2.8185017095406845e-05, |
| "loss": 0.016, |
| "step": 6010 |
| }, |
| { |
| "epoch": 0.6166769104691662, |
| "grad_norm": 0.2541959285736084, |
| "learning_rate": 2.8177125017177998e-05, |
| "loss": 0.0147, |
| "step": 6020 |
| }, |
| { |
| "epoch": 0.617701290719115, |
| "grad_norm": 0.2614126205444336, |
| "learning_rate": 2.8169216927499727e-05, |
| "loss": 0.0167, |
| "step": 6030 |
| }, |
| { |
| "epoch": 0.6187256709690637, |
| "grad_norm": 0.24808119237422943, |
| "learning_rate": 2.816129283598112e-05, |
| "loss": 0.0173, |
| "step": 6040 |
| }, |
| { |
| "epoch": 0.6197500512190125, |
| "grad_norm": 0.5369676947593689, |
| "learning_rate": 2.815335275225068e-05, |
| "loss": 0.0172, |
| "step": 6050 |
| }, |
| { |
| "epoch": 0.6207744314689613, |
| "grad_norm": 0.2833543121814728, |
| "learning_rate": 2.8145396685956374e-05, |
| "loss": 0.0158, |
| "step": 6060 |
| }, |
| { |
| "epoch": 0.62179881171891, |
| "grad_norm": 0.17974276840686798, |
| "learning_rate": 2.813742464676556e-05, |
| "loss": 0.0159, |
| "step": 6070 |
| }, |
| { |
| "epoch": 0.6228231919688588, |
| "grad_norm": 0.3221525549888611, |
| "learning_rate": 2.8129436644365025e-05, |
| "loss": 0.0148, |
| "step": 6080 |
| }, |
| { |
| "epoch": 0.6238475722188076, |
| "grad_norm": 0.47700434923171997, |
| "learning_rate": 2.8121432688460938e-05, |
| "loss": 0.0178, |
| "step": 6090 |
| }, |
| { |
| "epoch": 0.6248719524687564, |
| "grad_norm": 0.3448154330253601, |
| "learning_rate": 2.8113412788778867e-05, |
| "loss": 0.0147, |
| "step": 6100 |
| }, |
| { |
| "epoch": 0.6258963327187052, |
| "grad_norm": 0.35500189661979675, |
| "learning_rate": 2.8105376955063744e-05, |
| "loss": 0.0179, |
| "step": 6110 |
| }, |
| { |
| "epoch": 0.6269207129686539, |
| "grad_norm": 0.2931915521621704, |
| "learning_rate": 2.8097325197079862e-05, |
| "loss": 0.0195, |
| "step": 6120 |
| }, |
| { |
| "epoch": 0.6279450932186027, |
| "grad_norm": 0.40634453296661377, |
| "learning_rate": 2.808925752461087e-05, |
| "loss": 0.0119, |
| "step": 6130 |
| }, |
| { |
| "epoch": 0.6289694734685515, |
| "grad_norm": 0.41018009185791016, |
| "learning_rate": 2.808117394745975e-05, |
| "loss": 0.0164, |
| "step": 6140 |
| }, |
| { |
| "epoch": 0.6299938537185003, |
| "grad_norm": 0.26122966408729553, |
| "learning_rate": 2.807307447544881e-05, |
| "loss": 0.018, |
| "step": 6150 |
| }, |
| { |
| "epoch": 0.6310182339684491, |
| "grad_norm": 0.27016958594322205, |
| "learning_rate": 2.806495911841967e-05, |
| "loss": 0.0158, |
| "step": 6160 |
| }, |
| { |
| "epoch": 0.6320426142183979, |
| "grad_norm": 0.16065537929534912, |
| "learning_rate": 2.8056827886233255e-05, |
| "loss": 0.017, |
| "step": 6170 |
| }, |
| { |
| "epoch": 0.6330669944683467, |
| "grad_norm": 0.32265228033065796, |
| "learning_rate": 2.8048680788769783e-05, |
| "loss": 0.0158, |
| "step": 6180 |
| }, |
| { |
| "epoch": 0.6340913747182955, |
| "grad_norm": 0.3483600914478302, |
| "learning_rate": 2.8040517835928742e-05, |
| "loss": 0.0157, |
| "step": 6190 |
| }, |
| { |
| "epoch": 0.6351157549682442, |
| "grad_norm": 0.1955910325050354, |
| "learning_rate": 2.803233903762889e-05, |
| "loss": 0.0154, |
| "step": 6200 |
| }, |
| { |
| "epoch": 0.636140135218193, |
| "grad_norm": 0.3537892997264862, |
| "learning_rate": 2.8024144403808235e-05, |
| "loss": 0.0173, |
| "step": 6210 |
| }, |
| { |
| "epoch": 0.6371645154681418, |
| "grad_norm": 0.5894239544868469, |
| "learning_rate": 2.8015933944424037e-05, |
| "loss": 0.0142, |
| "step": 6220 |
| }, |
| { |
| "epoch": 0.6381888957180906, |
| "grad_norm": 0.16838586330413818, |
| "learning_rate": 2.800770766945277e-05, |
| "loss": 0.0141, |
| "step": 6230 |
| }, |
| { |
| "epoch": 0.6392132759680393, |
| "grad_norm": 0.2810184955596924, |
| "learning_rate": 2.7999465588890135e-05, |
| "loss": 0.0171, |
| "step": 6240 |
| }, |
| { |
| "epoch": 0.6402376562179881, |
| "grad_norm": 0.3369508683681488, |
| "learning_rate": 2.799120771275105e-05, |
| "loss": 0.0159, |
| "step": 6250 |
| }, |
| { |
| "epoch": 0.6412620364679369, |
| "grad_norm": 0.4554113447666168, |
| "learning_rate": 2.7982934051069597e-05, |
| "loss": 0.0151, |
| "step": 6260 |
| }, |
| { |
| "epoch": 0.6422864167178857, |
| "grad_norm": 0.1880011409521103, |
| "learning_rate": 2.7974644613899063e-05, |
| "loss": 0.016, |
| "step": 6270 |
| }, |
| { |
| "epoch": 0.6433107969678344, |
| "grad_norm": 0.25223255157470703, |
| "learning_rate": 2.796633941131189e-05, |
| "loss": 0.0148, |
| "step": 6280 |
| }, |
| { |
| "epoch": 0.6443351772177832, |
| "grad_norm": 0.8673241138458252, |
| "learning_rate": 2.795801845339969e-05, |
| "loss": 0.0182, |
| "step": 6290 |
| }, |
| { |
| "epoch": 0.645359557467732, |
| "grad_norm": 0.2727773189544678, |
| "learning_rate": 2.7949681750273208e-05, |
| "loss": 0.0153, |
| "step": 6300 |
| }, |
| { |
| "epoch": 0.6463839377176808, |
| "grad_norm": 0.200203537940979, |
| "learning_rate": 2.7941329312062322e-05, |
| "loss": 0.0173, |
| "step": 6310 |
| }, |
| { |
| "epoch": 0.6474083179676295, |
| "grad_norm": 0.24331219494342804, |
| "learning_rate": 2.7932961148916037e-05, |
| "loss": 0.0158, |
| "step": 6320 |
| }, |
| { |
| "epoch": 0.6484326982175783, |
| "grad_norm": 0.39296358823776245, |
| "learning_rate": 2.7924577271002456e-05, |
| "loss": 0.0203, |
| "step": 6330 |
| }, |
| { |
| "epoch": 0.6494570784675271, |
| "grad_norm": 0.24977879226207733, |
| "learning_rate": 2.7916177688508782e-05, |
| "loss": 0.0159, |
| "step": 6340 |
| }, |
| { |
| "epoch": 0.650481458717476, |
| "grad_norm": 0.6776654720306396, |
| "learning_rate": 2.7907762411641302e-05, |
| "loss": 0.0199, |
| "step": 6350 |
| }, |
| { |
| "epoch": 0.6515058389674248, |
| "grad_norm": 0.23866847157478333, |
| "learning_rate": 2.7899331450625367e-05, |
| "loss": 0.0133, |
| "step": 6360 |
| }, |
| { |
| "epoch": 0.6525302192173735, |
| "grad_norm": 0.39472654461860657, |
| "learning_rate": 2.7890884815705397e-05, |
| "loss": 0.0165, |
| "step": 6370 |
| }, |
| { |
| "epoch": 0.6535545994673223, |
| "grad_norm": 0.18122504651546478, |
| "learning_rate": 2.788242251714484e-05, |
| "loss": 0.0214, |
| "step": 6380 |
| }, |
| { |
| "epoch": 0.6545789797172711, |
| "grad_norm": 0.3844212591648102, |
| "learning_rate": 2.787394456522619e-05, |
| "loss": 0.0154, |
| "step": 6390 |
| }, |
| { |
| "epoch": 0.6556033599672199, |
| "grad_norm": 0.16730929911136627, |
| "learning_rate": 2.7865450970250965e-05, |
| "loss": 0.0126, |
| "step": 6400 |
| }, |
| { |
| "epoch": 0.6566277402171686, |
| "grad_norm": 0.24835821986198425, |
| "learning_rate": 2.7856941742539676e-05, |
| "loss": 0.0176, |
| "step": 6410 |
| }, |
| { |
| "epoch": 0.6576521204671174, |
| "grad_norm": 0.1991119682788849, |
| "learning_rate": 2.784841689243184e-05, |
| "loss": 0.0166, |
| "step": 6420 |
| }, |
| { |
| "epoch": 0.6586765007170662, |
| "grad_norm": 0.31169748306274414, |
| "learning_rate": 2.783987643028595e-05, |
| "loss": 0.0165, |
| "step": 6430 |
| }, |
| { |
| "epoch": 0.659700880967015, |
| "grad_norm": 0.17884479463100433, |
| "learning_rate": 2.783132036647948e-05, |
| "loss": 0.018, |
| "step": 6440 |
| }, |
| { |
| "epoch": 0.6607252612169637, |
| "grad_norm": 0.2945181429386139, |
| "learning_rate": 2.7822748711408852e-05, |
| "loss": 0.0163, |
| "step": 6450 |
| }, |
| { |
| "epoch": 0.6617496414669125, |
| "grad_norm": 0.14600443840026855, |
| "learning_rate": 2.7814161475489428e-05, |
| "loss": 0.0157, |
| "step": 6460 |
| }, |
| { |
| "epoch": 0.6627740217168613, |
| "grad_norm": 0.42704740166664124, |
| "learning_rate": 2.780555866915552e-05, |
| "loss": 0.0155, |
| "step": 6470 |
| }, |
| { |
| "epoch": 0.6637984019668101, |
| "grad_norm": 0.21523109078407288, |
| "learning_rate": 2.7796940302860338e-05, |
| "loss": 0.0167, |
| "step": 6480 |
| }, |
| { |
| "epoch": 0.6648227822167588, |
| "grad_norm": 0.1869623064994812, |
| "learning_rate": 2.7788306387076018e-05, |
| "loss": 0.0191, |
| "step": 6490 |
| }, |
| { |
| "epoch": 0.6658471624667076, |
| "grad_norm": 0.20242300629615784, |
| "learning_rate": 2.7779656932293584e-05, |
| "loss": 0.0178, |
| "step": 6500 |
| }, |
| { |
| "epoch": 0.6668715427166564, |
| "grad_norm": 0.22725418210029602, |
| "learning_rate": 2.777099194902293e-05, |
| "loss": 0.0144, |
| "step": 6510 |
| }, |
| { |
| "epoch": 0.6678959229666052, |
| "grad_norm": 0.21521858870983124, |
| "learning_rate": 2.776231144779283e-05, |
| "loss": 0.0182, |
| "step": 6520 |
| }, |
| { |
| "epoch": 0.6689203032165539, |
| "grad_norm": 1.017340064048767, |
| "learning_rate": 2.775361543915091e-05, |
| "loss": 0.0188, |
| "step": 6530 |
| }, |
| { |
| "epoch": 0.6699446834665027, |
| "grad_norm": 0.20178307592868805, |
| "learning_rate": 2.774490393366365e-05, |
| "loss": 0.0176, |
| "step": 6540 |
| }, |
| { |
| "epoch": 0.6709690637164516, |
| "grad_norm": 0.3553648591041565, |
| "learning_rate": 2.7736176941916346e-05, |
| "loss": 0.0175, |
| "step": 6550 |
| }, |
| { |
| "epoch": 0.6719934439664004, |
| "grad_norm": 0.27946317195892334, |
| "learning_rate": 2.772743447451311e-05, |
| "loss": 0.0152, |
| "step": 6560 |
| }, |
| { |
| "epoch": 0.6730178242163491, |
| "grad_norm": 0.37165507674217224, |
| "learning_rate": 2.7718676542076872e-05, |
| "loss": 0.0161, |
| "step": 6570 |
| }, |
| { |
| "epoch": 0.6740422044662979, |
| "grad_norm": 0.4469502568244934, |
| "learning_rate": 2.770990315524934e-05, |
| "loss": 0.0151, |
| "step": 6580 |
| }, |
| { |
| "epoch": 0.6750665847162467, |
| "grad_norm": 0.15135547518730164, |
| "learning_rate": 2.7701114324691015e-05, |
| "loss": 0.0171, |
| "step": 6590 |
| }, |
| { |
| "epoch": 0.6760909649661955, |
| "grad_norm": 0.2614862024784088, |
| "learning_rate": 2.7692310061081145e-05, |
| "loss": 0.0143, |
| "step": 6600 |
| }, |
| { |
| "epoch": 0.6771153452161442, |
| "grad_norm": 0.29889607429504395, |
| "learning_rate": 2.7683490375117742e-05, |
| "loss": 0.0163, |
| "step": 6610 |
| }, |
| { |
| "epoch": 0.678139725466093, |
| "grad_norm": 0.1480155885219574, |
| "learning_rate": 2.7674655277517556e-05, |
| "loss": 0.0178, |
| "step": 6620 |
| }, |
| { |
| "epoch": 0.6791641057160418, |
| "grad_norm": 0.3905628025531769, |
| "learning_rate": 2.7665804779016072e-05, |
| "loss": 0.0214, |
| "step": 6630 |
| }, |
| { |
| "epoch": 0.6801884859659906, |
| "grad_norm": 0.25485795736312866, |
| "learning_rate": 2.765693889036747e-05, |
| "loss": 0.015, |
| "step": 6640 |
| }, |
| { |
| "epoch": 0.6812128662159393, |
| "grad_norm": 0.24822254478931427, |
| "learning_rate": 2.764805762234465e-05, |
| "loss": 0.0174, |
| "step": 6650 |
| }, |
| { |
| "epoch": 0.6822372464658881, |
| "grad_norm": 0.2817930579185486, |
| "learning_rate": 2.7639160985739182e-05, |
| "loss": 0.0189, |
| "step": 6660 |
| }, |
| { |
| "epoch": 0.6832616267158369, |
| "grad_norm": 0.24020500481128693, |
| "learning_rate": 2.7630248991361317e-05, |
| "loss": 0.0149, |
| "step": 6670 |
| }, |
| { |
| "epoch": 0.6842860069657857, |
| "grad_norm": 0.3694416284561157, |
| "learning_rate": 2.762132165003998e-05, |
| "loss": 0.0162, |
| "step": 6680 |
| }, |
| { |
| "epoch": 0.6853103872157345, |
| "grad_norm": 0.19866929948329926, |
| "learning_rate": 2.7612378972622722e-05, |
| "loss": 0.0154, |
| "step": 6690 |
| }, |
| { |
| "epoch": 0.6863347674656832, |
| "grad_norm": 0.32851651310920715, |
| "learning_rate": 2.7603420969975743e-05, |
| "loss": 0.0142, |
| "step": 6700 |
| }, |
| { |
| "epoch": 0.687359147715632, |
| "grad_norm": 0.3275679349899292, |
| "learning_rate": 2.7594447652983862e-05, |
| "loss": 0.0186, |
| "step": 6710 |
| }, |
| { |
| "epoch": 0.6883835279655808, |
| "grad_norm": 0.2732580900192261, |
| "learning_rate": 2.7585459032550497e-05, |
| "loss": 0.015, |
| "step": 6720 |
| }, |
| { |
| "epoch": 0.6894079082155296, |
| "grad_norm": 0.229897141456604, |
| "learning_rate": 2.7576455119597684e-05, |
| "loss": 0.0145, |
| "step": 6730 |
| }, |
| { |
| "epoch": 0.6904322884654784, |
| "grad_norm": 0.19596907496452332, |
| "learning_rate": 2.7567435925066012e-05, |
| "loss": 0.015, |
| "step": 6740 |
| }, |
| { |
| "epoch": 0.6914566687154272, |
| "grad_norm": 0.6001933813095093, |
| "learning_rate": 2.7558401459914665e-05, |
| "loss": 0.0142, |
| "step": 6750 |
| }, |
| { |
| "epoch": 0.692481048965376, |
| "grad_norm": 0.4691894054412842, |
| "learning_rate": 2.7549351735121362e-05, |
| "loss": 0.0189, |
| "step": 6760 |
| }, |
| { |
| "epoch": 0.6935054292153248, |
| "grad_norm": 0.23567530512809753, |
| "learning_rate": 2.7540286761682377e-05, |
| "loss": 0.0175, |
| "step": 6770 |
| }, |
| { |
| "epoch": 0.6945298094652735, |
| "grad_norm": 0.25450170040130615, |
| "learning_rate": 2.75312065506125e-05, |
| "loss": 0.0166, |
| "step": 6780 |
| }, |
| { |
| "epoch": 0.6955541897152223, |
| "grad_norm": 0.37370869517326355, |
| "learning_rate": 2.7522111112945052e-05, |
| "loss": 0.0163, |
| "step": 6790 |
| }, |
| { |
| "epoch": 0.6965785699651711, |
| "grad_norm": 0.1539376825094223, |
| "learning_rate": 2.751300045973184e-05, |
| "loss": 0.0157, |
| "step": 6800 |
| }, |
| { |
| "epoch": 0.6976029502151199, |
| "grad_norm": 0.13564033806324005, |
| "learning_rate": 2.750387460204317e-05, |
| "loss": 0.0149, |
| "step": 6810 |
| }, |
| { |
| "epoch": 0.6986273304650686, |
| "grad_norm": 0.198135644197464, |
| "learning_rate": 2.7494733550967827e-05, |
| "loss": 0.0184, |
| "step": 6820 |
| }, |
| { |
| "epoch": 0.6996517107150174, |
| "grad_norm": 0.7528253197669983, |
| "learning_rate": 2.748557731761304e-05, |
| "loss": 0.0176, |
| "step": 6830 |
| }, |
| { |
| "epoch": 0.7006760909649662, |
| "grad_norm": 0.2603262662887573, |
| "learning_rate": 2.74764059131045e-05, |
| "loss": 0.015, |
| "step": 6840 |
| }, |
| { |
| "epoch": 0.701700471214915, |
| "grad_norm": 1.304443120956421, |
| "learning_rate": 2.7467219348586322e-05, |
| "loss": 0.0168, |
| "step": 6850 |
| }, |
| { |
| "epoch": 0.7027248514648637, |
| "grad_norm": 0.19223232567310333, |
| "learning_rate": 2.7458017635221056e-05, |
| "loss": 0.0153, |
| "step": 6860 |
| }, |
| { |
| "epoch": 0.7037492317148125, |
| "grad_norm": 0.1936415284872055, |
| "learning_rate": 2.7448800784189652e-05, |
| "loss": 0.0177, |
| "step": 6870 |
| }, |
| { |
| "epoch": 0.7047736119647613, |
| "grad_norm": 0.31571337580680847, |
| "learning_rate": 2.7439568806691448e-05, |
| "loss": 0.0136, |
| "step": 6880 |
| }, |
| { |
| "epoch": 0.7057979922147101, |
| "grad_norm": 0.26796650886535645, |
| "learning_rate": 2.7430321713944167e-05, |
| "loss": 0.0162, |
| "step": 6890 |
| }, |
| { |
| "epoch": 0.7068223724646588, |
| "grad_norm": 0.2187744826078415, |
| "learning_rate": 2.7421059517183906e-05, |
| "loss": 0.0157, |
| "step": 6900 |
| }, |
| { |
| "epoch": 0.7078467527146076, |
| "grad_norm": 1.2276657819747925, |
| "learning_rate": 2.7411782227665097e-05, |
| "loss": 0.015, |
| "step": 6910 |
| }, |
| { |
| "epoch": 0.7088711329645564, |
| "grad_norm": 0.19194194674491882, |
| "learning_rate": 2.7402489856660528e-05, |
| "loss": 0.0171, |
| "step": 6920 |
| }, |
| { |
| "epoch": 0.7098955132145052, |
| "grad_norm": 0.5658370852470398, |
| "learning_rate": 2.7393182415461307e-05, |
| "loss": 0.016, |
| "step": 6930 |
| }, |
| { |
| "epoch": 0.7109198934644541, |
| "grad_norm": 0.34448087215423584, |
| "learning_rate": 2.7383859915376853e-05, |
| "loss": 0.0188, |
| "step": 6940 |
| }, |
| { |
| "epoch": 0.7119442737144028, |
| "grad_norm": 0.23329411447048187, |
| "learning_rate": 2.7374522367734875e-05, |
| "loss": 0.0177, |
| "step": 6950 |
| }, |
| { |
| "epoch": 0.7129686539643516, |
| "grad_norm": 0.1834026575088501, |
| "learning_rate": 2.7365169783881378e-05, |
| "loss": 0.0161, |
| "step": 6960 |
| }, |
| { |
| "epoch": 0.7139930342143004, |
| "grad_norm": 0.2726532816886902, |
| "learning_rate": 2.735580217518064e-05, |
| "loss": 0.0158, |
| "step": 6970 |
| }, |
| { |
| "epoch": 0.7150174144642492, |
| "grad_norm": 0.26267561316490173, |
| "learning_rate": 2.7346419553015175e-05, |
| "loss": 0.016, |
| "step": 6980 |
| }, |
| { |
| "epoch": 0.7160417947141979, |
| "grad_norm": 0.36447128653526306, |
| "learning_rate": 2.733702192878577e-05, |
| "loss": 0.0171, |
| "step": 6990 |
| }, |
| { |
| "epoch": 0.7170661749641467, |
| "grad_norm": 0.2908358573913574, |
| "learning_rate": 2.732760931391141e-05, |
| "loss": 0.0192, |
| "step": 7000 |
| }, |
| { |
| "epoch": 0.7180905552140955, |
| "grad_norm": 0.15208639204502106, |
| "learning_rate": 2.731818171982932e-05, |
| "loss": 0.0145, |
| "step": 7010 |
| }, |
| { |
| "epoch": 0.7191149354640443, |
| "grad_norm": 0.20952358841896057, |
| "learning_rate": 2.730873915799491e-05, |
| "loss": 0.0123, |
| "step": 7020 |
| }, |
| { |
| "epoch": 0.720139315713993, |
| "grad_norm": 0.4895021319389343, |
| "learning_rate": 2.7299281639881783e-05, |
| "loss": 0.015, |
| "step": 7030 |
| }, |
| { |
| "epoch": 0.7211636959639418, |
| "grad_norm": 0.3789525032043457, |
| "learning_rate": 2.728980917698172e-05, |
| "loss": 0.0169, |
| "step": 7040 |
| }, |
| { |
| "epoch": 0.7221880762138906, |
| "grad_norm": 0.31804490089416504, |
| "learning_rate": 2.7280321780804655e-05, |
| "loss": 0.0162, |
| "step": 7050 |
| }, |
| { |
| "epoch": 0.7232124564638394, |
| "grad_norm": 0.26229700446128845, |
| "learning_rate": 2.7270819462878666e-05, |
| "loss": 0.018, |
| "step": 7060 |
| }, |
| { |
| "epoch": 0.7242368367137881, |
| "grad_norm": 0.7650478482246399, |
| "learning_rate": 2.7261302234749965e-05, |
| "loss": 0.0172, |
| "step": 7070 |
| }, |
| { |
| "epoch": 0.7252612169637369, |
| "grad_norm": 0.17814603447914124, |
| "learning_rate": 2.7251770107982888e-05, |
| "loss": 0.0208, |
| "step": 7080 |
| }, |
| { |
| "epoch": 0.7262855972136857, |
| "grad_norm": 0.5906654000282288, |
| "learning_rate": 2.7242223094159856e-05, |
| "loss": 0.0184, |
| "step": 7090 |
| }, |
| { |
| "epoch": 0.7273099774636345, |
| "grad_norm": 0.27973249554634094, |
| "learning_rate": 2.7232661204881408e-05, |
| "loss": 0.0159, |
| "step": 7100 |
| }, |
| { |
| "epoch": 0.7283343577135832, |
| "grad_norm": 0.3667912781238556, |
| "learning_rate": 2.7223084451766123e-05, |
| "loss": 0.0189, |
| "step": 7110 |
| }, |
| { |
| "epoch": 0.729358737963532, |
| "grad_norm": 0.4520585536956787, |
| "learning_rate": 2.721349284645067e-05, |
| "loss": 0.0147, |
| "step": 7120 |
| }, |
| { |
| "epoch": 0.7303831182134809, |
| "grad_norm": 0.4741721749305725, |
| "learning_rate": 2.7203886400589752e-05, |
| "loss": 0.0149, |
| "step": 7130 |
| }, |
| { |
| "epoch": 0.7314074984634297, |
| "grad_norm": 0.18878048658370972, |
| "learning_rate": 2.7194265125856105e-05, |
| "loss": 0.0202, |
| "step": 7140 |
| }, |
| { |
| "epoch": 0.7324318787133784, |
| "grad_norm": 0.2498127520084381, |
| "learning_rate": 2.718462903394049e-05, |
| "loss": 0.0157, |
| "step": 7150 |
| }, |
| { |
| "epoch": 0.7334562589633272, |
| "grad_norm": 0.3200870156288147, |
| "learning_rate": 2.7174978136551662e-05, |
| "loss": 0.0173, |
| "step": 7160 |
| }, |
| { |
| "epoch": 0.734480639213276, |
| "grad_norm": 0.22690846025943756, |
| "learning_rate": 2.716531244541637e-05, |
| "loss": 0.0173, |
| "step": 7170 |
| }, |
| { |
| "epoch": 0.7355050194632248, |
| "grad_norm": 0.2846194803714752, |
| "learning_rate": 2.715563197227934e-05, |
| "loss": 0.0171, |
| "step": 7180 |
| }, |
| { |
| "epoch": 0.7365293997131735, |
| "grad_norm": 0.5184846520423889, |
| "learning_rate": 2.714593672890327e-05, |
| "loss": 0.0156, |
| "step": 7190 |
| }, |
| { |
| "epoch": 0.7375537799631223, |
| "grad_norm": 0.14724493026733398, |
| "learning_rate": 2.7136226727068783e-05, |
| "loss": 0.0202, |
| "step": 7200 |
| }, |
| { |
| "epoch": 0.7385781602130711, |
| "grad_norm": 0.40960296988487244, |
| "learning_rate": 2.712650197857445e-05, |
| "loss": 0.017, |
| "step": 7210 |
| }, |
| { |
| "epoch": 0.7396025404630199, |
| "grad_norm": 0.358578085899353, |
| "learning_rate": 2.7116762495236763e-05, |
| "loss": 0.0156, |
| "step": 7220 |
| }, |
| { |
| "epoch": 0.7406269207129687, |
| "grad_norm": 0.16958242654800415, |
| "learning_rate": 2.7107008288890106e-05, |
| "loss": 0.0199, |
| "step": 7230 |
| }, |
| { |
| "epoch": 0.7416513009629174, |
| "grad_norm": 0.2522384822368622, |
| "learning_rate": 2.709723937138676e-05, |
| "loss": 0.0173, |
| "step": 7240 |
| }, |
| { |
| "epoch": 0.7426756812128662, |
| "grad_norm": 0.556548535823822, |
| "learning_rate": 2.7087455754596883e-05, |
| "loss": 0.0188, |
| "step": 7250 |
| }, |
| { |
| "epoch": 0.743700061462815, |
| "grad_norm": 0.18453767895698547, |
| "learning_rate": 2.7077657450408498e-05, |
| "loss": 0.0162, |
| "step": 7260 |
| }, |
| { |
| "epoch": 0.7447244417127638, |
| "grad_norm": 0.22701260447502136, |
| "learning_rate": 2.7067844470727456e-05, |
| "loss": 0.0164, |
| "step": 7270 |
| }, |
| { |
| "epoch": 0.7457488219627125, |
| "grad_norm": 0.4192267656326294, |
| "learning_rate": 2.7058016827477455e-05, |
| "loss": 0.0154, |
| "step": 7280 |
| }, |
| { |
| "epoch": 0.7467732022126613, |
| "grad_norm": 0.19477029144763947, |
| "learning_rate": 2.7048174532600018e-05, |
| "loss": 0.0162, |
| "step": 7290 |
| }, |
| { |
| "epoch": 0.7477975824626101, |
| "grad_norm": 0.3226083815097809, |
| "learning_rate": 2.7038317598054455e-05, |
| "loss": 0.0192, |
| "step": 7300 |
| }, |
| { |
| "epoch": 0.7488219627125589, |
| "grad_norm": 0.8070310354232788, |
| "learning_rate": 2.702844603581787e-05, |
| "loss": 0.017, |
| "step": 7310 |
| }, |
| { |
| "epoch": 0.7498463429625076, |
| "grad_norm": 0.17864570021629333, |
| "learning_rate": 2.701855985788514e-05, |
| "loss": 0.0154, |
| "step": 7320 |
| }, |
| { |
| "epoch": 0.7508707232124565, |
| "grad_norm": 0.4258253872394562, |
| "learning_rate": 2.7008659076268912e-05, |
| "loss": 0.0187, |
| "step": 7330 |
| }, |
| { |
| "epoch": 0.7518951034624053, |
| "grad_norm": 0.25371941924095154, |
| "learning_rate": 2.6998743702999562e-05, |
| "loss": 0.0157, |
| "step": 7340 |
| }, |
| { |
| "epoch": 0.7529194837123541, |
| "grad_norm": 0.26969289779663086, |
| "learning_rate": 2.6988813750125202e-05, |
| "loss": 0.0146, |
| "step": 7350 |
| }, |
| { |
| "epoch": 0.7539438639623028, |
| "grad_norm": 0.633540153503418, |
| "learning_rate": 2.6978869229711668e-05, |
| "loss": 0.0174, |
| "step": 7360 |
| }, |
| { |
| "epoch": 0.7549682442122516, |
| "grad_norm": 0.28244438767433167, |
| "learning_rate": 2.6968910153842484e-05, |
| "loss": 0.0181, |
| "step": 7370 |
| }, |
| { |
| "epoch": 0.7559926244622004, |
| "grad_norm": 0.7857756018638611, |
| "learning_rate": 2.6958936534618874e-05, |
| "loss": 0.016, |
| "step": 7380 |
| }, |
| { |
| "epoch": 0.7570170047121492, |
| "grad_norm": 0.2982165217399597, |
| "learning_rate": 2.694894838415971e-05, |
| "loss": 0.0178, |
| "step": 7390 |
| }, |
| { |
| "epoch": 0.7580413849620979, |
| "grad_norm": 0.22524069249629974, |
| "learning_rate": 2.6938945714601555e-05, |
| "loss": 0.0142, |
| "step": 7400 |
| }, |
| { |
| "epoch": 0.7590657652120467, |
| "grad_norm": 0.20735964179039001, |
| "learning_rate": 2.6928928538098587e-05, |
| "loss": 0.0147, |
| "step": 7410 |
| }, |
| { |
| "epoch": 0.7600901454619955, |
| "grad_norm": 0.4288942813873291, |
| "learning_rate": 2.6918896866822615e-05, |
| "loss": 0.0172, |
| "step": 7420 |
| }, |
| { |
| "epoch": 0.7611145257119443, |
| "grad_norm": 0.4393393099308014, |
| "learning_rate": 2.690885071296308e-05, |
| "loss": 0.0158, |
| "step": 7430 |
| }, |
| { |
| "epoch": 0.762138905961893, |
| "grad_norm": 0.3155905604362488, |
| "learning_rate": 2.6898790088726993e-05, |
| "loss": 0.0177, |
| "step": 7440 |
| }, |
| { |
| "epoch": 0.7631632862118418, |
| "grad_norm": 0.26497331261634827, |
| "learning_rate": 2.6888715006338968e-05, |
| "loss": 0.015, |
| "step": 7450 |
| }, |
| { |
| "epoch": 0.7641876664617906, |
| "grad_norm": 0.578819990158081, |
| "learning_rate": 2.6878625478041172e-05, |
| "loss": 0.0151, |
| "step": 7460 |
| }, |
| { |
| "epoch": 0.7652120467117394, |
| "grad_norm": 0.3057912588119507, |
| "learning_rate": 2.6868521516093342e-05, |
| "loss": 0.015, |
| "step": 7470 |
| }, |
| { |
| "epoch": 0.7662364269616881, |
| "grad_norm": 0.33452117443084717, |
| "learning_rate": 2.6858403132772745e-05, |
| "loss": 0.018, |
| "step": 7480 |
| }, |
| { |
| "epoch": 0.7672608072116369, |
| "grad_norm": 0.12242898344993591, |
| "learning_rate": 2.6848270340374162e-05, |
| "loss": 0.0137, |
| "step": 7490 |
| }, |
| { |
| "epoch": 0.7682851874615857, |
| "grad_norm": 0.4009722173213959, |
| "learning_rate": 2.68381231512099e-05, |
| "loss": 0.0152, |
| "step": 7500 |
| }, |
| { |
| "epoch": 0.7693095677115345, |
| "grad_norm": 0.25396695733070374, |
| "learning_rate": 2.6827961577609745e-05, |
| "loss": 0.0149, |
| "step": 7510 |
| }, |
| { |
| "epoch": 0.7703339479614834, |
| "grad_norm": 0.23192083835601807, |
| "learning_rate": 2.681778563192097e-05, |
| "loss": 0.0155, |
| "step": 7520 |
| }, |
| { |
| "epoch": 0.7713583282114321, |
| "grad_norm": 0.4426534175872803, |
| "learning_rate": 2.6807595326508308e-05, |
| "loss": 0.0146, |
| "step": 7530 |
| }, |
| { |
| "epoch": 0.7723827084613809, |
| "grad_norm": 0.26815804839134216, |
| "learning_rate": 2.679739067375394e-05, |
| "loss": 0.0182, |
| "step": 7540 |
| }, |
| { |
| "epoch": 0.7734070887113297, |
| "grad_norm": 0.6119135618209839, |
| "learning_rate": 2.6787171686057487e-05, |
| "loss": 0.0154, |
| "step": 7550 |
| }, |
| { |
| "epoch": 0.7744314689612785, |
| "grad_norm": 0.2958919107913971, |
| "learning_rate": 2.677693837583597e-05, |
| "loss": 0.0167, |
| "step": 7560 |
| }, |
| { |
| "epoch": 0.7754558492112272, |
| "grad_norm": 0.15216588973999023, |
| "learning_rate": 2.6766690755523836e-05, |
| "loss": 0.0142, |
| "step": 7570 |
| }, |
| { |
| "epoch": 0.776480229461176, |
| "grad_norm": 1.0104928016662598, |
| "learning_rate": 2.675642883757291e-05, |
| "loss": 0.0166, |
| "step": 7580 |
| }, |
| { |
| "epoch": 0.7775046097111248, |
| "grad_norm": 0.41246312856674194, |
| "learning_rate": 2.674615263445239e-05, |
| "loss": 0.0162, |
| "step": 7590 |
| }, |
| { |
| "epoch": 0.7785289899610736, |
| "grad_norm": 0.15693791210651398, |
| "learning_rate": 2.673586215864882e-05, |
| "loss": 0.0147, |
| "step": 7600 |
| }, |
| { |
| "epoch": 0.7795533702110223, |
| "grad_norm": 0.2697238624095917, |
| "learning_rate": 2.6725557422666113e-05, |
| "loss": 0.0165, |
| "step": 7610 |
| }, |
| { |
| "epoch": 0.7805777504609711, |
| "grad_norm": 0.12802690267562866, |
| "learning_rate": 2.6715238439025493e-05, |
| "loss": 0.0146, |
| "step": 7620 |
| }, |
| { |
| "epoch": 0.7816021307109199, |
| "grad_norm": 0.1555444598197937, |
| "learning_rate": 2.6704905220265493e-05, |
| "loss": 0.0143, |
| "step": 7630 |
| }, |
| { |
| "epoch": 0.7826265109608687, |
| "grad_norm": 0.32642653584480286, |
| "learning_rate": 2.6694557778941947e-05, |
| "loss": 0.0196, |
| "step": 7640 |
| }, |
| { |
| "epoch": 0.7836508912108174, |
| "grad_norm": 0.5561942458152771, |
| "learning_rate": 2.6684196127627977e-05, |
| "loss": 0.0178, |
| "step": 7650 |
| }, |
| { |
| "epoch": 0.7846752714607662, |
| "grad_norm": 0.20887190103530884, |
| "learning_rate": 2.6673820278913963e-05, |
| "loss": 0.0168, |
| "step": 7660 |
| }, |
| { |
| "epoch": 0.785699651710715, |
| "grad_norm": 0.24827495217323303, |
| "learning_rate": 2.6663430245407544e-05, |
| "loss": 0.0169, |
| "step": 7670 |
| }, |
| { |
| "epoch": 0.7867240319606638, |
| "grad_norm": 0.5800760388374329, |
| "learning_rate": 2.6653026039733585e-05, |
| "loss": 0.015, |
| "step": 7680 |
| }, |
| { |
| "epoch": 0.7877484122106125, |
| "grad_norm": 0.362886518239975, |
| "learning_rate": 2.6642607674534176e-05, |
| "loss": 0.0165, |
| "step": 7690 |
| }, |
| { |
| "epoch": 0.7887727924605613, |
| "grad_norm": 0.28075724840164185, |
| "learning_rate": 2.663217516246862e-05, |
| "loss": 0.0133, |
| "step": 7700 |
| }, |
| { |
| "epoch": 0.7897971727105101, |
| "grad_norm": 0.15478134155273438, |
| "learning_rate": 2.66217285162134e-05, |
| "loss": 0.0184, |
| "step": 7710 |
| }, |
| { |
| "epoch": 0.790821552960459, |
| "grad_norm": 0.33862605690956116, |
| "learning_rate": 2.6611267748462175e-05, |
| "loss": 0.0178, |
| "step": 7720 |
| }, |
| { |
| "epoch": 0.7918459332104077, |
| "grad_norm": 0.18788599967956543, |
| "learning_rate": 2.6600792871925764e-05, |
| "loss": 0.0164, |
| "step": 7730 |
| }, |
| { |
| "epoch": 0.7928703134603565, |
| "grad_norm": 0.4512311518192291, |
| "learning_rate": 2.659030389933213e-05, |
| "loss": 0.0133, |
| "step": 7740 |
| }, |
| { |
| "epoch": 0.7938946937103053, |
| "grad_norm": 0.4341576099395752, |
| "learning_rate": 2.6579800843426358e-05, |
| "loss": 0.0155, |
| "step": 7750 |
| }, |
| { |
| "epoch": 0.7949190739602541, |
| "grad_norm": 0.3127734661102295, |
| "learning_rate": 2.656928371697067e-05, |
| "loss": 0.0172, |
| "step": 7760 |
| }, |
| { |
| "epoch": 0.7959434542102028, |
| "grad_norm": 0.2984481453895569, |
| "learning_rate": 2.6558752532744346e-05, |
| "loss": 0.0165, |
| "step": 7770 |
| }, |
| { |
| "epoch": 0.7969678344601516, |
| "grad_norm": 0.11620308458805084, |
| "learning_rate": 2.654820730354377e-05, |
| "loss": 0.0156, |
| "step": 7780 |
| }, |
| { |
| "epoch": 0.7979922147101004, |
| "grad_norm": 0.5121631622314453, |
| "learning_rate": 2.6537648042182406e-05, |
| "loss": 0.0164, |
| "step": 7790 |
| }, |
| { |
| "epoch": 0.7990165949600492, |
| "grad_norm": 0.34094807505607605, |
| "learning_rate": 2.6527074761490735e-05, |
| "loss": 0.0174, |
| "step": 7800 |
| }, |
| { |
| "epoch": 0.800040975209998, |
| "grad_norm": 0.4180108904838562, |
| "learning_rate": 2.6516487474316303e-05, |
| "loss": 0.0168, |
| "step": 7810 |
| }, |
| { |
| "epoch": 0.8010653554599467, |
| "grad_norm": 0.20907804369926453, |
| "learning_rate": 2.650588619352365e-05, |
| "loss": 0.0127, |
| "step": 7820 |
| }, |
| { |
| "epoch": 0.8020897357098955, |
| "grad_norm": 0.3351478576660156, |
| "learning_rate": 2.649527093199434e-05, |
| "loss": 0.0157, |
| "step": 7830 |
| }, |
| { |
| "epoch": 0.8031141159598443, |
| "grad_norm": 0.23962736129760742, |
| "learning_rate": 2.6484641702626914e-05, |
| "loss": 0.0156, |
| "step": 7840 |
| }, |
| { |
| "epoch": 0.804138496209793, |
| "grad_norm": 0.2321198582649231, |
| "learning_rate": 2.647399851833689e-05, |
| "loss": 0.0229, |
| "step": 7850 |
| }, |
| { |
| "epoch": 0.8051628764597418, |
| "grad_norm": 0.1963765025138855, |
| "learning_rate": 2.6463341392056735e-05, |
| "loss": 0.0189, |
| "step": 7860 |
| }, |
| { |
| "epoch": 0.8061872567096906, |
| "grad_norm": 0.3131219744682312, |
| "learning_rate": 2.6452670336735867e-05, |
| "loss": 0.0177, |
| "step": 7870 |
| }, |
| { |
| "epoch": 0.8072116369596394, |
| "grad_norm": 0.23607558012008667, |
| "learning_rate": 2.644198536534063e-05, |
| "loss": 0.0181, |
| "step": 7880 |
| }, |
| { |
| "epoch": 0.8082360172095882, |
| "grad_norm": 0.16525885462760925, |
| "learning_rate": 2.6431286490854255e-05, |
| "loss": 0.0154, |
| "step": 7890 |
| }, |
| { |
| "epoch": 0.8092603974595369, |
| "grad_norm": 0.17901058495044708, |
| "learning_rate": 2.6420573726276903e-05, |
| "loss": 0.0136, |
| "step": 7900 |
| }, |
| { |
| "epoch": 0.8102847777094858, |
| "grad_norm": 0.3410714268684387, |
| "learning_rate": 2.6409847084625575e-05, |
| "loss": 0.0157, |
| "step": 7910 |
| }, |
| { |
| "epoch": 0.8113091579594346, |
| "grad_norm": 0.8034144043922424, |
| "learning_rate": 2.6399106578934167e-05, |
| "loss": 0.0145, |
| "step": 7920 |
| }, |
| { |
| "epoch": 0.8123335382093834, |
| "grad_norm": 0.2556256055831909, |
| "learning_rate": 2.63883522222534e-05, |
| "loss": 0.0152, |
| "step": 7930 |
| }, |
| { |
| "epoch": 0.8133579184593321, |
| "grad_norm": 0.3616190254688263, |
| "learning_rate": 2.6377584027650827e-05, |
| "loss": 0.0141, |
| "step": 7940 |
| }, |
| { |
| "epoch": 0.8143822987092809, |
| "grad_norm": 0.3378823697566986, |
| "learning_rate": 2.636680200821083e-05, |
| "loss": 0.0126, |
| "step": 7950 |
| }, |
| { |
| "epoch": 0.8154066789592297, |
| "grad_norm": 0.6798502802848816, |
| "learning_rate": 2.635600617703457e-05, |
| "loss": 0.0189, |
| "step": 7960 |
| }, |
| { |
| "epoch": 0.8164310592091785, |
| "grad_norm": 0.2865912616252899, |
| "learning_rate": 2.6345196547240007e-05, |
| "loss": 0.0152, |
| "step": 7970 |
| }, |
| { |
| "epoch": 0.8174554394591272, |
| "grad_norm": 0.5794517993927002, |
| "learning_rate": 2.6334373131961854e-05, |
| "loss": 0.0137, |
| "step": 7980 |
| }, |
| { |
| "epoch": 0.818479819709076, |
| "grad_norm": 0.2529003918170929, |
| "learning_rate": 2.6323535944351583e-05, |
| "loss": 0.0141, |
| "step": 7990 |
| }, |
| { |
| "epoch": 0.8195041999590248, |
| "grad_norm": 0.4019635319709778, |
| "learning_rate": 2.63126849975774e-05, |
| "loss": 0.016, |
| "step": 8000 |
| }, |
| { |
| "epoch": 0.8205285802089736, |
| "grad_norm": 0.27212825417518616, |
| "learning_rate": 2.6301820304824234e-05, |
| "loss": 0.0148, |
| "step": 8010 |
| }, |
| { |
| "epoch": 0.8215529604589223, |
| "grad_norm": 0.32897505164146423, |
| "learning_rate": 2.6290941879293703e-05, |
| "loss": 0.0158, |
| "step": 8020 |
| }, |
| { |
| "epoch": 0.8225773407088711, |
| "grad_norm": 0.18337197601795197, |
| "learning_rate": 2.628004973420412e-05, |
| "loss": 0.0143, |
| "step": 8030 |
| }, |
| { |
| "epoch": 0.8236017209588199, |
| "grad_norm": 0.16176874935626984, |
| "learning_rate": 2.626914388279047e-05, |
| "loss": 0.0159, |
| "step": 8040 |
| }, |
| { |
| "epoch": 0.8246261012087687, |
| "grad_norm": 0.3223077952861786, |
| "learning_rate": 2.6258224338304394e-05, |
| "loss": 0.0138, |
| "step": 8050 |
| }, |
| { |
| "epoch": 0.8256504814587174, |
| "grad_norm": 0.1674104928970337, |
| "learning_rate": 2.6247291114014166e-05, |
| "loss": 0.0189, |
| "step": 8060 |
| }, |
| { |
| "epoch": 0.8266748617086662, |
| "grad_norm": 0.5387569069862366, |
| "learning_rate": 2.6236344223204683e-05, |
| "loss": 0.0163, |
| "step": 8070 |
| }, |
| { |
| "epoch": 0.827699241958615, |
| "grad_norm": 0.18137380480766296, |
| "learning_rate": 2.6225383679177453e-05, |
| "loss": 0.0138, |
| "step": 8080 |
| }, |
| { |
| "epoch": 0.8287236222085638, |
| "grad_norm": 0.3350803256034851, |
| "learning_rate": 2.6214409495250568e-05, |
| "loss": 0.014, |
| "step": 8090 |
| }, |
| { |
| "epoch": 0.8297480024585125, |
| "grad_norm": 0.27069732546806335, |
| "learning_rate": 2.620342168475869e-05, |
| "loss": 0.0165, |
| "step": 8100 |
| }, |
| { |
| "epoch": 0.8307723827084614, |
| "grad_norm": 0.38835015892982483, |
| "learning_rate": 2.619242026105306e-05, |
| "loss": 0.0162, |
| "step": 8110 |
| }, |
| { |
| "epoch": 0.8317967629584102, |
| "grad_norm": 0.2788877487182617, |
| "learning_rate": 2.618140523750143e-05, |
| "loss": 0.014, |
| "step": 8120 |
| }, |
| { |
| "epoch": 0.832821143208359, |
| "grad_norm": 0.4050328731536865, |
| "learning_rate": 2.6170376627488103e-05, |
| "loss": 0.0174, |
| "step": 8130 |
| }, |
| { |
| "epoch": 0.8338455234583078, |
| "grad_norm": 0.2782834768295288, |
| "learning_rate": 2.6159334444413873e-05, |
| "loss": 0.0156, |
| "step": 8140 |
| }, |
| { |
| "epoch": 0.8348699037082565, |
| "grad_norm": 0.4955120086669922, |
| "learning_rate": 2.614827870169604e-05, |
| "loss": 0.0153, |
| "step": 8150 |
| }, |
| { |
| "epoch": 0.8358942839582053, |
| "grad_norm": 0.33219200372695923, |
| "learning_rate": 2.6137209412768367e-05, |
| "loss": 0.0167, |
| "step": 8160 |
| }, |
| { |
| "epoch": 0.8369186642081541, |
| "grad_norm": 0.6969621777534485, |
| "learning_rate": 2.612612659108109e-05, |
| "loss": 0.0193, |
| "step": 8170 |
| }, |
| { |
| "epoch": 0.8379430444581029, |
| "grad_norm": 0.26654475927352905, |
| "learning_rate": 2.611503025010088e-05, |
| "loss": 0.0173, |
| "step": 8180 |
| }, |
| { |
| "epoch": 0.8389674247080516, |
| "grad_norm": 0.19422677159309387, |
| "learning_rate": 2.6103920403310832e-05, |
| "loss": 0.0152, |
| "step": 8190 |
| }, |
| { |
| "epoch": 0.8399918049580004, |
| "grad_norm": 0.2187219262123108, |
| "learning_rate": 2.6092797064210465e-05, |
| "loss": 0.0141, |
| "step": 8200 |
| }, |
| { |
| "epoch": 0.8410161852079492, |
| "grad_norm": 0.19506877660751343, |
| "learning_rate": 2.6081660246315685e-05, |
| "loss": 0.0177, |
| "step": 8210 |
| }, |
| { |
| "epoch": 0.842040565457898, |
| "grad_norm": 0.35447773337364197, |
| "learning_rate": 2.6070509963158774e-05, |
| "loss": 0.0171, |
| "step": 8220 |
| }, |
| { |
| "epoch": 0.8430649457078467, |
| "grad_norm": 0.4142078161239624, |
| "learning_rate": 2.6059346228288375e-05, |
| "loss": 0.0134, |
| "step": 8230 |
| }, |
| { |
| "epoch": 0.8440893259577955, |
| "grad_norm": 0.2112116515636444, |
| "learning_rate": 2.604816905526948e-05, |
| "loss": 0.0154, |
| "step": 8240 |
| }, |
| { |
| "epoch": 0.8451137062077443, |
| "grad_norm": 0.3668624460697174, |
| "learning_rate": 2.603697845768341e-05, |
| "loss": 0.0154, |
| "step": 8250 |
| }, |
| { |
| "epoch": 0.8461380864576931, |
| "grad_norm": 0.343489408493042, |
| "learning_rate": 2.6025774449127792e-05, |
| "loss": 0.0158, |
| "step": 8260 |
| }, |
| { |
| "epoch": 0.8471624667076418, |
| "grad_norm": 0.370613694190979, |
| "learning_rate": 2.6014557043216562e-05, |
| "loss": 0.0141, |
| "step": 8270 |
| }, |
| { |
| "epoch": 0.8481868469575906, |
| "grad_norm": 0.23206596076488495, |
| "learning_rate": 2.6003326253579913e-05, |
| "loss": 0.0173, |
| "step": 8280 |
| }, |
| { |
| "epoch": 0.8492112272075394, |
| "grad_norm": 0.20969480276107788, |
| "learning_rate": 2.599208209386432e-05, |
| "loss": 0.0158, |
| "step": 8290 |
| }, |
| { |
| "epoch": 0.8502356074574883, |
| "grad_norm": 0.3885093629360199, |
| "learning_rate": 2.5980824577732502e-05, |
| "loss": 0.0138, |
| "step": 8300 |
| }, |
| { |
| "epoch": 0.851259987707437, |
| "grad_norm": 0.6133009791374207, |
| "learning_rate": 2.5969553718863395e-05, |
| "loss": 0.0137, |
| "step": 8310 |
| }, |
| { |
| "epoch": 0.8522843679573858, |
| "grad_norm": 0.1833907961845398, |
| "learning_rate": 2.595826953095216e-05, |
| "loss": 0.014, |
| "step": 8320 |
| }, |
| { |
| "epoch": 0.8533087482073346, |
| "grad_norm": 0.16815197467803955, |
| "learning_rate": 2.594697202771014e-05, |
| "loss": 0.0151, |
| "step": 8330 |
| }, |
| { |
| "epoch": 0.8543331284572834, |
| "grad_norm": 0.17033535242080688, |
| "learning_rate": 2.5935661222864888e-05, |
| "loss": 0.0137, |
| "step": 8340 |
| }, |
| { |
| "epoch": 0.8553575087072322, |
| "grad_norm": 0.2786332070827484, |
| "learning_rate": 2.592433713016008e-05, |
| "loss": 0.0153, |
| "step": 8350 |
| }, |
| { |
| "epoch": 0.8563818889571809, |
| "grad_norm": 0.6531735062599182, |
| "learning_rate": 2.5912999763355553e-05, |
| "loss": 0.0164, |
| "step": 8360 |
| }, |
| { |
| "epoch": 0.8574062692071297, |
| "grad_norm": 0.30043137073516846, |
| "learning_rate": 2.590164913622729e-05, |
| "loss": 0.0168, |
| "step": 8370 |
| }, |
| { |
| "epoch": 0.8584306494570785, |
| "grad_norm": 0.33304768800735474, |
| "learning_rate": 2.589028526256737e-05, |
| "loss": 0.0161, |
| "step": 8380 |
| }, |
| { |
| "epoch": 0.8594550297070273, |
| "grad_norm": 0.3742147386074066, |
| "learning_rate": 2.587890815618397e-05, |
| "loss": 0.0148, |
| "step": 8390 |
| }, |
| { |
| "epoch": 0.860479409956976, |
| "grad_norm": 0.21151430904865265, |
| "learning_rate": 2.5867517830901343e-05, |
| "loss": 0.0163, |
| "step": 8400 |
| }, |
| { |
| "epoch": 0.8615037902069248, |
| "grad_norm": 0.15247957408428192, |
| "learning_rate": 2.5856114300559812e-05, |
| "loss": 0.0157, |
| "step": 8410 |
| }, |
| { |
| "epoch": 0.8625281704568736, |
| "grad_norm": 0.19161395728588104, |
| "learning_rate": 2.584469757901574e-05, |
| "loss": 0.0176, |
| "step": 8420 |
| }, |
| { |
| "epoch": 0.8635525507068224, |
| "grad_norm": 0.4982970058917999, |
| "learning_rate": 2.5833267680141526e-05, |
| "loss": 0.0177, |
| "step": 8430 |
| }, |
| { |
| "epoch": 0.8645769309567711, |
| "grad_norm": 0.1791178286075592, |
| "learning_rate": 2.5821824617825564e-05, |
| "loss": 0.0171, |
| "step": 8440 |
| }, |
| { |
| "epoch": 0.8656013112067199, |
| "grad_norm": 0.48737144470214844, |
| "learning_rate": 2.5810368405972265e-05, |
| "loss": 0.0162, |
| "step": 8450 |
| }, |
| { |
| "epoch": 0.8666256914566687, |
| "grad_norm": 0.2120552361011505, |
| "learning_rate": 2.5798899058501996e-05, |
| "loss": 0.0128, |
| "step": 8460 |
| }, |
| { |
| "epoch": 0.8676500717066175, |
| "grad_norm": 0.1543828397989273, |
| "learning_rate": 2.578741658935111e-05, |
| "loss": 0.016, |
| "step": 8470 |
| }, |
| { |
| "epoch": 0.8686744519565662, |
| "grad_norm": 0.18222805857658386, |
| "learning_rate": 2.5775921012471875e-05, |
| "loss": 0.0161, |
| "step": 8480 |
| }, |
| { |
| "epoch": 0.869698832206515, |
| "grad_norm": 0.21760158240795135, |
| "learning_rate": 2.5764412341832505e-05, |
| "loss": 0.017, |
| "step": 8490 |
| }, |
| { |
| "epoch": 0.8707232124564639, |
| "grad_norm": 0.15017293393611908, |
| "learning_rate": 2.5752890591417125e-05, |
| "loss": 0.0171, |
| "step": 8500 |
| }, |
| { |
| "epoch": 0.8717475927064127, |
| "grad_norm": 0.17540723085403442, |
| "learning_rate": 2.5741355775225744e-05, |
| "loss": 0.0163, |
| "step": 8510 |
| }, |
| { |
| "epoch": 0.8727719729563614, |
| "grad_norm": 0.28646236658096313, |
| "learning_rate": 2.5729807907274254e-05, |
| "loss": 0.0136, |
| "step": 8520 |
| }, |
| { |
| "epoch": 0.8737963532063102, |
| "grad_norm": 0.33049267530441284, |
| "learning_rate": 2.5718247001594403e-05, |
| "loss": 0.015, |
| "step": 8530 |
| }, |
| { |
| "epoch": 0.874820733456259, |
| "grad_norm": 0.18711215257644653, |
| "learning_rate": 2.5706673072233776e-05, |
| "loss": 0.012, |
| "step": 8540 |
| }, |
| { |
| "epoch": 0.8758451137062078, |
| "grad_norm": 0.34098097681999207, |
| "learning_rate": 2.5695086133255795e-05, |
| "loss": 0.0127, |
| "step": 8550 |
| }, |
| { |
| "epoch": 0.8768694939561565, |
| "grad_norm": 0.32089564204216003, |
| "learning_rate": 2.5683486198739687e-05, |
| "loss": 0.0152, |
| "step": 8560 |
| }, |
| { |
| "epoch": 0.8778938742061053, |
| "grad_norm": 0.19191302359104156, |
| "learning_rate": 2.5671873282780452e-05, |
| "loss": 0.0129, |
| "step": 8570 |
| }, |
| { |
| "epoch": 0.8789182544560541, |
| "grad_norm": 0.38798224925994873, |
| "learning_rate": 2.5660247399488887e-05, |
| "loss": 0.0177, |
| "step": 8580 |
| }, |
| { |
| "epoch": 0.8799426347060029, |
| "grad_norm": 0.5275190472602844, |
| "learning_rate": 2.564860856299154e-05, |
| "loss": 0.014, |
| "step": 8590 |
| }, |
| { |
| "epoch": 0.8809670149559516, |
| "grad_norm": 0.2458644062280655, |
| "learning_rate": 2.5636956787430693e-05, |
| "loss": 0.0145, |
| "step": 8600 |
| }, |
| { |
| "epoch": 0.8819913952059004, |
| "grad_norm": 0.24447175860404968, |
| "learning_rate": 2.5625292086964346e-05, |
| "loss": 0.016, |
| "step": 8610 |
| }, |
| { |
| "epoch": 0.8830157754558492, |
| "grad_norm": 0.13792452216148376, |
| "learning_rate": 2.5613614475766212e-05, |
| "loss": 0.0158, |
| "step": 8620 |
| }, |
| { |
| "epoch": 0.884040155705798, |
| "grad_norm": 0.13795943558216095, |
| "learning_rate": 2.5601923968025694e-05, |
| "loss": 0.016, |
| "step": 8630 |
| }, |
| { |
| "epoch": 0.8850645359557467, |
| "grad_norm": 0.31792324781417847, |
| "learning_rate": 2.5590220577947862e-05, |
| "loss": 0.016, |
| "step": 8640 |
| }, |
| { |
| "epoch": 0.8860889162056955, |
| "grad_norm": 0.22578349709510803, |
| "learning_rate": 2.5578504319753427e-05, |
| "loss": 0.0132, |
| "step": 8650 |
| }, |
| { |
| "epoch": 0.8871132964556443, |
| "grad_norm": 0.24781551957130432, |
| "learning_rate": 2.556677520767876e-05, |
| "loss": 0.0152, |
| "step": 8660 |
| }, |
| { |
| "epoch": 0.8881376767055931, |
| "grad_norm": 0.3055349588394165, |
| "learning_rate": 2.5555033255975835e-05, |
| "loss": 0.0149, |
| "step": 8670 |
| }, |
| { |
| "epoch": 0.8891620569555418, |
| "grad_norm": 0.21134653687477112, |
| "learning_rate": 2.5543278478912228e-05, |
| "loss": 0.0139, |
| "step": 8680 |
| }, |
| { |
| "epoch": 0.8901864372054907, |
| "grad_norm": 0.32274129986763, |
| "learning_rate": 2.55315108907711e-05, |
| "loss": 0.0159, |
| "step": 8690 |
| }, |
| { |
| "epoch": 0.8912108174554395, |
| "grad_norm": 0.4169760048389435, |
| "learning_rate": 2.551973050585119e-05, |
| "loss": 0.0169, |
| "step": 8700 |
| }, |
| { |
| "epoch": 0.8922351977053883, |
| "grad_norm": 0.3176701068878174, |
| "learning_rate": 2.5507937338466768e-05, |
| "loss": 0.015, |
| "step": 8710 |
| }, |
| { |
| "epoch": 0.8932595779553371, |
| "grad_norm": 0.3472507894039154, |
| "learning_rate": 2.549613140294764e-05, |
| "loss": 0.0143, |
| "step": 8720 |
| }, |
| { |
| "epoch": 0.8942839582052858, |
| "grad_norm": 0.19957473874092102, |
| "learning_rate": 2.548431271363914e-05, |
| "loss": 0.0144, |
| "step": 8730 |
| }, |
| { |
| "epoch": 0.8953083384552346, |
| "grad_norm": 0.24656721949577332, |
| "learning_rate": 2.5472481284902092e-05, |
| "loss": 0.0172, |
| "step": 8740 |
| }, |
| { |
| "epoch": 0.8963327187051834, |
| "grad_norm": 0.2782464325428009, |
| "learning_rate": 2.5460637131112788e-05, |
| "loss": 0.0156, |
| "step": 8750 |
| }, |
| { |
| "epoch": 0.8973570989551322, |
| "grad_norm": 0.3610098361968994, |
| "learning_rate": 2.5448780266662995e-05, |
| "loss": 0.0145, |
| "step": 8760 |
| }, |
| { |
| "epoch": 0.8983814792050809, |
| "grad_norm": 0.2550366520881653, |
| "learning_rate": 2.5436910705959926e-05, |
| "loss": 0.012, |
| "step": 8770 |
| }, |
| { |
| "epoch": 0.8994058594550297, |
| "grad_norm": 0.18428152799606323, |
| "learning_rate": 2.542502846342621e-05, |
| "loss": 0.0135, |
| "step": 8780 |
| }, |
| { |
| "epoch": 0.9004302397049785, |
| "grad_norm": 0.16183240711688995, |
| "learning_rate": 2.5413133553499892e-05, |
| "loss": 0.0143, |
| "step": 8790 |
| }, |
| { |
| "epoch": 0.9014546199549273, |
| "grad_norm": 0.4008771777153015, |
| "learning_rate": 2.540122599063442e-05, |
| "loss": 0.0163, |
| "step": 8800 |
| }, |
| { |
| "epoch": 0.902479000204876, |
| "grad_norm": 0.18650273978710175, |
| "learning_rate": 2.538930578929859e-05, |
| "loss": 0.0125, |
| "step": 8810 |
| }, |
| { |
| "epoch": 0.9035033804548248, |
| "grad_norm": 0.15881310403347015, |
| "learning_rate": 2.537737296397658e-05, |
| "loss": 0.0138, |
| "step": 8820 |
| }, |
| { |
| "epoch": 0.9045277607047736, |
| "grad_norm": 0.263323038816452, |
| "learning_rate": 2.5365427529167903e-05, |
| "loss": 0.0149, |
| "step": 8830 |
| }, |
| { |
| "epoch": 0.9055521409547224, |
| "grad_norm": 0.2534765303134918, |
| "learning_rate": 2.5353469499387375e-05, |
| "loss": 0.0155, |
| "step": 8840 |
| }, |
| { |
| "epoch": 0.9065765212046711, |
| "grad_norm": 0.2501375079154968, |
| "learning_rate": 2.534149888916514e-05, |
| "loss": 0.0139, |
| "step": 8850 |
| }, |
| { |
| "epoch": 0.9076009014546199, |
| "grad_norm": 0.22129091620445251, |
| "learning_rate": 2.5329515713046616e-05, |
| "loss": 0.0143, |
| "step": 8860 |
| }, |
| { |
| "epoch": 0.9086252817045687, |
| "grad_norm": 0.3225857615470886, |
| "learning_rate": 2.531751998559249e-05, |
| "loss": 0.0158, |
| "step": 8870 |
| }, |
| { |
| "epoch": 0.9096496619545175, |
| "grad_norm": 0.32638853788375854, |
| "learning_rate": 2.5305511721378703e-05, |
| "loss": 0.0143, |
| "step": 8880 |
| }, |
| { |
| "epoch": 0.9106740422044663, |
| "grad_norm": 0.24320584535598755, |
| "learning_rate": 2.529349093499642e-05, |
| "loss": 0.0149, |
| "step": 8890 |
| }, |
| { |
| "epoch": 0.9116984224544151, |
| "grad_norm": 0.2126363217830658, |
| "learning_rate": 2.528145764105205e-05, |
| "loss": 0.0146, |
| "step": 8900 |
| }, |
| { |
| "epoch": 0.9127228027043639, |
| "grad_norm": 0.2942073941230774, |
| "learning_rate": 2.5269411854167166e-05, |
| "loss": 0.0164, |
| "step": 8910 |
| }, |
| { |
| "epoch": 0.9137471829543127, |
| "grad_norm": 0.15426021814346313, |
| "learning_rate": 2.5257353588978534e-05, |
| "loss": 0.0164, |
| "step": 8920 |
| }, |
| { |
| "epoch": 0.9147715632042615, |
| "grad_norm": 0.28633952140808105, |
| "learning_rate": 2.5245282860138083e-05, |
| "loss": 0.0137, |
| "step": 8930 |
| }, |
| { |
| "epoch": 0.9157959434542102, |
| "grad_norm": 0.3243907392024994, |
| "learning_rate": 2.5233199682312894e-05, |
| "loss": 0.016, |
| "step": 8940 |
| }, |
| { |
| "epoch": 0.916820323704159, |
| "grad_norm": 0.4934837818145752, |
| "learning_rate": 2.5221104070185163e-05, |
| "loss": 0.0149, |
| "step": 8950 |
| }, |
| { |
| "epoch": 0.9178447039541078, |
| "grad_norm": 0.523448646068573, |
| "learning_rate": 2.5208996038452204e-05, |
| "loss": 0.0152, |
| "step": 8960 |
| }, |
| { |
| "epoch": 0.9188690842040566, |
| "grad_norm": 0.25466033816337585, |
| "learning_rate": 2.5196875601826403e-05, |
| "loss": 0.0162, |
| "step": 8970 |
| }, |
| { |
| "epoch": 0.9198934644540053, |
| "grad_norm": 0.12161422520875931, |
| "learning_rate": 2.518474277503525e-05, |
| "loss": 0.014, |
| "step": 8980 |
| }, |
| { |
| "epoch": 0.9209178447039541, |
| "grad_norm": 0.20086221396923065, |
| "learning_rate": 2.5172597572821266e-05, |
| "loss": 0.0142, |
| "step": 8990 |
| }, |
| { |
| "epoch": 0.9219422249539029, |
| "grad_norm": 0.4037904441356659, |
| "learning_rate": 2.5160440009942015e-05, |
| "loss": 0.0144, |
| "step": 9000 |
| }, |
| { |
| "epoch": 0.9229666052038517, |
| "grad_norm": 0.27131617069244385, |
| "learning_rate": 2.514827010117009e-05, |
| "loss": 0.0193, |
| "step": 9010 |
| }, |
| { |
| "epoch": 0.9239909854538004, |
| "grad_norm": 0.4256954491138458, |
| "learning_rate": 2.5136087861293068e-05, |
| "loss": 0.016, |
| "step": 9020 |
| }, |
| { |
| "epoch": 0.9250153657037492, |
| "grad_norm": 0.5509855151176453, |
| "learning_rate": 2.5123893305113525e-05, |
| "loss": 0.0154, |
| "step": 9030 |
| }, |
| { |
| "epoch": 0.926039745953698, |
| "grad_norm": 0.243845596909523, |
| "learning_rate": 2.5111686447448997e-05, |
| "loss": 0.0151, |
| "step": 9040 |
| }, |
| { |
| "epoch": 0.9270641262036468, |
| "grad_norm": 0.3576467037200928, |
| "learning_rate": 2.5099467303131966e-05, |
| "loss": 0.0196, |
| "step": 9050 |
| }, |
| { |
| "epoch": 0.9280885064535955, |
| "grad_norm": 0.3384329080581665, |
| "learning_rate": 2.508723588700984e-05, |
| "loss": 0.0149, |
| "step": 9060 |
| }, |
| { |
| "epoch": 0.9291128867035443, |
| "grad_norm": 0.1435217559337616, |
| "learning_rate": 2.5074992213944952e-05, |
| "loss": 0.0161, |
| "step": 9070 |
| }, |
| { |
| "epoch": 0.9301372669534932, |
| "grad_norm": 0.1920420080423355, |
| "learning_rate": 2.506273629881451e-05, |
| "loss": 0.0164, |
| "step": 9080 |
| }, |
| { |
| "epoch": 0.931161647203442, |
| "grad_norm": 0.1583794206380844, |
| "learning_rate": 2.5050468156510617e-05, |
| "loss": 0.0146, |
| "step": 9090 |
| }, |
| { |
| "epoch": 0.9321860274533907, |
| "grad_norm": 0.23996804654598236, |
| "learning_rate": 2.5038187801940215e-05, |
| "loss": 0.0147, |
| "step": 9100 |
| }, |
| { |
| "epoch": 0.9332104077033395, |
| "grad_norm": 0.2926790714263916, |
| "learning_rate": 2.5025895250025088e-05, |
| "loss": 0.0181, |
| "step": 9110 |
| }, |
| { |
| "epoch": 0.9342347879532883, |
| "grad_norm": 0.5305456519126892, |
| "learning_rate": 2.501359051570186e-05, |
| "loss": 0.0146, |
| "step": 9120 |
| }, |
| { |
| "epoch": 0.9352591682032371, |
| "grad_norm": 0.21375571191310883, |
| "learning_rate": 2.500127361392193e-05, |
| "loss": 0.0163, |
| "step": 9130 |
| }, |
| { |
| "epoch": 0.9362835484531858, |
| "grad_norm": 0.30986547470092773, |
| "learning_rate": 2.49889445596515e-05, |
| "loss": 0.0207, |
| "step": 9140 |
| }, |
| { |
| "epoch": 0.9373079287031346, |
| "grad_norm": 0.2224327027797699, |
| "learning_rate": 2.497660336787154e-05, |
| "loss": 0.014, |
| "step": 9150 |
| }, |
| { |
| "epoch": 0.9383323089530834, |
| "grad_norm": 0.24256621301174164, |
| "learning_rate": 2.496425005357775e-05, |
| "loss": 0.0139, |
| "step": 9160 |
| }, |
| { |
| "epoch": 0.9393566892030322, |
| "grad_norm": 0.30102643370628357, |
| "learning_rate": 2.4951884631780578e-05, |
| "loss": 0.0151, |
| "step": 9170 |
| }, |
| { |
| "epoch": 0.9403810694529809, |
| "grad_norm": 0.3082323372364044, |
| "learning_rate": 2.493950711750518e-05, |
| "loss": 0.0156, |
| "step": 9180 |
| }, |
| { |
| "epoch": 0.9414054497029297, |
| "grad_norm": 0.2706078886985779, |
| "learning_rate": 2.49271175257914e-05, |
| "loss": 0.0144, |
| "step": 9190 |
| }, |
| { |
| "epoch": 0.9424298299528785, |
| "grad_norm": 0.40828177332878113, |
| "learning_rate": 2.4914715871693757e-05, |
| "loss": 0.0171, |
| "step": 9200 |
| }, |
| { |
| "epoch": 0.9434542102028273, |
| "grad_norm": 0.20656433701515198, |
| "learning_rate": 2.4902302170281438e-05, |
| "loss": 0.0161, |
| "step": 9210 |
| }, |
| { |
| "epoch": 0.944478590452776, |
| "grad_norm": 0.31153976917266846, |
| "learning_rate": 2.4889876436638252e-05, |
| "loss": 0.0168, |
| "step": 9220 |
| }, |
| { |
| "epoch": 0.9455029707027248, |
| "grad_norm": 0.3474360704421997, |
| "learning_rate": 2.4877438685862643e-05, |
| "loss": 0.0162, |
| "step": 9230 |
| }, |
| { |
| "epoch": 0.9465273509526736, |
| "grad_norm": 0.26078999042510986, |
| "learning_rate": 2.486498893306765e-05, |
| "loss": 0.0137, |
| "step": 9240 |
| }, |
| { |
| "epoch": 0.9475517312026224, |
| "grad_norm": 0.2430412471294403, |
| "learning_rate": 2.485252719338089e-05, |
| "loss": 0.0191, |
| "step": 9250 |
| }, |
| { |
| "epoch": 0.9485761114525711, |
| "grad_norm": 0.17388616502285004, |
| "learning_rate": 2.4840053481944563e-05, |
| "loss": 0.0148, |
| "step": 9260 |
| }, |
| { |
| "epoch": 0.9496004917025199, |
| "grad_norm": 0.22588755190372467, |
| "learning_rate": 2.4827567813915398e-05, |
| "loss": 0.0126, |
| "step": 9270 |
| }, |
| { |
| "epoch": 0.9506248719524688, |
| "grad_norm": 0.31199705600738525, |
| "learning_rate": 2.481507020446466e-05, |
| "loss": 0.0136, |
| "step": 9280 |
| }, |
| { |
| "epoch": 0.9516492522024176, |
| "grad_norm": 0.3027561604976654, |
| "learning_rate": 2.4802560668778122e-05, |
| "loss": 0.0155, |
| "step": 9290 |
| }, |
| { |
| "epoch": 0.9526736324523664, |
| "grad_norm": 0.20310090482234955, |
| "learning_rate": 2.479003922205605e-05, |
| "loss": 0.0167, |
| "step": 9300 |
| }, |
| { |
| "epoch": 0.9536980127023151, |
| "grad_norm": 0.13221105933189392, |
| "learning_rate": 2.4777505879513183e-05, |
| "loss": 0.0122, |
| "step": 9310 |
| }, |
| { |
| "epoch": 0.9547223929522639, |
| "grad_norm": 0.16453291475772858, |
| "learning_rate": 2.4764960656378714e-05, |
| "loss": 0.0157, |
| "step": 9320 |
| }, |
| { |
| "epoch": 0.9557467732022127, |
| "grad_norm": 0.6800953149795532, |
| "learning_rate": 2.4752403567896274e-05, |
| "loss": 0.0157, |
| "step": 9330 |
| }, |
| { |
| "epoch": 0.9567711534521615, |
| "grad_norm": 0.594478189945221, |
| "learning_rate": 2.47398346293239e-05, |
| "loss": 0.0157, |
| "step": 9340 |
| }, |
| { |
| "epoch": 0.9577955337021102, |
| "grad_norm": 0.4451906681060791, |
| "learning_rate": 2.4727253855934055e-05, |
| "loss": 0.0178, |
| "step": 9350 |
| }, |
| { |
| "epoch": 0.958819913952059, |
| "grad_norm": 0.3534000515937805, |
| "learning_rate": 2.4714661263013548e-05, |
| "loss": 0.0183, |
| "step": 9360 |
| }, |
| { |
| "epoch": 0.9598442942020078, |
| "grad_norm": 0.1739826649427414, |
| "learning_rate": 2.4702056865863575e-05, |
| "loss": 0.0155, |
| "step": 9370 |
| }, |
| { |
| "epoch": 0.9608686744519566, |
| "grad_norm": 0.25467702746391296, |
| "learning_rate": 2.4689440679799672e-05, |
| "loss": 0.0168, |
| "step": 9380 |
| }, |
| { |
| "epoch": 0.9618930547019053, |
| "grad_norm": 0.36115628480911255, |
| "learning_rate": 2.4676812720151678e-05, |
| "loss": 0.0106, |
| "step": 9390 |
| }, |
| { |
| "epoch": 0.9629174349518541, |
| "grad_norm": 0.7658531665802002, |
| "learning_rate": 2.466417300226377e-05, |
| "loss": 0.0157, |
| "step": 9400 |
| }, |
| { |
| "epoch": 0.9639418152018029, |
| "grad_norm": 0.3737786114215851, |
| "learning_rate": 2.4651521541494394e-05, |
| "loss": 0.0143, |
| "step": 9410 |
| }, |
| { |
| "epoch": 0.9649661954517517, |
| "grad_norm": 0.16693508625030518, |
| "learning_rate": 2.4638858353216267e-05, |
| "loss": 0.0158, |
| "step": 9420 |
| }, |
| { |
| "epoch": 0.9659905757017004, |
| "grad_norm": 0.4617614150047302, |
| "learning_rate": 2.4626183452816355e-05, |
| "loss": 0.0177, |
| "step": 9430 |
| }, |
| { |
| "epoch": 0.9670149559516492, |
| "grad_norm": 0.41547277569770813, |
| "learning_rate": 2.4613496855695855e-05, |
| "loss": 0.0167, |
| "step": 9440 |
| }, |
| { |
| "epoch": 0.968039336201598, |
| "grad_norm": 0.21641899645328522, |
| "learning_rate": 2.4600798577270183e-05, |
| "loss": 0.015, |
| "step": 9450 |
| }, |
| { |
| "epoch": 0.9690637164515468, |
| "grad_norm": 0.22797243297100067, |
| "learning_rate": 2.4588088632968944e-05, |
| "loss": 0.0156, |
| "step": 9460 |
| }, |
| { |
| "epoch": 0.9700880967014957, |
| "grad_norm": 0.4429326355457306, |
| "learning_rate": 2.4575367038235916e-05, |
| "loss": 0.0131, |
| "step": 9470 |
| }, |
| { |
| "epoch": 0.9711124769514444, |
| "grad_norm": 0.4062148332595825, |
| "learning_rate": 2.4562633808529043e-05, |
| "loss": 0.0153, |
| "step": 9480 |
| }, |
| { |
| "epoch": 0.9721368572013932, |
| "grad_norm": 0.10158292204141617, |
| "learning_rate": 2.4549888959320396e-05, |
| "loss": 0.0143, |
| "step": 9490 |
| }, |
| { |
| "epoch": 0.973161237451342, |
| "grad_norm": 0.3680323660373688, |
| "learning_rate": 2.4537132506096162e-05, |
| "loss": 0.0169, |
| "step": 9500 |
| }, |
| { |
| "epoch": 0.9741856177012908, |
| "grad_norm": 0.4091793894767761, |
| "learning_rate": 2.452436446435664e-05, |
| "loss": 0.013, |
| "step": 9510 |
| }, |
| { |
| "epoch": 0.9752099979512395, |
| "grad_norm": 0.2390587478876114, |
| "learning_rate": 2.451158484961621e-05, |
| "loss": 0.0223, |
| "step": 9520 |
| }, |
| { |
| "epoch": 0.9762343782011883, |
| "grad_norm": 0.19762200117111206, |
| "learning_rate": 2.4498793677403304e-05, |
| "loss": 0.0142, |
| "step": 9530 |
| }, |
| { |
| "epoch": 0.9772587584511371, |
| "grad_norm": 0.28891947865486145, |
| "learning_rate": 2.4485990963260398e-05, |
| "loss": 0.014, |
| "step": 9540 |
| }, |
| { |
| "epoch": 0.9782831387010859, |
| "grad_norm": 0.18218865990638733, |
| "learning_rate": 2.4473176722744e-05, |
| "loss": 0.0143, |
| "step": 9550 |
| }, |
| { |
| "epoch": 0.9793075189510346, |
| "grad_norm": 0.18870890140533447, |
| "learning_rate": 2.446035097142463e-05, |
| "loss": 0.0166, |
| "step": 9560 |
| }, |
| { |
| "epoch": 0.9803318992009834, |
| "grad_norm": 0.1908000409603119, |
| "learning_rate": 2.444751372488677e-05, |
| "loss": 0.0159, |
| "step": 9570 |
| }, |
| { |
| "epoch": 0.9813562794509322, |
| "grad_norm": 0.3086000978946686, |
| "learning_rate": 2.443466499872889e-05, |
| "loss": 0.015, |
| "step": 9580 |
| }, |
| { |
| "epoch": 0.982380659700881, |
| "grad_norm": 0.44861286878585815, |
| "learning_rate": 2.4421804808563413e-05, |
| "loss": 0.0166, |
| "step": 9590 |
| }, |
| { |
| "epoch": 0.9834050399508297, |
| "grad_norm": 0.294913649559021, |
| "learning_rate": 2.440893317001667e-05, |
| "loss": 0.0152, |
| "step": 9600 |
| }, |
| { |
| "epoch": 0.9844294202007785, |
| "grad_norm": 0.13306112587451935, |
| "learning_rate": 2.4396050098728925e-05, |
| "loss": 0.0166, |
| "step": 9610 |
| }, |
| { |
| "epoch": 0.9854538004507273, |
| "grad_norm": 0.7456767559051514, |
| "learning_rate": 2.4383155610354314e-05, |
| "loss": 0.0172, |
| "step": 9620 |
| }, |
| { |
| "epoch": 0.986478180700676, |
| "grad_norm": 0.2886616587638855, |
| "learning_rate": 2.437024972056087e-05, |
| "loss": 0.018, |
| "step": 9630 |
| }, |
| { |
| "epoch": 0.9875025609506248, |
| "grad_norm": 0.2099754959344864, |
| "learning_rate": 2.4357332445030452e-05, |
| "loss": 0.0132, |
| "step": 9640 |
| }, |
| { |
| "epoch": 0.9885269412005736, |
| "grad_norm": 0.5140033960342407, |
| "learning_rate": 2.434440379945878e-05, |
| "loss": 0.0163, |
| "step": 9650 |
| }, |
| { |
| "epoch": 0.9895513214505224, |
| "grad_norm": 0.15983477234840393, |
| "learning_rate": 2.4331463799555366e-05, |
| "loss": 0.0147, |
| "step": 9660 |
| }, |
| { |
| "epoch": 0.9905757017004713, |
| "grad_norm": 0.3319350481033325, |
| "learning_rate": 2.4318512461043544e-05, |
| "loss": 0.0143, |
| "step": 9670 |
| }, |
| { |
| "epoch": 0.99160008195042, |
| "grad_norm": 0.2642180919647217, |
| "learning_rate": 2.4305549799660397e-05, |
| "loss": 0.0154, |
| "step": 9680 |
| }, |
| { |
| "epoch": 0.9926244622003688, |
| "grad_norm": 0.23326200246810913, |
| "learning_rate": 2.429257583115679e-05, |
| "loss": 0.0167, |
| "step": 9690 |
| }, |
| { |
| "epoch": 0.9936488424503176, |
| "grad_norm": 0.23949231207370758, |
| "learning_rate": 2.4279590571297316e-05, |
| "loss": 0.0188, |
| "step": 9700 |
| }, |
| { |
| "epoch": 0.9946732227002664, |
| "grad_norm": 0.16241538524627686, |
| "learning_rate": 2.4266594035860288e-05, |
| "loss": 0.0114, |
| "step": 9710 |
| }, |
| { |
| "epoch": 0.9956976029502151, |
| "grad_norm": 0.18831419944763184, |
| "learning_rate": 2.425358624063773e-05, |
| "loss": 0.0141, |
| "step": 9720 |
| }, |
| { |
| "epoch": 0.9967219832001639, |
| "grad_norm": 0.2672021985054016, |
| "learning_rate": 2.424056720143532e-05, |
| "loss": 0.0164, |
| "step": 9730 |
| }, |
| { |
| "epoch": 0.9977463634501127, |
| "grad_norm": 0.2342095673084259, |
| "learning_rate": 2.422753693407244e-05, |
| "loss": 0.0157, |
| "step": 9740 |
| }, |
| { |
| "epoch": 0.9987707437000615, |
| "grad_norm": 0.44341105222702026, |
| "learning_rate": 2.421449545438209e-05, |
| "loss": 0.0123, |
| "step": 9750 |
| }, |
| { |
| "epoch": 0.9997951239500102, |
| "grad_norm": 0.2807219326496124, |
| "learning_rate": 2.4201442778210887e-05, |
| "loss": 0.0184, |
| "step": 9760 |
| }, |
| { |
| "epoch": 1.000819504199959, |
| "grad_norm": 0.30685994029045105, |
| "learning_rate": 2.418837892141907e-05, |
| "loss": 0.0141, |
| "step": 9770 |
| }, |
| { |
| "epoch": 1.0018438844499078, |
| "grad_norm": 0.1825302690267563, |
| "learning_rate": 2.4175303899880458e-05, |
| "loss": 0.0141, |
| "step": 9780 |
| }, |
| { |
| "epoch": 1.0028682646998566, |
| "grad_norm": 0.1875874549150467, |
| "learning_rate": 2.4162217729482427e-05, |
| "loss": 0.0125, |
| "step": 9790 |
| }, |
| { |
| "epoch": 1.0038926449498053, |
| "grad_norm": 0.2336321622133255, |
| "learning_rate": 2.4149120426125918e-05, |
| "loss": 0.0165, |
| "step": 9800 |
| }, |
| { |
| "epoch": 1.0049170251997541, |
| "grad_norm": 0.2906849682331085, |
| "learning_rate": 2.413601200572538e-05, |
| "loss": 0.0176, |
| "step": 9810 |
| }, |
| { |
| "epoch": 1.005941405449703, |
| "grad_norm": 0.3306921124458313, |
| "learning_rate": 2.4122892484208794e-05, |
| "loss": 0.0161, |
| "step": 9820 |
| }, |
| { |
| "epoch": 1.0069657856996517, |
| "grad_norm": 0.30475977063179016, |
| "learning_rate": 2.4109761877517603e-05, |
| "loss": 0.0167, |
| "step": 9830 |
| }, |
| { |
| "epoch": 1.0079901659496004, |
| "grad_norm": 0.301849901676178, |
| "learning_rate": 2.4096620201606737e-05, |
| "loss": 0.015, |
| "step": 9840 |
| }, |
| { |
| "epoch": 1.0090145461995492, |
| "grad_norm": 0.39280790090560913, |
| "learning_rate": 2.4083467472444573e-05, |
| "loss": 0.0157, |
| "step": 9850 |
| }, |
| { |
| "epoch": 1.010038926449498, |
| "grad_norm": 0.30809351801872253, |
| "learning_rate": 2.4070303706012912e-05, |
| "loss": 0.0152, |
| "step": 9860 |
| }, |
| { |
| "epoch": 1.0110633066994468, |
| "grad_norm": 0.4052877724170685, |
| "learning_rate": 2.4057128918306975e-05, |
| "loss": 0.0166, |
| "step": 9870 |
| }, |
| { |
| "epoch": 1.0120876869493955, |
| "grad_norm": 0.36830392479896545, |
| "learning_rate": 2.4043943125335373e-05, |
| "loss": 0.0168, |
| "step": 9880 |
| }, |
| { |
| "epoch": 1.0131120671993443, |
| "grad_norm": 0.250506728887558, |
| "learning_rate": 2.403074634312008e-05, |
| "loss": 0.0193, |
| "step": 9890 |
| }, |
| { |
| "epoch": 1.014136447449293, |
| "grad_norm": 0.17920617759227753, |
| "learning_rate": 2.4017538587696442e-05, |
| "loss": 0.0168, |
| "step": 9900 |
| }, |
| { |
| "epoch": 1.0151608276992419, |
| "grad_norm": 0.2513575553894043, |
| "learning_rate": 2.4004319875113118e-05, |
| "loss": 0.0151, |
| "step": 9910 |
| }, |
| { |
| "epoch": 1.0161852079491906, |
| "grad_norm": 0.2233298420906067, |
| "learning_rate": 2.3991090221432096e-05, |
| "loss": 0.0159, |
| "step": 9920 |
| }, |
| { |
| "epoch": 1.0172095881991394, |
| "grad_norm": 0.19957461953163147, |
| "learning_rate": 2.397784964272865e-05, |
| "loss": 0.0156, |
| "step": 9930 |
| }, |
| { |
| "epoch": 1.0182339684490882, |
| "grad_norm": 0.32318419218063354, |
| "learning_rate": 2.3964598155091335e-05, |
| "loss": 0.0131, |
| "step": 9940 |
| }, |
| { |
| "epoch": 1.019258348699037, |
| "grad_norm": 0.3026696443557739, |
| "learning_rate": 2.3951335774621952e-05, |
| "loss": 0.0147, |
| "step": 9950 |
| }, |
| { |
| "epoch": 1.020282728948986, |
| "grad_norm": 0.2060929834842682, |
| "learning_rate": 2.3938062517435548e-05, |
| "loss": 0.0208, |
| "step": 9960 |
| }, |
| { |
| "epoch": 1.0213071091989347, |
| "grad_norm": 0.4713757634162903, |
| "learning_rate": 2.392477839966038e-05, |
| "loss": 0.014, |
| "step": 9970 |
| }, |
| { |
| "epoch": 1.0223314894488835, |
| "grad_norm": 0.22139571607112885, |
| "learning_rate": 2.3911483437437904e-05, |
| "loss": 0.0167, |
| "step": 9980 |
| }, |
| { |
| "epoch": 1.0233558696988323, |
| "grad_norm": 0.4890003800392151, |
| "learning_rate": 2.3898177646922757e-05, |
| "loss": 0.0131, |
| "step": 9990 |
| }, |
| { |
| "epoch": 1.024380249948781, |
| "grad_norm": 0.22033756971359253, |
| "learning_rate": 2.3884861044282722e-05, |
| "loss": 0.0139, |
| "step": 10000 |
| }, |
| { |
| "epoch": 1.0254046301987298, |
| "grad_norm": 0.5444302558898926, |
| "learning_rate": 2.3871533645698732e-05, |
| "loss": 0.0133, |
| "step": 10010 |
| }, |
| { |
| "epoch": 1.0264290104486786, |
| "grad_norm": 0.1380189061164856, |
| "learning_rate": 2.3858195467364833e-05, |
| "loss": 0.0153, |
| "step": 10020 |
| }, |
| { |
| "epoch": 1.0274533906986274, |
| "grad_norm": 0.23673130571842194, |
| "learning_rate": 2.3844846525488166e-05, |
| "loss": 0.0175, |
| "step": 10030 |
| }, |
| { |
| "epoch": 1.0284777709485762, |
| "grad_norm": 0.15531551837921143, |
| "learning_rate": 2.3831486836288957e-05, |
| "loss": 0.0136, |
| "step": 10040 |
| }, |
| { |
| "epoch": 1.029502151198525, |
| "grad_norm": 0.22529898583889008, |
| "learning_rate": 2.381811641600048e-05, |
| "loss": 0.0142, |
| "step": 10050 |
| }, |
| { |
| "epoch": 1.0305265314484737, |
| "grad_norm": 0.19423753023147583, |
| "learning_rate": 2.3804735280869073e-05, |
| "loss": 0.0137, |
| "step": 10060 |
| }, |
| { |
| "epoch": 1.0315509116984225, |
| "grad_norm": 0.4878730773925781, |
| "learning_rate": 2.3791343447154064e-05, |
| "loss": 0.0161, |
| "step": 10070 |
| }, |
| { |
| "epoch": 1.0325752919483713, |
| "grad_norm": 0.25544533133506775, |
| "learning_rate": 2.37779409311278e-05, |
| "loss": 0.0139, |
| "step": 10080 |
| }, |
| { |
| "epoch": 1.03359967219832, |
| "grad_norm": 0.1452128291130066, |
| "learning_rate": 2.3764527749075596e-05, |
| "loss": 0.0142, |
| "step": 10090 |
| }, |
| { |
| "epoch": 1.0346240524482688, |
| "grad_norm": 0.14237940311431885, |
| "learning_rate": 2.3751103917295735e-05, |
| "loss": 0.015, |
| "step": 10100 |
| }, |
| { |
| "epoch": 1.0356484326982176, |
| "grad_norm": 0.33134031295776367, |
| "learning_rate": 2.373766945209944e-05, |
| "loss": 0.014, |
| "step": 10110 |
| }, |
| { |
| "epoch": 1.0366728129481664, |
| "grad_norm": 0.1765003800392151, |
| "learning_rate": 2.3724224369810856e-05, |
| "loss": 0.014, |
| "step": 10120 |
| }, |
| { |
| "epoch": 1.0376971931981152, |
| "grad_norm": 0.3016045391559601, |
| "learning_rate": 2.3710768686767015e-05, |
| "loss": 0.011, |
| "step": 10130 |
| }, |
| { |
| "epoch": 1.038721573448064, |
| "grad_norm": 0.1968013197183609, |
| "learning_rate": 2.3697302419317856e-05, |
| "loss": 0.0127, |
| "step": 10140 |
| }, |
| { |
| "epoch": 1.0397459536980127, |
| "grad_norm": 0.272747665643692, |
| "learning_rate": 2.3683825583826146e-05, |
| "loss": 0.0164, |
| "step": 10150 |
| }, |
| { |
| "epoch": 1.0407703339479615, |
| "grad_norm": 0.2112593650817871, |
| "learning_rate": 2.3670338196667528e-05, |
| "loss": 0.013, |
| "step": 10160 |
| }, |
| { |
| "epoch": 1.0417947141979103, |
| "grad_norm": 0.3381974399089813, |
| "learning_rate": 2.3656840274230443e-05, |
| "loss": 0.0127, |
| "step": 10170 |
| }, |
| { |
| "epoch": 1.042819094447859, |
| "grad_norm": 0.16871069371700287, |
| "learning_rate": 2.3643331832916133e-05, |
| "loss": 0.0136, |
| "step": 10180 |
| }, |
| { |
| "epoch": 1.0438434746978078, |
| "grad_norm": 0.1627000868320465, |
| "learning_rate": 2.362981288913863e-05, |
| "loss": 0.0143, |
| "step": 10190 |
| }, |
| { |
| "epoch": 1.0448678549477566, |
| "grad_norm": 0.2683081030845642, |
| "learning_rate": 2.3616283459324737e-05, |
| "loss": 0.0138, |
| "step": 10200 |
| }, |
| { |
| "epoch": 1.0458922351977054, |
| "grad_norm": 0.37282848358154297, |
| "learning_rate": 2.3602743559913973e-05, |
| "loss": 0.0145, |
| "step": 10210 |
| }, |
| { |
| "epoch": 1.0469166154476541, |
| "grad_norm": 0.15974655747413635, |
| "learning_rate": 2.35891932073586e-05, |
| "loss": 0.0149, |
| "step": 10220 |
| }, |
| { |
| "epoch": 1.047940995697603, |
| "grad_norm": 0.22651997208595276, |
| "learning_rate": 2.357563241812357e-05, |
| "loss": 0.0162, |
| "step": 10230 |
| }, |
| { |
| "epoch": 1.0489653759475517, |
| "grad_norm": 0.8356772661209106, |
| "learning_rate": 2.3562061208686522e-05, |
| "loss": 0.0142, |
| "step": 10240 |
| }, |
| { |
| "epoch": 1.0499897561975005, |
| "grad_norm": 0.3883328437805176, |
| "learning_rate": 2.354847959553776e-05, |
| "loss": 0.0125, |
| "step": 10250 |
| }, |
| { |
| "epoch": 1.0510141364474492, |
| "grad_norm": 0.6201087832450867, |
| "learning_rate": 2.353488759518022e-05, |
| "loss": 0.0138, |
| "step": 10260 |
| }, |
| { |
| "epoch": 1.052038516697398, |
| "grad_norm": 0.7259970903396606, |
| "learning_rate": 2.352128522412946e-05, |
| "loss": 0.0156, |
| "step": 10270 |
| }, |
| { |
| "epoch": 1.0530628969473468, |
| "grad_norm": 0.19446827471256256, |
| "learning_rate": 2.3507672498913653e-05, |
| "loss": 0.0135, |
| "step": 10280 |
| }, |
| { |
| "epoch": 1.0540872771972956, |
| "grad_norm": 0.25560879707336426, |
| "learning_rate": 2.3494049436073537e-05, |
| "loss": 0.0149, |
| "step": 10290 |
| }, |
| { |
| "epoch": 1.0551116574472443, |
| "grad_norm": 0.15617823600769043, |
| "learning_rate": 2.348041605216242e-05, |
| "loss": 0.0136, |
| "step": 10300 |
| }, |
| { |
| "epoch": 1.056136037697193, |
| "grad_norm": 0.5250183343887329, |
| "learning_rate": 2.3466772363746146e-05, |
| "loss": 0.0142, |
| "step": 10310 |
| }, |
| { |
| "epoch": 1.0571604179471419, |
| "grad_norm": 0.48453715443611145, |
| "learning_rate": 2.345311838740309e-05, |
| "loss": 0.0159, |
| "step": 10320 |
| }, |
| { |
| "epoch": 1.0581847981970907, |
| "grad_norm": 0.40211451053619385, |
| "learning_rate": 2.343945413972411e-05, |
| "loss": 0.0146, |
| "step": 10330 |
| }, |
| { |
| "epoch": 1.0592091784470394, |
| "grad_norm": 0.19622018933296204, |
| "learning_rate": 2.3425779637312568e-05, |
| "loss": 0.015, |
| "step": 10340 |
| }, |
| { |
| "epoch": 1.0602335586969884, |
| "grad_norm": 0.19315792620182037, |
| "learning_rate": 2.3412094896784264e-05, |
| "loss": 0.0141, |
| "step": 10350 |
| }, |
| { |
| "epoch": 1.0612579389469372, |
| "grad_norm": 0.2764706611633301, |
| "learning_rate": 2.339839993476745e-05, |
| "loss": 0.0152, |
| "step": 10360 |
| }, |
| { |
| "epoch": 1.062282319196886, |
| "grad_norm": 0.13211829960346222, |
| "learning_rate": 2.3384694767902804e-05, |
| "loss": 0.0147, |
| "step": 10370 |
| }, |
| { |
| "epoch": 1.0633066994468348, |
| "grad_norm": 0.2646767199039459, |
| "learning_rate": 2.3370979412843384e-05, |
| "loss": 0.0115, |
| "step": 10380 |
| }, |
| { |
| "epoch": 1.0643310796967835, |
| "grad_norm": 0.7087327241897583, |
| "learning_rate": 2.3357253886254648e-05, |
| "loss": 0.0186, |
| "step": 10390 |
| }, |
| { |
| "epoch": 1.0653554599467323, |
| "grad_norm": 0.13776163756847382, |
| "learning_rate": 2.3343518204814407e-05, |
| "loss": 0.0144, |
| "step": 10400 |
| }, |
| { |
| "epoch": 1.066379840196681, |
| "grad_norm": 0.5583887100219727, |
| "learning_rate": 2.3329772385212798e-05, |
| "loss": 0.0144, |
| "step": 10410 |
| }, |
| { |
| "epoch": 1.0674042204466299, |
| "grad_norm": 0.18559670448303223, |
| "learning_rate": 2.3316016444152296e-05, |
| "loss": 0.0144, |
| "step": 10420 |
| }, |
| { |
| "epoch": 1.0684286006965786, |
| "grad_norm": 0.24973586201667786, |
| "learning_rate": 2.3302250398347668e-05, |
| "loss": 0.017, |
| "step": 10430 |
| }, |
| { |
| "epoch": 1.0694529809465274, |
| "grad_norm": 0.3422272801399231, |
| "learning_rate": 2.328847426452595e-05, |
| "loss": 0.016, |
| "step": 10440 |
| }, |
| { |
| "epoch": 1.0704773611964762, |
| "grad_norm": 0.18503795564174652, |
| "learning_rate": 2.3274688059426456e-05, |
| "loss": 0.0135, |
| "step": 10450 |
| }, |
| { |
| "epoch": 1.071501741446425, |
| "grad_norm": 0.24961815774440765, |
| "learning_rate": 2.3260891799800715e-05, |
| "loss": 0.0124, |
| "step": 10460 |
| }, |
| { |
| "epoch": 1.0725261216963737, |
| "grad_norm": 0.18159185349941254, |
| "learning_rate": 2.3247085502412482e-05, |
| "loss": 0.0132, |
| "step": 10470 |
| }, |
| { |
| "epoch": 1.0735505019463225, |
| "grad_norm": 0.2926952540874481, |
| "learning_rate": 2.3233269184037717e-05, |
| "loss": 0.0117, |
| "step": 10480 |
| }, |
| { |
| "epoch": 1.0745748821962713, |
| "grad_norm": 0.26654332876205444, |
| "learning_rate": 2.3219442861464547e-05, |
| "loss": 0.0148, |
| "step": 10490 |
| }, |
| { |
| "epoch": 1.07559926244622, |
| "grad_norm": 0.1997380256652832, |
| "learning_rate": 2.3205606551493255e-05, |
| "loss": 0.0126, |
| "step": 10500 |
| }, |
| { |
| "epoch": 1.0766236426961688, |
| "grad_norm": 0.24683986604213715, |
| "learning_rate": 2.319176027093627e-05, |
| "loss": 0.0159, |
| "step": 10510 |
| }, |
| { |
| "epoch": 1.0776480229461176, |
| "grad_norm": 0.16246971487998962, |
| "learning_rate": 2.3177904036618117e-05, |
| "loss": 0.0126, |
| "step": 10520 |
| }, |
| { |
| "epoch": 1.0786724031960664, |
| "grad_norm": 0.25972244143486023, |
| "learning_rate": 2.316403786537544e-05, |
| "loss": 0.0152, |
| "step": 10530 |
| }, |
| { |
| "epoch": 1.0796967834460152, |
| "grad_norm": 0.23779071867465973, |
| "learning_rate": 2.3150161774056933e-05, |
| "loss": 0.0149, |
| "step": 10540 |
| }, |
| { |
| "epoch": 1.080721163695964, |
| "grad_norm": 0.21360233426094055, |
| "learning_rate": 2.313627577952336e-05, |
| "loss": 0.0142, |
| "step": 10550 |
| }, |
| { |
| "epoch": 1.0817455439459127, |
| "grad_norm": 0.43647119402885437, |
| "learning_rate": 2.312237989864752e-05, |
| "loss": 0.0146, |
| "step": 10560 |
| }, |
| { |
| "epoch": 1.0827699241958615, |
| "grad_norm": 0.4088636338710785, |
| "learning_rate": 2.3108474148314205e-05, |
| "loss": 0.0144, |
| "step": 10570 |
| }, |
| { |
| "epoch": 1.0837943044458103, |
| "grad_norm": 0.18082693219184875, |
| "learning_rate": 2.309455854542023e-05, |
| "loss": 0.0184, |
| "step": 10580 |
| }, |
| { |
| "epoch": 1.084818684695759, |
| "grad_norm": 0.14174991846084595, |
| "learning_rate": 2.308063310687435e-05, |
| "loss": 0.0188, |
| "step": 10590 |
| }, |
| { |
| "epoch": 1.0858430649457078, |
| "grad_norm": 0.40196195244789124, |
| "learning_rate": 2.306669784959729e-05, |
| "loss": 0.0148, |
| "step": 10600 |
| }, |
| { |
| "epoch": 1.0868674451956566, |
| "grad_norm": 0.21454943716526031, |
| "learning_rate": 2.3052752790521703e-05, |
| "loss": 0.0158, |
| "step": 10610 |
| }, |
| { |
| "epoch": 1.0878918254456054, |
| "grad_norm": 0.9748015999794006, |
| "learning_rate": 2.3038797946592152e-05, |
| "loss": 0.0135, |
| "step": 10620 |
| }, |
| { |
| "epoch": 1.0889162056955541, |
| "grad_norm": 0.2503933310508728, |
| "learning_rate": 2.3024833334765087e-05, |
| "loss": 0.0137, |
| "step": 10630 |
| }, |
| { |
| "epoch": 1.089940585945503, |
| "grad_norm": 0.7081142663955688, |
| "learning_rate": 2.301085897200883e-05, |
| "loss": 0.0149, |
| "step": 10640 |
| }, |
| { |
| "epoch": 1.0909649661954517, |
| "grad_norm": 0.2045382559299469, |
| "learning_rate": 2.299687487530354e-05, |
| "loss": 0.0152, |
| "step": 10650 |
| }, |
| { |
| "epoch": 1.0919893464454005, |
| "grad_norm": 0.1558094024658203, |
| "learning_rate": 2.2982881061641228e-05, |
| "loss": 0.0129, |
| "step": 10660 |
| }, |
| { |
| "epoch": 1.0930137266953492, |
| "grad_norm": 0.2741594910621643, |
| "learning_rate": 2.296887754802569e-05, |
| "loss": 0.0133, |
| "step": 10670 |
| }, |
| { |
| "epoch": 1.094038106945298, |
| "grad_norm": 0.14990895986557007, |
| "learning_rate": 2.295486435147251e-05, |
| "loss": 0.0134, |
| "step": 10680 |
| }, |
| { |
| "epoch": 1.0950624871952468, |
| "grad_norm": 0.14671127498149872, |
| "learning_rate": 2.294084148900905e-05, |
| "loss": 0.0141, |
| "step": 10690 |
| }, |
| { |
| "epoch": 1.0960868674451956, |
| "grad_norm": 0.13317522406578064, |
| "learning_rate": 2.292680897767441e-05, |
| "loss": 0.0139, |
| "step": 10700 |
| }, |
| { |
| "epoch": 1.0971112476951443, |
| "grad_norm": 0.4665707051753998, |
| "learning_rate": 2.291276683451941e-05, |
| "loss": 0.0133, |
| "step": 10710 |
| }, |
| { |
| "epoch": 1.0981356279450931, |
| "grad_norm": 0.3788447678089142, |
| "learning_rate": 2.2898715076606584e-05, |
| "loss": 0.0127, |
| "step": 10720 |
| }, |
| { |
| "epoch": 1.099160008195042, |
| "grad_norm": 0.20421528816223145, |
| "learning_rate": 2.2884653721010135e-05, |
| "loss": 0.0126, |
| "step": 10730 |
| }, |
| { |
| "epoch": 1.100184388444991, |
| "grad_norm": 0.21685431897640228, |
| "learning_rate": 2.2870582784815946e-05, |
| "loss": 0.0121, |
| "step": 10740 |
| }, |
| { |
| "epoch": 1.1012087686949394, |
| "grad_norm": 0.14899574220180511, |
| "learning_rate": 2.2856502285121525e-05, |
| "loss": 0.0164, |
| "step": 10750 |
| }, |
| { |
| "epoch": 1.1022331489448884, |
| "grad_norm": 0.16324041783809662, |
| "learning_rate": 2.2842412239036004e-05, |
| "loss": 0.0144, |
| "step": 10760 |
| }, |
| { |
| "epoch": 1.1032575291948372, |
| "grad_norm": 0.19038313627243042, |
| "learning_rate": 2.2828312663680125e-05, |
| "loss": 0.0142, |
| "step": 10770 |
| }, |
| { |
| "epoch": 1.104281909444786, |
| "grad_norm": 0.25058111548423767, |
| "learning_rate": 2.2814203576186194e-05, |
| "loss": 0.0166, |
| "step": 10780 |
| }, |
| { |
| "epoch": 1.1053062896947348, |
| "grad_norm": 0.28681570291519165, |
| "learning_rate": 2.2800084993698093e-05, |
| "loss": 0.0128, |
| "step": 10790 |
| }, |
| { |
| "epoch": 1.1063306699446835, |
| "grad_norm": 0.24675802886486053, |
| "learning_rate": 2.2785956933371224e-05, |
| "loss": 0.0117, |
| "step": 10800 |
| }, |
| { |
| "epoch": 1.1073550501946323, |
| "grad_norm": 0.2977922856807709, |
| "learning_rate": 2.2771819412372505e-05, |
| "loss": 0.0166, |
| "step": 10810 |
| }, |
| { |
| "epoch": 1.108379430444581, |
| "grad_norm": 0.14832274615764618, |
| "learning_rate": 2.2757672447880374e-05, |
| "loss": 0.0103, |
| "step": 10820 |
| }, |
| { |
| "epoch": 1.1094038106945299, |
| "grad_norm": 0.16521455347537994, |
| "learning_rate": 2.2743516057084712e-05, |
| "loss": 0.0138, |
| "step": 10830 |
| }, |
| { |
| "epoch": 1.1104281909444786, |
| "grad_norm": 0.16540171205997467, |
| "learning_rate": 2.2729350257186877e-05, |
| "loss": 0.0123, |
| "step": 10840 |
| }, |
| { |
| "epoch": 1.1114525711944274, |
| "grad_norm": 0.20303165912628174, |
| "learning_rate": 2.2715175065399655e-05, |
| "loss": 0.0175, |
| "step": 10850 |
| }, |
| { |
| "epoch": 1.1124769514443762, |
| "grad_norm": 0.3671548366546631, |
| "learning_rate": 2.270099049894723e-05, |
| "loss": 0.013, |
| "step": 10860 |
| }, |
| { |
| "epoch": 1.113501331694325, |
| "grad_norm": 0.1604735404253006, |
| "learning_rate": 2.268679657506519e-05, |
| "loss": 0.0116, |
| "step": 10870 |
| }, |
| { |
| "epoch": 1.1145257119442737, |
| "grad_norm": 0.3387209177017212, |
| "learning_rate": 2.2672593311000498e-05, |
| "loss": 0.0123, |
| "step": 10880 |
| }, |
| { |
| "epoch": 1.1155500921942225, |
| "grad_norm": 0.39342233538627625, |
| "learning_rate": 2.2658380724011452e-05, |
| "loss": 0.0146, |
| "step": 10890 |
| }, |
| { |
| "epoch": 1.1165744724441713, |
| "grad_norm": 0.5959771275520325, |
| "learning_rate": 2.264415883136769e-05, |
| "loss": 0.0149, |
| "step": 10900 |
| }, |
| { |
| "epoch": 1.11759885269412, |
| "grad_norm": 0.48341694474220276, |
| "learning_rate": 2.262992765035015e-05, |
| "loss": 0.0154, |
| "step": 10910 |
| }, |
| { |
| "epoch": 1.1186232329440688, |
| "grad_norm": 0.6976805329322815, |
| "learning_rate": 2.2615687198251062e-05, |
| "loss": 0.0132, |
| "step": 10920 |
| }, |
| { |
| "epoch": 1.1196476131940176, |
| "grad_norm": 0.24011868238449097, |
| "learning_rate": 2.2601437492373914e-05, |
| "loss": 0.0141, |
| "step": 10930 |
| }, |
| { |
| "epoch": 1.1206719934439664, |
| "grad_norm": 0.4414004385471344, |
| "learning_rate": 2.2587178550033444e-05, |
| "loss": 0.0145, |
| "step": 10940 |
| }, |
| { |
| "epoch": 1.1216963736939152, |
| "grad_norm": 0.1627817004919052, |
| "learning_rate": 2.2572910388555615e-05, |
| "loss": 0.0133, |
| "step": 10950 |
| }, |
| { |
| "epoch": 1.122720753943864, |
| "grad_norm": 0.12267287820577621, |
| "learning_rate": 2.255863302527758e-05, |
| "loss": 0.0144, |
| "step": 10960 |
| }, |
| { |
| "epoch": 1.1237451341938127, |
| "grad_norm": 0.23989459872245789, |
| "learning_rate": 2.2544346477547685e-05, |
| "loss": 0.0151, |
| "step": 10970 |
| }, |
| { |
| "epoch": 1.1247695144437615, |
| "grad_norm": 0.2303701788187027, |
| "learning_rate": 2.253005076272544e-05, |
| "loss": 0.0153, |
| "step": 10980 |
| }, |
| { |
| "epoch": 1.1257938946937103, |
| "grad_norm": 0.14976780116558075, |
| "learning_rate": 2.2515745898181477e-05, |
| "loss": 0.017, |
| "step": 10990 |
| }, |
| { |
| "epoch": 1.126818274943659, |
| "grad_norm": 0.25021934509277344, |
| "learning_rate": 2.250143190129756e-05, |
| "loss": 0.0126, |
| "step": 11000 |
| }, |
| { |
| "epoch": 1.1278426551936078, |
| "grad_norm": 0.27183032035827637, |
| "learning_rate": 2.2487108789466547e-05, |
| "loss": 0.013, |
| "step": 11010 |
| }, |
| { |
| "epoch": 1.1288670354435566, |
| "grad_norm": 0.35330551862716675, |
| "learning_rate": 2.247277658009236e-05, |
| "loss": 0.0166, |
| "step": 11020 |
| }, |
| { |
| "epoch": 1.1298914156935054, |
| "grad_norm": 0.31392136216163635, |
| "learning_rate": 2.245843529059e-05, |
| "loss": 0.012, |
| "step": 11030 |
| }, |
| { |
| "epoch": 1.1309157959434541, |
| "grad_norm": 0.14260123670101166, |
| "learning_rate": 2.244408493838547e-05, |
| "loss": 0.0136, |
| "step": 11040 |
| }, |
| { |
| "epoch": 1.131940176193403, |
| "grad_norm": 0.1784290075302124, |
| "learning_rate": 2.242972554091581e-05, |
| "loss": 0.0142, |
| "step": 11050 |
| }, |
| { |
| "epoch": 1.1329645564433517, |
| "grad_norm": 0.39298418164253235, |
| "learning_rate": 2.2415357115629045e-05, |
| "loss": 0.0138, |
| "step": 11060 |
| }, |
| { |
| "epoch": 1.1339889366933005, |
| "grad_norm": 0.18088243901729584, |
| "learning_rate": 2.2400979679984162e-05, |
| "loss": 0.0109, |
| "step": 11070 |
| }, |
| { |
| "epoch": 1.1350133169432493, |
| "grad_norm": 0.19047708809375763, |
| "learning_rate": 2.2386593251451096e-05, |
| "loss": 0.0157, |
| "step": 11080 |
| }, |
| { |
| "epoch": 1.136037697193198, |
| "grad_norm": 0.21795836091041565, |
| "learning_rate": 2.2372197847510722e-05, |
| "loss": 0.0172, |
| "step": 11090 |
| }, |
| { |
| "epoch": 1.1370620774431468, |
| "grad_norm": 0.184831902384758, |
| "learning_rate": 2.2357793485654813e-05, |
| "loss": 0.0139, |
| "step": 11100 |
| }, |
| { |
| "epoch": 1.1380864576930958, |
| "grad_norm": 0.24509915709495544, |
| "learning_rate": 2.2343380183386017e-05, |
| "loss": 0.0129, |
| "step": 11110 |
| }, |
| { |
| "epoch": 1.1391108379430444, |
| "grad_norm": 0.1284109205007553, |
| "learning_rate": 2.232895795821787e-05, |
| "loss": 0.0142, |
| "step": 11120 |
| }, |
| { |
| "epoch": 1.1401352181929933, |
| "grad_norm": 0.23673005402088165, |
| "learning_rate": 2.2314526827674723e-05, |
| "loss": 0.013, |
| "step": 11130 |
| }, |
| { |
| "epoch": 1.141159598442942, |
| "grad_norm": 0.1618281453847885, |
| "learning_rate": 2.230008680929177e-05, |
| "loss": 0.0117, |
| "step": 11140 |
| }, |
| { |
| "epoch": 1.142183978692891, |
| "grad_norm": 0.2070453017950058, |
| "learning_rate": 2.2285637920614987e-05, |
| "loss": 0.0137, |
| "step": 11150 |
| }, |
| { |
| "epoch": 1.1432083589428397, |
| "grad_norm": 0.5458146333694458, |
| "learning_rate": 2.2271180179201143e-05, |
| "loss": 0.0156, |
| "step": 11160 |
| }, |
| { |
| "epoch": 1.1442327391927885, |
| "grad_norm": 0.18877986073493958, |
| "learning_rate": 2.225671360261775e-05, |
| "loss": 0.0139, |
| "step": 11170 |
| }, |
| { |
| "epoch": 1.1452571194427372, |
| "grad_norm": 0.23341993987560272, |
| "learning_rate": 2.2242238208443068e-05, |
| "loss": 0.0159, |
| "step": 11180 |
| }, |
| { |
| "epoch": 1.146281499692686, |
| "grad_norm": 0.1796734780073166, |
| "learning_rate": 2.2227754014266063e-05, |
| "loss": 0.0147, |
| "step": 11190 |
| }, |
| { |
| "epoch": 1.1473058799426348, |
| "grad_norm": 0.18588471412658691, |
| "learning_rate": 2.22132610376864e-05, |
| "loss": 0.0185, |
| "step": 11200 |
| }, |
| { |
| "epoch": 1.1483302601925836, |
| "grad_norm": 0.41951805353164673, |
| "learning_rate": 2.219875929631441e-05, |
| "loss": 0.0122, |
| "step": 11210 |
| }, |
| { |
| "epoch": 1.1493546404425323, |
| "grad_norm": 0.15966898202896118, |
| "learning_rate": 2.2184248807771074e-05, |
| "loss": 0.013, |
| "step": 11220 |
| }, |
| { |
| "epoch": 1.150379020692481, |
| "grad_norm": 0.3857877850532532, |
| "learning_rate": 2.216972958968801e-05, |
| "loss": 0.015, |
| "step": 11230 |
| }, |
| { |
| "epoch": 1.1514034009424299, |
| "grad_norm": 0.7261695265769958, |
| "learning_rate": 2.215520165970743e-05, |
| "loss": 0.0163, |
| "step": 11240 |
| }, |
| { |
| "epoch": 1.1524277811923787, |
| "grad_norm": 0.4193238317966461, |
| "learning_rate": 2.2140665035482143e-05, |
| "loss": 0.0154, |
| "step": 11250 |
| }, |
| { |
| "epoch": 1.1534521614423274, |
| "grad_norm": 0.29892492294311523, |
| "learning_rate": 2.2126119734675518e-05, |
| "loss": 0.0145, |
| "step": 11260 |
| }, |
| { |
| "epoch": 1.1544765416922762, |
| "grad_norm": 0.21619601547718048, |
| "learning_rate": 2.2111565774961463e-05, |
| "loss": 0.0141, |
| "step": 11270 |
| }, |
| { |
| "epoch": 1.155500921942225, |
| "grad_norm": 0.23915475606918335, |
| "learning_rate": 2.2097003174024417e-05, |
| "loss": 0.0129, |
| "step": 11280 |
| }, |
| { |
| "epoch": 1.1565253021921738, |
| "grad_norm": 0.6679841876029968, |
| "learning_rate": 2.2082431949559315e-05, |
| "loss": 0.0137, |
| "step": 11290 |
| }, |
| { |
| "epoch": 1.1575496824421225, |
| "grad_norm": 0.28376731276512146, |
| "learning_rate": 2.2067852119271558e-05, |
| "loss": 0.0125, |
| "step": 11300 |
| }, |
| { |
| "epoch": 1.1585740626920713, |
| "grad_norm": 0.16744264960289001, |
| "learning_rate": 2.2053263700877026e-05, |
| "loss": 0.0132, |
| "step": 11310 |
| }, |
| { |
| "epoch": 1.15959844294202, |
| "grad_norm": 0.422137975692749, |
| "learning_rate": 2.2038666712102015e-05, |
| "loss": 0.0134, |
| "step": 11320 |
| }, |
| { |
| "epoch": 1.1606228231919689, |
| "grad_norm": 0.30519652366638184, |
| "learning_rate": 2.2024061170683246e-05, |
| "loss": 0.017, |
| "step": 11330 |
| }, |
| { |
| "epoch": 1.1616472034419176, |
| "grad_norm": 0.3746437728404999, |
| "learning_rate": 2.200944709436783e-05, |
| "loss": 0.0142, |
| "step": 11340 |
| }, |
| { |
| "epoch": 1.1626715836918664, |
| "grad_norm": 0.23608554899692535, |
| "learning_rate": 2.1994824500913246e-05, |
| "loss": 0.0167, |
| "step": 11350 |
| }, |
| { |
| "epoch": 1.1636959639418152, |
| "grad_norm": 0.23530133068561554, |
| "learning_rate": 2.1980193408087322e-05, |
| "loss": 0.0167, |
| "step": 11360 |
| }, |
| { |
| "epoch": 1.164720344191764, |
| "grad_norm": 0.5710408091545105, |
| "learning_rate": 2.1965553833668224e-05, |
| "loss": 0.011, |
| "step": 11370 |
| }, |
| { |
| "epoch": 1.1657447244417127, |
| "grad_norm": 0.373941034078598, |
| "learning_rate": 2.195090579544441e-05, |
| "loss": 0.0146, |
| "step": 11380 |
| }, |
| { |
| "epoch": 1.1667691046916615, |
| "grad_norm": 0.17240716516971588, |
| "learning_rate": 2.193624931121462e-05, |
| "loss": 0.015, |
| "step": 11390 |
| }, |
| { |
| "epoch": 1.1677934849416103, |
| "grad_norm": 0.3455638885498047, |
| "learning_rate": 2.1921584398787872e-05, |
| "loss": 0.0157, |
| "step": 11400 |
| }, |
| { |
| "epoch": 1.168817865191559, |
| "grad_norm": 0.2570631802082062, |
| "learning_rate": 2.1906911075983418e-05, |
| "loss": 0.0138, |
| "step": 11410 |
| }, |
| { |
| "epoch": 1.1698422454415078, |
| "grad_norm": 0.3442557454109192, |
| "learning_rate": 2.1892229360630728e-05, |
| "loss": 0.0141, |
| "step": 11420 |
| }, |
| { |
| "epoch": 1.1708666256914566, |
| "grad_norm": 0.3789099156856537, |
| "learning_rate": 2.187753927056947e-05, |
| "loss": 0.0144, |
| "step": 11430 |
| }, |
| { |
| "epoch": 1.1718910059414054, |
| "grad_norm": 0.2516676187515259, |
| "learning_rate": 2.186284082364949e-05, |
| "loss": 0.0176, |
| "step": 11440 |
| }, |
| { |
| "epoch": 1.1729153861913542, |
| "grad_norm": 0.4092411696910858, |
| "learning_rate": 2.1848134037730786e-05, |
| "loss": 0.016, |
| "step": 11450 |
| }, |
| { |
| "epoch": 1.173939766441303, |
| "grad_norm": 0.17516003549098969, |
| "learning_rate": 2.183341893068349e-05, |
| "loss": 0.0133, |
| "step": 11460 |
| }, |
| { |
| "epoch": 1.1749641466912517, |
| "grad_norm": 0.2080044001340866, |
| "learning_rate": 2.1818695520387842e-05, |
| "loss": 0.0158, |
| "step": 11470 |
| }, |
| { |
| "epoch": 1.1759885269412005, |
| "grad_norm": 0.6698071360588074, |
| "learning_rate": 2.1803963824734173e-05, |
| "loss": 0.0143, |
| "step": 11480 |
| }, |
| { |
| "epoch": 1.1770129071911493, |
| "grad_norm": 0.8943179249763489, |
| "learning_rate": 2.1789223861622887e-05, |
| "loss": 0.013, |
| "step": 11490 |
| }, |
| { |
| "epoch": 1.1780372874410983, |
| "grad_norm": 0.32973042130470276, |
| "learning_rate": 2.1774475648964428e-05, |
| "loss": 0.0125, |
| "step": 11500 |
| }, |
| { |
| "epoch": 1.1790616676910468, |
| "grad_norm": 0.1794813573360443, |
| "learning_rate": 2.1759719204679267e-05, |
| "loss": 0.0137, |
| "step": 11510 |
| }, |
| { |
| "epoch": 1.1800860479409958, |
| "grad_norm": 0.47253793478012085, |
| "learning_rate": 2.174495454669787e-05, |
| "loss": 0.0161, |
| "step": 11520 |
| }, |
| { |
| "epoch": 1.1811104281909444, |
| "grad_norm": 1.2170356512069702, |
| "learning_rate": 2.1730181692960692e-05, |
| "loss": 0.0151, |
| "step": 11530 |
| }, |
| { |
| "epoch": 1.1821348084408934, |
| "grad_norm": 0.31432008743286133, |
| "learning_rate": 2.171540066141814e-05, |
| "loss": 0.0139, |
| "step": 11540 |
| }, |
| { |
| "epoch": 1.1831591886908421, |
| "grad_norm": 0.20243075489997864, |
| "learning_rate": 2.170061147003056e-05, |
| "loss": 0.0137, |
| "step": 11550 |
| }, |
| { |
| "epoch": 1.184183568940791, |
| "grad_norm": 0.18004322052001953, |
| "learning_rate": 2.1685814136768225e-05, |
| "loss": 0.012, |
| "step": 11560 |
| }, |
| { |
| "epoch": 1.1852079491907397, |
| "grad_norm": 0.24439069628715515, |
| "learning_rate": 2.167100867961128e-05, |
| "loss": 0.0142, |
| "step": 11570 |
| }, |
| { |
| "epoch": 1.1862323294406885, |
| "grad_norm": 0.28954413533210754, |
| "learning_rate": 2.1656195116549756e-05, |
| "loss": 0.0137, |
| "step": 11580 |
| }, |
| { |
| "epoch": 1.1872567096906372, |
| "grad_norm": 0.6649225354194641, |
| "learning_rate": 2.1641373465583527e-05, |
| "loss": 0.0149, |
| "step": 11590 |
| }, |
| { |
| "epoch": 1.188281089940586, |
| "grad_norm": 0.15726172924041748, |
| "learning_rate": 2.1626543744722292e-05, |
| "loss": 0.0125, |
| "step": 11600 |
| }, |
| { |
| "epoch": 1.1893054701905348, |
| "grad_norm": 0.593467652797699, |
| "learning_rate": 2.161170597198557e-05, |
| "loss": 0.0155, |
| "step": 11610 |
| }, |
| { |
| "epoch": 1.1903298504404836, |
| "grad_norm": 0.4286263585090637, |
| "learning_rate": 2.1596860165402645e-05, |
| "loss": 0.0145, |
| "step": 11620 |
| }, |
| { |
| "epoch": 1.1913542306904323, |
| "grad_norm": 0.196187824010849, |
| "learning_rate": 2.1582006343012576e-05, |
| "loss": 0.0129, |
| "step": 11630 |
| }, |
| { |
| "epoch": 1.1923786109403811, |
| "grad_norm": 0.18965528905391693, |
| "learning_rate": 2.156714452286416e-05, |
| "loss": 0.0146, |
| "step": 11640 |
| }, |
| { |
| "epoch": 1.1934029911903299, |
| "grad_norm": 0.2510643005371094, |
| "learning_rate": 2.155227472301591e-05, |
| "loss": 0.0144, |
| "step": 11650 |
| }, |
| { |
| "epoch": 1.1944273714402787, |
| "grad_norm": 0.24945417046546936, |
| "learning_rate": 2.1537396961536027e-05, |
| "loss": 0.0128, |
| "step": 11660 |
| }, |
| { |
| "epoch": 1.1954517516902274, |
| "grad_norm": 0.24243700504302979, |
| "learning_rate": 2.1522511256502403e-05, |
| "loss": 0.0139, |
| "step": 11670 |
| }, |
| { |
| "epoch": 1.1964761319401762, |
| "grad_norm": 0.4482676088809967, |
| "learning_rate": 2.1507617626002566e-05, |
| "loss": 0.0133, |
| "step": 11680 |
| }, |
| { |
| "epoch": 1.197500512190125, |
| "grad_norm": 0.7937094569206238, |
| "learning_rate": 2.1492716088133685e-05, |
| "loss": 0.0136, |
| "step": 11690 |
| }, |
| { |
| "epoch": 1.1985248924400738, |
| "grad_norm": 0.26940056681632996, |
| "learning_rate": 2.1477806661002536e-05, |
| "loss": 0.0158, |
| "step": 11700 |
| }, |
| { |
| "epoch": 1.1995492726900225, |
| "grad_norm": 0.20709142088890076, |
| "learning_rate": 2.146288936272548e-05, |
| "loss": 0.014, |
| "step": 11710 |
| }, |
| { |
| "epoch": 1.2005736529399713, |
| "grad_norm": 0.5331788659095764, |
| "learning_rate": 2.1447964211428433e-05, |
| "loss": 0.0105, |
| "step": 11720 |
| }, |
| { |
| "epoch": 1.20159803318992, |
| "grad_norm": 0.4134519100189209, |
| "learning_rate": 2.1433031225246865e-05, |
| "loss": 0.0166, |
| "step": 11730 |
| }, |
| { |
| "epoch": 1.2026224134398689, |
| "grad_norm": 0.20942330360412598, |
| "learning_rate": 2.1418090422325766e-05, |
| "loss": 0.015, |
| "step": 11740 |
| }, |
| { |
| "epoch": 1.2036467936898176, |
| "grad_norm": 0.9071303009986877, |
| "learning_rate": 2.1403141820819623e-05, |
| "loss": 0.0157, |
| "step": 11750 |
| }, |
| { |
| "epoch": 1.2046711739397664, |
| "grad_norm": 0.3482285737991333, |
| "learning_rate": 2.1388185438892377e-05, |
| "loss": 0.0148, |
| "step": 11760 |
| }, |
| { |
| "epoch": 1.2056955541897152, |
| "grad_norm": 0.20598337054252625, |
| "learning_rate": 2.1373221294717465e-05, |
| "loss": 0.0125, |
| "step": 11770 |
| }, |
| { |
| "epoch": 1.206719934439664, |
| "grad_norm": 0.2921450436115265, |
| "learning_rate": 2.1358249406477724e-05, |
| "loss": 0.0129, |
| "step": 11780 |
| }, |
| { |
| "epoch": 1.2077443146896127, |
| "grad_norm": 0.38918814063072205, |
| "learning_rate": 2.1343269792365408e-05, |
| "loss": 0.013, |
| "step": 11790 |
| }, |
| { |
| "epoch": 1.2087686949395615, |
| "grad_norm": 0.17799413204193115, |
| "learning_rate": 2.132828247058217e-05, |
| "loss": 0.0141, |
| "step": 11800 |
| }, |
| { |
| "epoch": 1.2097930751895103, |
| "grad_norm": 0.19724677503108978, |
| "learning_rate": 2.1313287459339004e-05, |
| "loss": 0.0122, |
| "step": 11810 |
| }, |
| { |
| "epoch": 1.210817455439459, |
| "grad_norm": 0.2886558771133423, |
| "learning_rate": 2.1298284776856276e-05, |
| "loss": 0.015, |
| "step": 11820 |
| }, |
| { |
| "epoch": 1.2118418356894078, |
| "grad_norm": 0.41239988803863525, |
| "learning_rate": 2.128327444136365e-05, |
| "loss": 0.0135, |
| "step": 11830 |
| }, |
| { |
| "epoch": 1.2128662159393566, |
| "grad_norm": 0.27603039145469666, |
| "learning_rate": 2.1268256471100105e-05, |
| "loss": 0.0129, |
| "step": 11840 |
| }, |
| { |
| "epoch": 1.2138905961893054, |
| "grad_norm": 0.25559672713279724, |
| "learning_rate": 2.1253230884313893e-05, |
| "loss": 0.0146, |
| "step": 11850 |
| }, |
| { |
| "epoch": 1.2149149764392542, |
| "grad_norm": 0.6757758259773254, |
| "learning_rate": 2.123819769926251e-05, |
| "loss": 0.0144, |
| "step": 11860 |
| }, |
| { |
| "epoch": 1.215939356689203, |
| "grad_norm": 0.1627424955368042, |
| "learning_rate": 2.1223156934212702e-05, |
| "loss": 0.0172, |
| "step": 11870 |
| }, |
| { |
| "epoch": 1.2169637369391517, |
| "grad_norm": 0.3842763602733612, |
| "learning_rate": 2.1208108607440418e-05, |
| "loss": 0.0126, |
| "step": 11880 |
| }, |
| { |
| "epoch": 1.2179881171891007, |
| "grad_norm": 0.32898038625717163, |
| "learning_rate": 2.119305273723078e-05, |
| "loss": 0.0124, |
| "step": 11890 |
| }, |
| { |
| "epoch": 1.2190124974390493, |
| "grad_norm": 0.1357547640800476, |
| "learning_rate": 2.1177989341878106e-05, |
| "loss": 0.0112, |
| "step": 11900 |
| }, |
| { |
| "epoch": 1.2200368776889983, |
| "grad_norm": 0.2292754203081131, |
| "learning_rate": 2.116291843968583e-05, |
| "loss": 0.0153, |
| "step": 11910 |
| }, |
| { |
| "epoch": 1.2210612579389468, |
| "grad_norm": 0.15151797235012054, |
| "learning_rate": 2.114784004896653e-05, |
| "loss": 0.0132, |
| "step": 11920 |
| }, |
| { |
| "epoch": 1.2220856381888958, |
| "grad_norm": 0.19245703518390656, |
| "learning_rate": 2.113275418804187e-05, |
| "loss": 0.0131, |
| "step": 11930 |
| }, |
| { |
| "epoch": 1.2231100184388446, |
| "grad_norm": 0.20630048215389252, |
| "learning_rate": 2.1117660875242583e-05, |
| "loss": 0.0149, |
| "step": 11940 |
| }, |
| { |
| "epoch": 1.2241343986887934, |
| "grad_norm": 0.2435985654592514, |
| "learning_rate": 2.1102560128908478e-05, |
| "loss": 0.0157, |
| "step": 11950 |
| }, |
| { |
| "epoch": 1.2251587789387421, |
| "grad_norm": 0.3530169129371643, |
| "learning_rate": 2.1087451967388384e-05, |
| "loss": 0.0143, |
| "step": 11960 |
| }, |
| { |
| "epoch": 1.226183159188691, |
| "grad_norm": 0.5680727362632751, |
| "learning_rate": 2.1072336409040135e-05, |
| "loss": 0.0132, |
| "step": 11970 |
| }, |
| { |
| "epoch": 1.2272075394386397, |
| "grad_norm": 0.7829403877258301, |
| "learning_rate": 2.1057213472230573e-05, |
| "loss": 0.0201, |
| "step": 11980 |
| }, |
| { |
| "epoch": 1.2282319196885885, |
| "grad_norm": 0.1574874073266983, |
| "learning_rate": 2.1042083175335478e-05, |
| "loss": 0.0129, |
| "step": 11990 |
| }, |
| { |
| "epoch": 1.2292562999385372, |
| "grad_norm": 0.13306061923503876, |
| "learning_rate": 2.1026945536739595e-05, |
| "loss": 0.0148, |
| "step": 12000 |
| }, |
| { |
| "epoch": 1.230280680188486, |
| "grad_norm": 0.2387513965368271, |
| "learning_rate": 2.1011800574836586e-05, |
| "loss": 0.0135, |
| "step": 12010 |
| }, |
| { |
| "epoch": 1.2313050604384348, |
| "grad_norm": 0.20105187594890594, |
| "learning_rate": 2.0996648308028992e-05, |
| "loss": 0.0128, |
| "step": 12020 |
| }, |
| { |
| "epoch": 1.2323294406883836, |
| "grad_norm": 0.4251648783683777, |
| "learning_rate": 2.0981488754728266e-05, |
| "loss": 0.0132, |
| "step": 12030 |
| }, |
| { |
| "epoch": 1.2333538209383323, |
| "grad_norm": 0.2234528809785843, |
| "learning_rate": 2.0966321933354685e-05, |
| "loss": 0.0134, |
| "step": 12040 |
| }, |
| { |
| "epoch": 1.2343782011882811, |
| "grad_norm": 0.36645951867103577, |
| "learning_rate": 2.095114786233736e-05, |
| "loss": 0.0134, |
| "step": 12050 |
| }, |
| { |
| "epoch": 1.23540258143823, |
| "grad_norm": 0.25636523962020874, |
| "learning_rate": 2.0935966560114236e-05, |
| "loss": 0.0137, |
| "step": 12060 |
| }, |
| { |
| "epoch": 1.2364269616881787, |
| "grad_norm": 0.5063742995262146, |
| "learning_rate": 2.092077804513201e-05, |
| "loss": 0.0122, |
| "step": 12070 |
| }, |
| { |
| "epoch": 1.2374513419381274, |
| "grad_norm": 0.2536717653274536, |
| "learning_rate": 2.0905582335846167e-05, |
| "loss": 0.0139, |
| "step": 12080 |
| }, |
| { |
| "epoch": 1.2384757221880762, |
| "grad_norm": 0.14671505987644196, |
| "learning_rate": 2.0890379450720927e-05, |
| "loss": 0.0153, |
| "step": 12090 |
| }, |
| { |
| "epoch": 1.239500102438025, |
| "grad_norm": 0.1218695119023323, |
| "learning_rate": 2.087516940822923e-05, |
| "loss": 0.013, |
| "step": 12100 |
| }, |
| { |
| "epoch": 1.2405244826879738, |
| "grad_norm": 0.24892984330654144, |
| "learning_rate": 2.085995222685271e-05, |
| "loss": 0.0161, |
| "step": 12110 |
| }, |
| { |
| "epoch": 1.2415488629379225, |
| "grad_norm": 0.31986692547798157, |
| "learning_rate": 2.084472792508167e-05, |
| "loss": 0.0137, |
| "step": 12120 |
| }, |
| { |
| "epoch": 1.2425732431878713, |
| "grad_norm": 0.22231939435005188, |
| "learning_rate": 2.082949652141508e-05, |
| "loss": 0.0142, |
| "step": 12130 |
| }, |
| { |
| "epoch": 1.24359762343782, |
| "grad_norm": 0.3626723289489746, |
| "learning_rate": 2.081425803436054e-05, |
| "loss": 0.015, |
| "step": 12140 |
| }, |
| { |
| "epoch": 1.2446220036877689, |
| "grad_norm": 0.32705554366111755, |
| "learning_rate": 2.079901248243423e-05, |
| "loss": 0.0154, |
| "step": 12150 |
| }, |
| { |
| "epoch": 1.2456463839377176, |
| "grad_norm": 0.47443267703056335, |
| "learning_rate": 2.078375988416095e-05, |
| "loss": 0.0145, |
| "step": 12160 |
| }, |
| { |
| "epoch": 1.2466707641876664, |
| "grad_norm": 0.15707479417324066, |
| "learning_rate": 2.076850025807404e-05, |
| "loss": 0.0164, |
| "step": 12170 |
| }, |
| { |
| "epoch": 1.2476951444376152, |
| "grad_norm": 0.3174891769886017, |
| "learning_rate": 2.0753233622715374e-05, |
| "loss": 0.0134, |
| "step": 12180 |
| }, |
| { |
| "epoch": 1.248719524687564, |
| "grad_norm": 0.23719307780265808, |
| "learning_rate": 2.073795999663536e-05, |
| "loss": 0.0138, |
| "step": 12190 |
| }, |
| { |
| "epoch": 1.2497439049375128, |
| "grad_norm": 0.24166852235794067, |
| "learning_rate": 2.07226793983929e-05, |
| "loss": 0.0131, |
| "step": 12200 |
| }, |
| { |
| "epoch": 1.2507682851874615, |
| "grad_norm": 0.12627413868904114, |
| "learning_rate": 2.070739184655536e-05, |
| "loss": 0.0127, |
| "step": 12210 |
| }, |
| { |
| "epoch": 1.2517926654374103, |
| "grad_norm": 0.7690250873565674, |
| "learning_rate": 2.069209735969855e-05, |
| "loss": 0.0161, |
| "step": 12220 |
| }, |
| { |
| "epoch": 1.252817045687359, |
| "grad_norm": 0.8240934610366821, |
| "learning_rate": 2.067679595640672e-05, |
| "loss": 0.0154, |
| "step": 12230 |
| }, |
| { |
| "epoch": 1.2538414259373079, |
| "grad_norm": 0.2521305978298187, |
| "learning_rate": 2.066148765527252e-05, |
| "loss": 0.0125, |
| "step": 12240 |
| }, |
| { |
| "epoch": 1.2548658061872566, |
| "grad_norm": 0.6687923669815063, |
| "learning_rate": 2.0646172474896976e-05, |
| "loss": 0.0122, |
| "step": 12250 |
| }, |
| { |
| "epoch": 1.2558901864372056, |
| "grad_norm": 0.23636089265346527, |
| "learning_rate": 2.063085043388947e-05, |
| "loss": 0.0147, |
| "step": 12260 |
| }, |
| { |
| "epoch": 1.2569145666871542, |
| "grad_norm": 0.23536916077136993, |
| "learning_rate": 2.0615521550867732e-05, |
| "loss": 0.013, |
| "step": 12270 |
| }, |
| { |
| "epoch": 1.2579389469371032, |
| "grad_norm": 0.1843791902065277, |
| "learning_rate": 2.0600185844457803e-05, |
| "loss": 0.013, |
| "step": 12280 |
| }, |
| { |
| "epoch": 1.2589633271870517, |
| "grad_norm": 0.2729344367980957, |
| "learning_rate": 2.058484333329401e-05, |
| "loss": 0.0143, |
| "step": 12290 |
| }, |
| { |
| "epoch": 1.2599877074370007, |
| "grad_norm": 0.34917712211608887, |
| "learning_rate": 2.0569494036018955e-05, |
| "loss": 0.0175, |
| "step": 12300 |
| }, |
| { |
| "epoch": 1.2610120876869493, |
| "grad_norm": 0.21293897926807404, |
| "learning_rate": 2.055413797128347e-05, |
| "loss": 0.0163, |
| "step": 12310 |
| }, |
| { |
| "epoch": 1.2620364679368983, |
| "grad_norm": 0.1807810217142105, |
| "learning_rate": 2.0538775157746634e-05, |
| "loss": 0.0151, |
| "step": 12320 |
| }, |
| { |
| "epoch": 1.2630608481868468, |
| "grad_norm": 0.45775702595710754, |
| "learning_rate": 2.0523405614075706e-05, |
| "loss": 0.0153, |
| "step": 12330 |
| }, |
| { |
| "epoch": 1.2640852284367958, |
| "grad_norm": 0.3105059862136841, |
| "learning_rate": 2.0508029358946132e-05, |
| "loss": 0.0133, |
| "step": 12340 |
| }, |
| { |
| "epoch": 1.2651096086867446, |
| "grad_norm": 0.26105406880378723, |
| "learning_rate": 2.0492646411041515e-05, |
| "loss": 0.0153, |
| "step": 12350 |
| }, |
| { |
| "epoch": 1.2661339889366934, |
| "grad_norm": 0.123297318816185, |
| "learning_rate": 2.0477256789053587e-05, |
| "loss": 0.0135, |
| "step": 12360 |
| }, |
| { |
| "epoch": 1.2671583691866422, |
| "grad_norm": 0.2418176233768463, |
| "learning_rate": 2.0461860511682194e-05, |
| "loss": 0.0138, |
| "step": 12370 |
| }, |
| { |
| "epoch": 1.268182749436591, |
| "grad_norm": 0.2089763730764389, |
| "learning_rate": 2.0446457597635254e-05, |
| "loss": 0.0154, |
| "step": 12380 |
| }, |
| { |
| "epoch": 1.2692071296865397, |
| "grad_norm": 0.3082645833492279, |
| "learning_rate": 2.0431048065628763e-05, |
| "loss": 0.0182, |
| "step": 12390 |
| }, |
| { |
| "epoch": 1.2702315099364885, |
| "grad_norm": 0.4531726539134979, |
| "learning_rate": 2.0415631934386764e-05, |
| "loss": 0.0153, |
| "step": 12400 |
| }, |
| { |
| "epoch": 1.2712558901864373, |
| "grad_norm": 0.23231953382492065, |
| "learning_rate": 2.04002092226413e-05, |
| "loss": 0.0158, |
| "step": 12410 |
| }, |
| { |
| "epoch": 1.272280270436386, |
| "grad_norm": 0.3832436501979828, |
| "learning_rate": 2.038477994913242e-05, |
| "loss": 0.0151, |
| "step": 12420 |
| }, |
| { |
| "epoch": 1.2733046506863348, |
| "grad_norm": 0.20042504370212555, |
| "learning_rate": 2.036934413260816e-05, |
| "loss": 0.0142, |
| "step": 12430 |
| }, |
| { |
| "epoch": 1.2743290309362836, |
| "grad_norm": 0.2518218755722046, |
| "learning_rate": 2.0353901791824473e-05, |
| "loss": 0.0142, |
| "step": 12440 |
| }, |
| { |
| "epoch": 1.2753534111862324, |
| "grad_norm": 0.36102497577667236, |
| "learning_rate": 2.0338452945545275e-05, |
| "loss": 0.014, |
| "step": 12450 |
| }, |
| { |
| "epoch": 1.2763777914361811, |
| "grad_norm": 0.18341436982154846, |
| "learning_rate": 2.0322997612542366e-05, |
| "loss": 0.0134, |
| "step": 12460 |
| }, |
| { |
| "epoch": 1.27740217168613, |
| "grad_norm": 0.3438948690891266, |
| "learning_rate": 2.0307535811595427e-05, |
| "loss": 0.014, |
| "step": 12470 |
| }, |
| { |
| "epoch": 1.2784265519360787, |
| "grad_norm": 0.42463043332099915, |
| "learning_rate": 2.0292067561492013e-05, |
| "loss": 0.0148, |
| "step": 12480 |
| }, |
| { |
| "epoch": 1.2794509321860275, |
| "grad_norm": 0.2740517258644104, |
| "learning_rate": 2.02765928810275e-05, |
| "loss": 0.0139, |
| "step": 12490 |
| }, |
| { |
| "epoch": 1.2804753124359762, |
| "grad_norm": 0.2259746789932251, |
| "learning_rate": 2.0261111789005087e-05, |
| "loss": 0.013, |
| "step": 12500 |
| }, |
| { |
| "epoch": 1.281499692685925, |
| "grad_norm": 0.21311938762664795, |
| "learning_rate": 2.024562430423576e-05, |
| "loss": 0.0152, |
| "step": 12510 |
| }, |
| { |
| "epoch": 1.2825240729358738, |
| "grad_norm": 0.3844241797924042, |
| "learning_rate": 2.0230130445538273e-05, |
| "loss": 0.0143, |
| "step": 12520 |
| }, |
| { |
| "epoch": 1.2835484531858226, |
| "grad_norm": 0.7356323003768921, |
| "learning_rate": 2.0214630231739127e-05, |
| "loss": 0.0149, |
| "step": 12530 |
| }, |
| { |
| "epoch": 1.2845728334357713, |
| "grad_norm": 0.22849901020526886, |
| "learning_rate": 2.0199123681672543e-05, |
| "loss": 0.0133, |
| "step": 12540 |
| }, |
| { |
| "epoch": 1.28559721368572, |
| "grad_norm": 0.337579607963562, |
| "learning_rate": 2.0183610814180437e-05, |
| "loss": 0.0136, |
| "step": 12550 |
| }, |
| { |
| "epoch": 1.2866215939356689, |
| "grad_norm": 0.4233008623123169, |
| "learning_rate": 2.016809164811241e-05, |
| "loss": 0.0125, |
| "step": 12560 |
| }, |
| { |
| "epoch": 1.2876459741856177, |
| "grad_norm": 0.1709970384836197, |
| "learning_rate": 2.015256620232571e-05, |
| "loss": 0.0124, |
| "step": 12570 |
| }, |
| { |
| "epoch": 1.2886703544355664, |
| "grad_norm": 0.47781774401664734, |
| "learning_rate": 2.0137034495685222e-05, |
| "loss": 0.0132, |
| "step": 12580 |
| }, |
| { |
| "epoch": 1.2896947346855152, |
| "grad_norm": 0.23690733313560486, |
| "learning_rate": 2.012149654706343e-05, |
| "loss": 0.015, |
| "step": 12590 |
| }, |
| { |
| "epoch": 1.290719114935464, |
| "grad_norm": 0.32302945852279663, |
| "learning_rate": 2.010595237534041e-05, |
| "loss": 0.015, |
| "step": 12600 |
| }, |
| { |
| "epoch": 1.2917434951854128, |
| "grad_norm": 0.1918087899684906, |
| "learning_rate": 2.0090401999403793e-05, |
| "loss": 0.0147, |
| "step": 12610 |
| }, |
| { |
| "epoch": 1.2927678754353615, |
| "grad_norm": 0.9559868574142456, |
| "learning_rate": 2.007484543814875e-05, |
| "loss": 0.016, |
| "step": 12620 |
| }, |
| { |
| "epoch": 1.2937922556853103, |
| "grad_norm": 0.3171558976173401, |
| "learning_rate": 2.0059282710477972e-05, |
| "loss": 0.014, |
| "step": 12630 |
| }, |
| { |
| "epoch": 1.294816635935259, |
| "grad_norm": 0.3252876400947571, |
| "learning_rate": 2.0043713835301647e-05, |
| "loss": 0.014, |
| "step": 12640 |
| }, |
| { |
| "epoch": 1.295841016185208, |
| "grad_norm": 0.18855667114257812, |
| "learning_rate": 2.0028138831537417e-05, |
| "loss": 0.0132, |
| "step": 12650 |
| }, |
| { |
| "epoch": 1.2968653964351566, |
| "grad_norm": 0.2330816388130188, |
| "learning_rate": 2.0012557718110387e-05, |
| "loss": 0.0157, |
| "step": 12660 |
| }, |
| { |
| "epoch": 1.2978897766851056, |
| "grad_norm": 0.3085130751132965, |
| "learning_rate": 1.9996970513953077e-05, |
| "loss": 0.0148, |
| "step": 12670 |
| }, |
| { |
| "epoch": 1.2989141569350542, |
| "grad_norm": 0.20268884301185608, |
| "learning_rate": 1.9981377238005405e-05, |
| "loss": 0.0135, |
| "step": 12680 |
| }, |
| { |
| "epoch": 1.2999385371850032, |
| "grad_norm": 0.1907525211572647, |
| "learning_rate": 1.9965777909214674e-05, |
| "loss": 0.0144, |
| "step": 12690 |
| }, |
| { |
| "epoch": 1.3009629174349517, |
| "grad_norm": 0.28147563338279724, |
| "learning_rate": 1.9950172546535545e-05, |
| "loss": 0.0143, |
| "step": 12700 |
| }, |
| { |
| "epoch": 1.3019872976849007, |
| "grad_norm": 0.15537601709365845, |
| "learning_rate": 1.9934561168930002e-05, |
| "loss": 0.014, |
| "step": 12710 |
| }, |
| { |
| "epoch": 1.3030116779348493, |
| "grad_norm": 0.6331644058227539, |
| "learning_rate": 1.991894379536734e-05, |
| "loss": 0.0173, |
| "step": 12720 |
| }, |
| { |
| "epoch": 1.3040360581847983, |
| "grad_norm": 0.18725502490997314, |
| "learning_rate": 1.9903320444824137e-05, |
| "loss": 0.0142, |
| "step": 12730 |
| }, |
| { |
| "epoch": 1.305060438434747, |
| "grad_norm": 0.4730084538459778, |
| "learning_rate": 1.9887691136284245e-05, |
| "loss": 0.0186, |
| "step": 12740 |
| }, |
| { |
| "epoch": 1.3060848186846958, |
| "grad_norm": 0.30225059390068054, |
| "learning_rate": 1.9872055888738745e-05, |
| "loss": 0.0122, |
| "step": 12750 |
| }, |
| { |
| "epoch": 1.3071091989346446, |
| "grad_norm": 0.34531739354133606, |
| "learning_rate": 1.985641472118593e-05, |
| "loss": 0.0174, |
| "step": 12760 |
| }, |
| { |
| "epoch": 1.3081335791845934, |
| "grad_norm": 0.15478704869747162, |
| "learning_rate": 1.9840767652631294e-05, |
| "loss": 0.014, |
| "step": 12770 |
| }, |
| { |
| "epoch": 1.3091579594345422, |
| "grad_norm": 0.44145479798316956, |
| "learning_rate": 1.9825114702087513e-05, |
| "loss": 0.0138, |
| "step": 12780 |
| }, |
| { |
| "epoch": 1.310182339684491, |
| "grad_norm": 0.21648284792900085, |
| "learning_rate": 1.9809455888574388e-05, |
| "loss": 0.0118, |
| "step": 12790 |
| }, |
| { |
| "epoch": 1.3112067199344397, |
| "grad_norm": 0.13805831968784332, |
| "learning_rate": 1.9793791231118864e-05, |
| "loss": 0.0158, |
| "step": 12800 |
| }, |
| { |
| "epoch": 1.3122311001843885, |
| "grad_norm": 0.6065384745597839, |
| "learning_rate": 1.9778120748754958e-05, |
| "loss": 0.0142, |
| "step": 12810 |
| }, |
| { |
| "epoch": 1.3132554804343373, |
| "grad_norm": 0.3178669214248657, |
| "learning_rate": 1.97624444605238e-05, |
| "loss": 0.0176, |
| "step": 12820 |
| }, |
| { |
| "epoch": 1.314279860684286, |
| "grad_norm": 0.3320910930633545, |
| "learning_rate": 1.9746762385473552e-05, |
| "loss": 0.0136, |
| "step": 12830 |
| }, |
| { |
| "epoch": 1.3153042409342348, |
| "grad_norm": 0.4127451479434967, |
| "learning_rate": 1.9731074542659414e-05, |
| "loss": 0.0128, |
| "step": 12840 |
| }, |
| { |
| "epoch": 1.3163286211841836, |
| "grad_norm": 0.20876151323318481, |
| "learning_rate": 1.9715380951143603e-05, |
| "loss": 0.0127, |
| "step": 12850 |
| }, |
| { |
| "epoch": 1.3173530014341324, |
| "grad_norm": 0.36355075240135193, |
| "learning_rate": 1.9699681629995295e-05, |
| "loss": 0.0143, |
| "step": 12860 |
| }, |
| { |
| "epoch": 1.3183773816840811, |
| "grad_norm": 0.23014618456363678, |
| "learning_rate": 1.968397659829066e-05, |
| "loss": 0.0147, |
| "step": 12870 |
| }, |
| { |
| "epoch": 1.31940176193403, |
| "grad_norm": 0.21605123579502106, |
| "learning_rate": 1.966826587511279e-05, |
| "loss": 0.0111, |
| "step": 12880 |
| }, |
| { |
| "epoch": 1.3204261421839787, |
| "grad_norm": 0.22227388620376587, |
| "learning_rate": 1.965254947955169e-05, |
| "loss": 0.0135, |
| "step": 12890 |
| }, |
| { |
| "epoch": 1.3214505224339275, |
| "grad_norm": 0.6599821448326111, |
| "learning_rate": 1.963682743070427e-05, |
| "loss": 0.0136, |
| "step": 12900 |
| }, |
| { |
| "epoch": 1.3224749026838762, |
| "grad_norm": 0.27596062421798706, |
| "learning_rate": 1.962109974767429e-05, |
| "loss": 0.0133, |
| "step": 12910 |
| }, |
| { |
| "epoch": 1.323499282933825, |
| "grad_norm": 0.27250346541404724, |
| "learning_rate": 1.960536644957238e-05, |
| "loss": 0.0145, |
| "step": 12920 |
| }, |
| { |
| "epoch": 1.3245236631837738, |
| "grad_norm": 0.25176575779914856, |
| "learning_rate": 1.9589627555515974e-05, |
| "loss": 0.0149, |
| "step": 12930 |
| }, |
| { |
| "epoch": 1.3255480434337226, |
| "grad_norm": 0.44439294934272766, |
| "learning_rate": 1.9573883084629317e-05, |
| "loss": 0.0134, |
| "step": 12940 |
| }, |
| { |
| "epoch": 1.3265724236836713, |
| "grad_norm": 0.4393136501312256, |
| "learning_rate": 1.9558133056043422e-05, |
| "loss": 0.0133, |
| "step": 12950 |
| }, |
| { |
| "epoch": 1.3275968039336201, |
| "grad_norm": 0.16233214735984802, |
| "learning_rate": 1.9542377488896062e-05, |
| "loss": 0.0135, |
| "step": 12960 |
| }, |
| { |
| "epoch": 1.328621184183569, |
| "grad_norm": 0.3534053862094879, |
| "learning_rate": 1.9526616402331733e-05, |
| "loss": 0.0149, |
| "step": 12970 |
| }, |
| { |
| "epoch": 1.3296455644335177, |
| "grad_norm": 0.42510831356048584, |
| "learning_rate": 1.9510849815501638e-05, |
| "loss": 0.0124, |
| "step": 12980 |
| }, |
| { |
| "epoch": 1.3306699446834664, |
| "grad_norm": 0.32156234979629517, |
| "learning_rate": 1.949507774756368e-05, |
| "loss": 0.0146, |
| "step": 12990 |
| }, |
| { |
| "epoch": 1.3316943249334152, |
| "grad_norm": 0.6286745071411133, |
| "learning_rate": 1.9479300217682394e-05, |
| "loss": 0.0119, |
| "step": 13000 |
| }, |
| { |
| "epoch": 1.332718705183364, |
| "grad_norm": 0.3272535800933838, |
| "learning_rate": 1.946351724502898e-05, |
| "loss": 0.0161, |
| "step": 13010 |
| }, |
| { |
| "epoch": 1.3337430854333128, |
| "grad_norm": 0.9224309921264648, |
| "learning_rate": 1.9447728848781227e-05, |
| "loss": 0.0131, |
| "step": 13020 |
| }, |
| { |
| "epoch": 1.3347674656832615, |
| "grad_norm": 0.3104749023914337, |
| "learning_rate": 1.943193504812353e-05, |
| "loss": 0.0168, |
| "step": 13030 |
| }, |
| { |
| "epoch": 1.3357918459332105, |
| "grad_norm": 0.23240035772323608, |
| "learning_rate": 1.9416135862246846e-05, |
| "loss": 0.0142, |
| "step": 13040 |
| }, |
| { |
| "epoch": 1.336816226183159, |
| "grad_norm": 0.21963346004486084, |
| "learning_rate": 1.9400331310348673e-05, |
| "loss": 0.0154, |
| "step": 13050 |
| }, |
| { |
| "epoch": 1.337840606433108, |
| "grad_norm": 0.13591913878917694, |
| "learning_rate": 1.9384521411633033e-05, |
| "loss": 0.0135, |
| "step": 13060 |
| }, |
| { |
| "epoch": 1.3388649866830566, |
| "grad_norm": 0.25451046228408813, |
| "learning_rate": 1.936870618531045e-05, |
| "loss": 0.0116, |
| "step": 13070 |
| }, |
| { |
| "epoch": 1.3398893669330056, |
| "grad_norm": 0.1676362007856369, |
| "learning_rate": 1.9352885650597904e-05, |
| "loss": 0.0144, |
| "step": 13080 |
| }, |
| { |
| "epoch": 1.3409137471829542, |
| "grad_norm": 0.14009907841682434, |
| "learning_rate": 1.9337059826718847e-05, |
| "loss": 0.0147, |
| "step": 13090 |
| }, |
| { |
| "epoch": 1.3419381274329032, |
| "grad_norm": 0.2648715078830719, |
| "learning_rate": 1.932122873290314e-05, |
| "loss": 0.0117, |
| "step": 13100 |
| }, |
| { |
| "epoch": 1.3429625076828517, |
| "grad_norm": 0.3067002594470978, |
| "learning_rate": 1.9305392388387054e-05, |
| "loss": 0.0108, |
| "step": 13110 |
| }, |
| { |
| "epoch": 1.3439868879328007, |
| "grad_norm": 0.4735851287841797, |
| "learning_rate": 1.928955081241325e-05, |
| "loss": 0.0172, |
| "step": 13120 |
| }, |
| { |
| "epoch": 1.3450112681827495, |
| "grad_norm": 0.26623180508613586, |
| "learning_rate": 1.9273704024230725e-05, |
| "loss": 0.0151, |
| "step": 13130 |
| }, |
| { |
| "epoch": 1.3460356484326983, |
| "grad_norm": 0.2680647373199463, |
| "learning_rate": 1.925785204309483e-05, |
| "loss": 0.0144, |
| "step": 13140 |
| }, |
| { |
| "epoch": 1.347060028682647, |
| "grad_norm": 0.16013659536838531, |
| "learning_rate": 1.9241994888267212e-05, |
| "loss": 0.0118, |
| "step": 13150 |
| }, |
| { |
| "epoch": 1.3480844089325958, |
| "grad_norm": 0.1385974884033203, |
| "learning_rate": 1.9226132579015815e-05, |
| "loss": 0.0145, |
| "step": 13160 |
| }, |
| { |
| "epoch": 1.3491087891825446, |
| "grad_norm": 0.3810504972934723, |
| "learning_rate": 1.9210265134614832e-05, |
| "loss": 0.0159, |
| "step": 13170 |
| }, |
| { |
| "epoch": 1.3501331694324934, |
| "grad_norm": 0.3630194067955017, |
| "learning_rate": 1.9194392574344712e-05, |
| "loss": 0.013, |
| "step": 13180 |
| }, |
| { |
| "epoch": 1.3511575496824422, |
| "grad_norm": 0.20464670658111572, |
| "learning_rate": 1.9178514917492103e-05, |
| "loss": 0.0126, |
| "step": 13190 |
| }, |
| { |
| "epoch": 1.352181929932391, |
| "grad_norm": 0.2362581193447113, |
| "learning_rate": 1.9162632183349866e-05, |
| "loss": 0.0124, |
| "step": 13200 |
| }, |
| { |
| "epoch": 1.3532063101823397, |
| "grad_norm": 0.10466092824935913, |
| "learning_rate": 1.914674439121701e-05, |
| "loss": 0.0103, |
| "step": 13210 |
| }, |
| { |
| "epoch": 1.3542306904322885, |
| "grad_norm": 0.341568261384964, |
| "learning_rate": 1.9130851560398715e-05, |
| "loss": 0.0166, |
| "step": 13220 |
| }, |
| { |
| "epoch": 1.3552550706822373, |
| "grad_norm": 0.8375012278556824, |
| "learning_rate": 1.911495371020625e-05, |
| "loss": 0.0121, |
| "step": 13230 |
| }, |
| { |
| "epoch": 1.356279450932186, |
| "grad_norm": 0.8242742419242859, |
| "learning_rate": 1.909905085995702e-05, |
| "loss": 0.0113, |
| "step": 13240 |
| }, |
| { |
| "epoch": 1.3573038311821348, |
| "grad_norm": 0.2532491981983185, |
| "learning_rate": 1.908314302897448e-05, |
| "loss": 0.0143, |
| "step": 13250 |
| }, |
| { |
| "epoch": 1.3583282114320836, |
| "grad_norm": 0.4292425513267517, |
| "learning_rate": 1.9067230236588142e-05, |
| "loss": 0.0149, |
| "step": 13260 |
| }, |
| { |
| "epoch": 1.3593525916820324, |
| "grad_norm": 0.1518702358007431, |
| "learning_rate": 1.905131250213355e-05, |
| "loss": 0.0119, |
| "step": 13270 |
| }, |
| { |
| "epoch": 1.3603769719319811, |
| "grad_norm": 0.14751629531383514, |
| "learning_rate": 1.9035389844952256e-05, |
| "loss": 0.0119, |
| "step": 13280 |
| }, |
| { |
| "epoch": 1.36140135218193, |
| "grad_norm": 0.16413085162639618, |
| "learning_rate": 1.901946228439179e-05, |
| "loss": 0.0104, |
| "step": 13290 |
| }, |
| { |
| "epoch": 1.3624257324318787, |
| "grad_norm": 0.18580059707164764, |
| "learning_rate": 1.9003529839805637e-05, |
| "loss": 0.0127, |
| "step": 13300 |
| }, |
| { |
| "epoch": 1.3634501126818275, |
| "grad_norm": 0.15212629735469818, |
| "learning_rate": 1.8987592530553217e-05, |
| "loss": 0.0127, |
| "step": 13310 |
| }, |
| { |
| "epoch": 1.3644744929317763, |
| "grad_norm": 0.391001433134079, |
| "learning_rate": 1.897165037599987e-05, |
| "loss": 0.0138, |
| "step": 13320 |
| }, |
| { |
| "epoch": 1.365498873181725, |
| "grad_norm": 0.5556850433349609, |
| "learning_rate": 1.8955703395516814e-05, |
| "loss": 0.0129, |
| "step": 13330 |
| }, |
| { |
| "epoch": 1.3665232534316738, |
| "grad_norm": 0.41252401471138, |
| "learning_rate": 1.8939751608481134e-05, |
| "loss": 0.0155, |
| "step": 13340 |
| }, |
| { |
| "epoch": 1.3675476336816226, |
| "grad_norm": 0.262264609336853, |
| "learning_rate": 1.892379503427575e-05, |
| "loss": 0.0118, |
| "step": 13350 |
| }, |
| { |
| "epoch": 1.3685720139315714, |
| "grad_norm": 0.2740829885005951, |
| "learning_rate": 1.8907833692289412e-05, |
| "loss": 0.0135, |
| "step": 13360 |
| }, |
| { |
| "epoch": 1.3695963941815201, |
| "grad_norm": 0.6438356637954712, |
| "learning_rate": 1.889186760191666e-05, |
| "loss": 0.0127, |
| "step": 13370 |
| }, |
| { |
| "epoch": 1.370620774431469, |
| "grad_norm": 0.1788123995065689, |
| "learning_rate": 1.887589678255779e-05, |
| "loss": 0.0118, |
| "step": 13380 |
| }, |
| { |
| "epoch": 1.3716451546814177, |
| "grad_norm": 0.4639585018157959, |
| "learning_rate": 1.8859921253618852e-05, |
| "loss": 0.0128, |
| "step": 13390 |
| }, |
| { |
| "epoch": 1.3726695349313665, |
| "grad_norm": 0.20991815626621246, |
| "learning_rate": 1.884394103451163e-05, |
| "loss": 0.0158, |
| "step": 13400 |
| }, |
| { |
| "epoch": 1.3736939151813152, |
| "grad_norm": 0.1605006605386734, |
| "learning_rate": 1.8827956144653583e-05, |
| "loss": 0.0139, |
| "step": 13410 |
| }, |
| { |
| "epoch": 1.374718295431264, |
| "grad_norm": 0.844539999961853, |
| "learning_rate": 1.8811966603467874e-05, |
| "loss": 0.0149, |
| "step": 13420 |
| }, |
| { |
| "epoch": 1.375742675681213, |
| "grad_norm": 0.32398971915245056, |
| "learning_rate": 1.8795972430383293e-05, |
| "loss": 0.0133, |
| "step": 13430 |
| }, |
| { |
| "epoch": 1.3767670559311616, |
| "grad_norm": 0.21985100209712982, |
| "learning_rate": 1.8779973644834273e-05, |
| "loss": 0.0138, |
| "step": 13440 |
| }, |
| { |
| "epoch": 1.3777914361811106, |
| "grad_norm": 1.0425817966461182, |
| "learning_rate": 1.876397026626085e-05, |
| "loss": 0.0122, |
| "step": 13450 |
| }, |
| { |
| "epoch": 1.378815816431059, |
| "grad_norm": 0.45722296833992004, |
| "learning_rate": 1.8747962314108633e-05, |
| "loss": 0.0126, |
| "step": 13460 |
| }, |
| { |
| "epoch": 1.379840196681008, |
| "grad_norm": 0.256424218416214, |
| "learning_rate": 1.873194980782879e-05, |
| "loss": 0.0148, |
| "step": 13470 |
| }, |
| { |
| "epoch": 1.3808645769309567, |
| "grad_norm": 0.17218205332756042, |
| "learning_rate": 1.871593276687804e-05, |
| "loss": 0.0138, |
| "step": 13480 |
| }, |
| { |
| "epoch": 1.3818889571809057, |
| "grad_norm": 0.7575616836547852, |
| "learning_rate": 1.869991121071858e-05, |
| "loss": 0.0139, |
| "step": 13490 |
| }, |
| { |
| "epoch": 1.3829133374308542, |
| "grad_norm": 0.17170336842536926, |
| "learning_rate": 1.8683885158818117e-05, |
| "loss": 0.0135, |
| "step": 13500 |
| }, |
| { |
| "epoch": 1.3839377176808032, |
| "grad_norm": 0.2610829472541809, |
| "learning_rate": 1.8667854630649825e-05, |
| "loss": 0.0142, |
| "step": 13510 |
| }, |
| { |
| "epoch": 1.3849620979307518, |
| "grad_norm": 0.17902614176273346, |
| "learning_rate": 1.86518196456923e-05, |
| "loss": 0.0133, |
| "step": 13520 |
| }, |
| { |
| "epoch": 1.3859864781807008, |
| "grad_norm": 0.2590945363044739, |
| "learning_rate": 1.8635780223429565e-05, |
| "loss": 0.0135, |
| "step": 13530 |
| }, |
| { |
| "epoch": 1.3870108584306495, |
| "grad_norm": 0.3471442759037018, |
| "learning_rate": 1.8619736383351017e-05, |
| "loss": 0.0144, |
| "step": 13540 |
| }, |
| { |
| "epoch": 1.3880352386805983, |
| "grad_norm": 0.2686975300312042, |
| "learning_rate": 1.8603688144951454e-05, |
| "loss": 0.0115, |
| "step": 13550 |
| }, |
| { |
| "epoch": 1.389059618930547, |
| "grad_norm": 0.21529005467891693, |
| "learning_rate": 1.8587635527730983e-05, |
| "loss": 0.0155, |
| "step": 13560 |
| }, |
| { |
| "epoch": 1.3900839991804959, |
| "grad_norm": 0.4831743538379669, |
| "learning_rate": 1.8571578551195056e-05, |
| "loss": 0.0163, |
| "step": 13570 |
| }, |
| { |
| "epoch": 1.3911083794304446, |
| "grad_norm": 0.25540825724601746, |
| "learning_rate": 1.855551723485441e-05, |
| "loss": 0.0155, |
| "step": 13580 |
| }, |
| { |
| "epoch": 1.3921327596803934, |
| "grad_norm": 0.3135828971862793, |
| "learning_rate": 1.8539451598225066e-05, |
| "loss": 0.0111, |
| "step": 13590 |
| }, |
| { |
| "epoch": 1.3931571399303422, |
| "grad_norm": 0.44960108399391174, |
| "learning_rate": 1.8523381660828276e-05, |
| "loss": 0.0137, |
| "step": 13600 |
| }, |
| { |
| "epoch": 1.394181520180291, |
| "grad_norm": 0.11005840450525284, |
| "learning_rate": 1.850730744219054e-05, |
| "loss": 0.0162, |
| "step": 13610 |
| }, |
| { |
| "epoch": 1.3952059004302397, |
| "grad_norm": 0.32107317447662354, |
| "learning_rate": 1.849122896184354e-05, |
| "loss": 0.0141, |
| "step": 13620 |
| }, |
| { |
| "epoch": 1.3962302806801885, |
| "grad_norm": 0.38131994009017944, |
| "learning_rate": 1.8475146239324146e-05, |
| "loss": 0.0135, |
| "step": 13630 |
| }, |
| { |
| "epoch": 1.3972546609301373, |
| "grad_norm": 0.2464354783296585, |
| "learning_rate": 1.8459059294174396e-05, |
| "loss": 0.0155, |
| "step": 13640 |
| }, |
| { |
| "epoch": 1.398279041180086, |
| "grad_norm": 0.13476291298866272, |
| "learning_rate": 1.844296814594143e-05, |
| "loss": 0.0135, |
| "step": 13650 |
| }, |
| { |
| "epoch": 1.3993034214300348, |
| "grad_norm": 0.21349047124385834, |
| "learning_rate": 1.8426872814177516e-05, |
| "loss": 0.0124, |
| "step": 13660 |
| }, |
| { |
| "epoch": 1.4003278016799836, |
| "grad_norm": 0.4448750913143158, |
| "learning_rate": 1.8410773318440003e-05, |
| "loss": 0.0146, |
| "step": 13670 |
| }, |
| { |
| "epoch": 1.4013521819299324, |
| "grad_norm": 0.21708209812641144, |
| "learning_rate": 1.839466967829129e-05, |
| "loss": 0.0117, |
| "step": 13680 |
| }, |
| { |
| "epoch": 1.4023765621798812, |
| "grad_norm": 0.17444850504398346, |
| "learning_rate": 1.8378561913298822e-05, |
| "loss": 0.0126, |
| "step": 13690 |
| }, |
| { |
| "epoch": 1.40340094242983, |
| "grad_norm": 0.4328935742378235, |
| "learning_rate": 1.8362450043035053e-05, |
| "loss": 0.0139, |
| "step": 13700 |
| }, |
| { |
| "epoch": 1.4044253226797787, |
| "grad_norm": 0.3198728561401367, |
| "learning_rate": 1.834633408707742e-05, |
| "loss": 0.0142, |
| "step": 13710 |
| }, |
| { |
| "epoch": 1.4054497029297275, |
| "grad_norm": 0.6568001508712769, |
| "learning_rate": 1.8330214065008336e-05, |
| "loss": 0.0139, |
| "step": 13720 |
| }, |
| { |
| "epoch": 1.4064740831796763, |
| "grad_norm": 0.22537779808044434, |
| "learning_rate": 1.831408999641514e-05, |
| "loss": 0.0124, |
| "step": 13730 |
| }, |
| { |
| "epoch": 1.407498463429625, |
| "grad_norm": 0.2351643443107605, |
| "learning_rate": 1.8297961900890107e-05, |
| "loss": 0.0135, |
| "step": 13740 |
| }, |
| { |
| "epoch": 1.4085228436795738, |
| "grad_norm": 0.6375037431716919, |
| "learning_rate": 1.8281829798030385e-05, |
| "loss": 0.0147, |
| "step": 13750 |
| }, |
| { |
| "epoch": 1.4095472239295226, |
| "grad_norm": 0.4508446753025055, |
| "learning_rate": 1.8265693707438002e-05, |
| "loss": 0.0111, |
| "step": 13760 |
| }, |
| { |
| "epoch": 1.4105716041794714, |
| "grad_norm": 0.5748338103294373, |
| "learning_rate": 1.824955364871982e-05, |
| "loss": 0.0112, |
| "step": 13770 |
| }, |
| { |
| "epoch": 1.4115959844294201, |
| "grad_norm": 0.43466609716415405, |
| "learning_rate": 1.8233409641487553e-05, |
| "loss": 0.0146, |
| "step": 13780 |
| }, |
| { |
| "epoch": 1.412620364679369, |
| "grad_norm": 0.4040360450744629, |
| "learning_rate": 1.8217261705357677e-05, |
| "loss": 0.0145, |
| "step": 13790 |
| }, |
| { |
| "epoch": 1.4136447449293177, |
| "grad_norm": 0.12479580193758011, |
| "learning_rate": 1.8201109859951462e-05, |
| "loss": 0.0109, |
| "step": 13800 |
| }, |
| { |
| "epoch": 1.4146691251792665, |
| "grad_norm": 0.10730087012052536, |
| "learning_rate": 1.818495412489492e-05, |
| "loss": 0.0143, |
| "step": 13810 |
| }, |
| { |
| "epoch": 1.4156935054292155, |
| "grad_norm": 0.3391306400299072, |
| "learning_rate": 1.8168794519818802e-05, |
| "loss": 0.0144, |
| "step": 13820 |
| }, |
| { |
| "epoch": 1.416717885679164, |
| "grad_norm": 0.24596834182739258, |
| "learning_rate": 1.8152631064358537e-05, |
| "loss": 0.0112, |
| "step": 13830 |
| }, |
| { |
| "epoch": 1.417742265929113, |
| "grad_norm": 0.5312104821205139, |
| "learning_rate": 1.8136463778154258e-05, |
| "loss": 0.0139, |
| "step": 13840 |
| }, |
| { |
| "epoch": 1.4187666461790616, |
| "grad_norm": 0.2699443995952606, |
| "learning_rate": 1.812029268085074e-05, |
| "loss": 0.0124, |
| "step": 13850 |
| }, |
| { |
| "epoch": 1.4197910264290106, |
| "grad_norm": 0.25075387954711914, |
| "learning_rate": 1.8104117792097387e-05, |
| "loss": 0.0138, |
| "step": 13860 |
| }, |
| { |
| "epoch": 1.4208154066789591, |
| "grad_norm": 0.4274173974990845, |
| "learning_rate": 1.808793913154822e-05, |
| "loss": 0.0134, |
| "step": 13870 |
| }, |
| { |
| "epoch": 1.4218397869289081, |
| "grad_norm": 0.20575514435768127, |
| "learning_rate": 1.8071756718861838e-05, |
| "loss": 0.0143, |
| "step": 13880 |
| }, |
| { |
| "epoch": 1.4228641671788567, |
| "grad_norm": 0.3132922649383545, |
| "learning_rate": 1.805557057370139e-05, |
| "loss": 0.0129, |
| "step": 13890 |
| }, |
| { |
| "epoch": 1.4238885474288057, |
| "grad_norm": 0.7305237650871277, |
| "learning_rate": 1.8039380715734574e-05, |
| "loss": 0.0135, |
| "step": 13900 |
| }, |
| { |
| "epoch": 1.4249129276787542, |
| "grad_norm": 0.10184314101934433, |
| "learning_rate": 1.802318716463359e-05, |
| "loss": 0.0119, |
| "step": 13910 |
| }, |
| { |
| "epoch": 1.4259373079287032, |
| "grad_norm": 0.27214860916137695, |
| "learning_rate": 1.8006989940075138e-05, |
| "loss": 0.0141, |
| "step": 13920 |
| }, |
| { |
| "epoch": 1.426961688178652, |
| "grad_norm": 0.5004453659057617, |
| "learning_rate": 1.7990789061740368e-05, |
| "loss": 0.0144, |
| "step": 13930 |
| }, |
| { |
| "epoch": 1.4279860684286008, |
| "grad_norm": 0.17023734748363495, |
| "learning_rate": 1.7974584549314875e-05, |
| "loss": 0.0118, |
| "step": 13940 |
| }, |
| { |
| "epoch": 1.4290104486785495, |
| "grad_norm": 0.11566552519798279, |
| "learning_rate": 1.7958376422488672e-05, |
| "loss": 0.0136, |
| "step": 13950 |
| }, |
| { |
| "epoch": 1.4300348289284983, |
| "grad_norm": 0.29088500142097473, |
| "learning_rate": 1.7942164700956156e-05, |
| "loss": 0.0149, |
| "step": 13960 |
| }, |
| { |
| "epoch": 1.431059209178447, |
| "grad_norm": 0.475600004196167, |
| "learning_rate": 1.79259494044161e-05, |
| "loss": 0.0128, |
| "step": 13970 |
| }, |
| { |
| "epoch": 1.4320835894283959, |
| "grad_norm": 0.6676848530769348, |
| "learning_rate": 1.7909730552571617e-05, |
| "loss": 0.0143, |
| "step": 13980 |
| }, |
| { |
| "epoch": 1.4331079696783446, |
| "grad_norm": 0.17428959906101227, |
| "learning_rate": 1.7893508165130143e-05, |
| "loss": 0.013, |
| "step": 13990 |
| }, |
| { |
| "epoch": 1.4341323499282934, |
| "grad_norm": 0.757672905921936, |
| "learning_rate": 1.787728226180341e-05, |
| "loss": 0.0129, |
| "step": 14000 |
| }, |
| { |
| "epoch": 1.4351567301782422, |
| "grad_norm": 0.280303031206131, |
| "learning_rate": 1.786105286230742e-05, |
| "loss": 0.0124, |
| "step": 14010 |
| }, |
| { |
| "epoch": 1.436181110428191, |
| "grad_norm": 0.4819464683532715, |
| "learning_rate": 1.784481998636242e-05, |
| "loss": 0.0119, |
| "step": 14020 |
| }, |
| { |
| "epoch": 1.4372054906781397, |
| "grad_norm": 1.2588610649108887, |
| "learning_rate": 1.782858365369289e-05, |
| "loss": 0.0123, |
| "step": 14030 |
| }, |
| { |
| "epoch": 1.4382298709280885, |
| "grad_norm": 0.17123635113239288, |
| "learning_rate": 1.78123438840275e-05, |
| "loss": 0.0118, |
| "step": 14040 |
| }, |
| { |
| "epoch": 1.4392542511780373, |
| "grad_norm": 0.17428705096244812, |
| "learning_rate": 1.779610069709911e-05, |
| "loss": 0.0135, |
| "step": 14050 |
| }, |
| { |
| "epoch": 1.440278631427986, |
| "grad_norm": 0.20843759179115295, |
| "learning_rate": 1.7779854112644713e-05, |
| "loss": 0.0132, |
| "step": 14060 |
| }, |
| { |
| "epoch": 1.4413030116779348, |
| "grad_norm": 0.169826477766037, |
| "learning_rate": 1.7763604150405446e-05, |
| "loss": 0.0148, |
| "step": 14070 |
| }, |
| { |
| "epoch": 1.4423273919278836, |
| "grad_norm": 0.2196032553911209, |
| "learning_rate": 1.7747350830126548e-05, |
| "loss": 0.0099, |
| "step": 14080 |
| }, |
| { |
| "epoch": 1.4433517721778324, |
| "grad_norm": 0.8376290798187256, |
| "learning_rate": 1.7731094171557332e-05, |
| "loss": 0.0132, |
| "step": 14090 |
| }, |
| { |
| "epoch": 1.4443761524277812, |
| "grad_norm": 0.1784752756357193, |
| "learning_rate": 1.771483419445117e-05, |
| "loss": 0.0127, |
| "step": 14100 |
| }, |
| { |
| "epoch": 1.44540053267773, |
| "grad_norm": 0.7512127757072449, |
| "learning_rate": 1.7698570918565474e-05, |
| "loss": 0.0138, |
| "step": 14110 |
| }, |
| { |
| "epoch": 1.4464249129276787, |
| "grad_norm": 0.15940797328948975, |
| "learning_rate": 1.768230436366165e-05, |
| "loss": 0.0109, |
| "step": 14120 |
| }, |
| { |
| "epoch": 1.4474492931776275, |
| "grad_norm": 0.38637787103652954, |
| "learning_rate": 1.7666034549505102e-05, |
| "loss": 0.0126, |
| "step": 14130 |
| }, |
| { |
| "epoch": 1.4484736734275763, |
| "grad_norm": 0.5113162398338318, |
| "learning_rate": 1.7649761495865185e-05, |
| "loss": 0.0114, |
| "step": 14140 |
| }, |
| { |
| "epoch": 1.449498053677525, |
| "grad_norm": 0.2935347259044647, |
| "learning_rate": 1.7633485222515188e-05, |
| "loss": 0.0136, |
| "step": 14150 |
| }, |
| { |
| "epoch": 1.4505224339274738, |
| "grad_norm": 0.2337443083524704, |
| "learning_rate": 1.761720574923233e-05, |
| "loss": 0.0146, |
| "step": 14160 |
| }, |
| { |
| "epoch": 1.4515468141774226, |
| "grad_norm": 0.5570079684257507, |
| "learning_rate": 1.7600923095797704e-05, |
| "loss": 0.0115, |
| "step": 14170 |
| }, |
| { |
| "epoch": 1.4525711944273714, |
| "grad_norm": 0.41292503476142883, |
| "learning_rate": 1.758463728199626e-05, |
| "loss": 0.0137, |
| "step": 14180 |
| }, |
| { |
| "epoch": 1.4535955746773201, |
| "grad_norm": 0.36874985694885254, |
| "learning_rate": 1.7568348327616803e-05, |
| "loss": 0.0153, |
| "step": 14190 |
| }, |
| { |
| "epoch": 1.454619954927269, |
| "grad_norm": 0.19937613606452942, |
| "learning_rate": 1.755205625245195e-05, |
| "loss": 0.0119, |
| "step": 14200 |
| }, |
| { |
| "epoch": 1.455644335177218, |
| "grad_norm": 0.9493114948272705, |
| "learning_rate": 1.7535761076298104e-05, |
| "loss": 0.0163, |
| "step": 14210 |
| }, |
| { |
| "epoch": 1.4566687154271665, |
| "grad_norm": 0.20606830716133118, |
| "learning_rate": 1.7519462818955448e-05, |
| "loss": 0.0138, |
| "step": 14220 |
| }, |
| { |
| "epoch": 1.4576930956771155, |
| "grad_norm": 0.18265338242053986, |
| "learning_rate": 1.7503161500227896e-05, |
| "loss": 0.0141, |
| "step": 14230 |
| }, |
| { |
| "epoch": 1.458717475927064, |
| "grad_norm": 0.24758809804916382, |
| "learning_rate": 1.7486857139923093e-05, |
| "loss": 0.0125, |
| "step": 14240 |
| }, |
| { |
| "epoch": 1.459741856177013, |
| "grad_norm": 0.5712595582008362, |
| "learning_rate": 1.747054975785236e-05, |
| "loss": 0.0169, |
| "step": 14250 |
| }, |
| { |
| "epoch": 1.4607662364269616, |
| "grad_norm": 0.5265136957168579, |
| "learning_rate": 1.745423937383072e-05, |
| "loss": 0.0124, |
| "step": 14260 |
| }, |
| { |
| "epoch": 1.4617906166769106, |
| "grad_norm": 0.5060036182403564, |
| "learning_rate": 1.7437926007676825e-05, |
| "loss": 0.0135, |
| "step": 14270 |
| }, |
| { |
| "epoch": 1.4628149969268591, |
| "grad_norm": 0.34388792514801025, |
| "learning_rate": 1.742160967921294e-05, |
| "loss": 0.0111, |
| "step": 14280 |
| }, |
| { |
| "epoch": 1.4638393771768081, |
| "grad_norm": 0.26012033224105835, |
| "learning_rate": 1.7405290408264954e-05, |
| "loss": 0.0135, |
| "step": 14290 |
| }, |
| { |
| "epoch": 1.4648637574267567, |
| "grad_norm": 0.21177008748054504, |
| "learning_rate": 1.738896821466232e-05, |
| "loss": 0.0138, |
| "step": 14300 |
| }, |
| { |
| "epoch": 1.4658881376767057, |
| "grad_norm": 0.19501881301403046, |
| "learning_rate": 1.737264311823804e-05, |
| "loss": 0.0128, |
| "step": 14310 |
| }, |
| { |
| "epoch": 1.4669125179266544, |
| "grad_norm": 0.1536514312028885, |
| "learning_rate": 1.7356315138828645e-05, |
| "loss": 0.0153, |
| "step": 14320 |
| }, |
| { |
| "epoch": 1.4679368981766032, |
| "grad_norm": 0.5302589535713196, |
| "learning_rate": 1.7339984296274174e-05, |
| "loss": 0.0159, |
| "step": 14330 |
| }, |
| { |
| "epoch": 1.468961278426552, |
| "grad_norm": 0.4712163209915161, |
| "learning_rate": 1.7323650610418134e-05, |
| "loss": 0.0139, |
| "step": 14340 |
| }, |
| { |
| "epoch": 1.4699856586765008, |
| "grad_norm": 0.1864139437675476, |
| "learning_rate": 1.730731410110749e-05, |
| "loss": 0.0126, |
| "step": 14350 |
| }, |
| { |
| "epoch": 1.4710100389264495, |
| "grad_norm": 0.6358790397644043, |
| "learning_rate": 1.7290974788192655e-05, |
| "loss": 0.0121, |
| "step": 14360 |
| }, |
| { |
| "epoch": 1.4720344191763983, |
| "grad_norm": 0.2657432556152344, |
| "learning_rate": 1.727463269152743e-05, |
| "loss": 0.0166, |
| "step": 14370 |
| }, |
| { |
| "epoch": 1.473058799426347, |
| "grad_norm": 0.13608287274837494, |
| "learning_rate": 1.7258287830968996e-05, |
| "loss": 0.0131, |
| "step": 14380 |
| }, |
| { |
| "epoch": 1.4740831796762959, |
| "grad_norm": 0.4240320920944214, |
| "learning_rate": 1.7241940226377908e-05, |
| "loss": 0.0126, |
| "step": 14390 |
| }, |
| { |
| "epoch": 1.4751075599262446, |
| "grad_norm": 0.3134903311729431, |
| "learning_rate": 1.7225589897618043e-05, |
| "loss": 0.0134, |
| "step": 14400 |
| }, |
| { |
| "epoch": 1.4761319401761934, |
| "grad_norm": 0.5216298699378967, |
| "learning_rate": 1.7209236864556588e-05, |
| "loss": 0.0147, |
| "step": 14410 |
| }, |
| { |
| "epoch": 1.4771563204261422, |
| "grad_norm": 0.3353475034236908, |
| "learning_rate": 1.719288114706403e-05, |
| "loss": 0.0111, |
| "step": 14420 |
| }, |
| { |
| "epoch": 1.478180700676091, |
| "grad_norm": 0.40981176495552063, |
| "learning_rate": 1.7176522765014104e-05, |
| "loss": 0.0123, |
| "step": 14430 |
| }, |
| { |
| "epoch": 1.4792050809260398, |
| "grad_norm": 0.42023205757141113, |
| "learning_rate": 1.7160161738283782e-05, |
| "loss": 0.0135, |
| "step": 14440 |
| }, |
| { |
| "epoch": 1.4802294611759885, |
| "grad_norm": 0.22594283521175385, |
| "learning_rate": 1.714379808675326e-05, |
| "loss": 0.015, |
| "step": 14450 |
| }, |
| { |
| "epoch": 1.4812538414259373, |
| "grad_norm": 0.19290079176425934, |
| "learning_rate": 1.712743183030592e-05, |
| "loss": 0.0137, |
| "step": 14460 |
| }, |
| { |
| "epoch": 1.482278221675886, |
| "grad_norm": 0.691516101360321, |
| "learning_rate": 1.7111062988828294e-05, |
| "loss": 0.0144, |
| "step": 14470 |
| }, |
| { |
| "epoch": 1.4833026019258349, |
| "grad_norm": 0.16169840097427368, |
| "learning_rate": 1.709469158221008e-05, |
| "loss": 0.0145, |
| "step": 14480 |
| }, |
| { |
| "epoch": 1.4843269821757836, |
| "grad_norm": 0.20352774858474731, |
| "learning_rate": 1.7078317630344083e-05, |
| "loss": 0.0129, |
| "step": 14490 |
| }, |
| { |
| "epoch": 1.4853513624257324, |
| "grad_norm": 0.517322301864624, |
| "learning_rate": 1.706194115312618e-05, |
| "loss": 0.0166, |
| "step": 14500 |
| }, |
| { |
| "epoch": 1.4863757426756812, |
| "grad_norm": 0.20517213642597198, |
| "learning_rate": 1.704556217045536e-05, |
| "loss": 0.0111, |
| "step": 14510 |
| }, |
| { |
| "epoch": 1.48740012292563, |
| "grad_norm": 0.5912905335426331, |
| "learning_rate": 1.7029180702233612e-05, |
| "loss": 0.0115, |
| "step": 14520 |
| }, |
| { |
| "epoch": 1.4884245031755787, |
| "grad_norm": 0.9042195677757263, |
| "learning_rate": 1.7012796768365975e-05, |
| "loss": 0.0132, |
| "step": 14530 |
| }, |
| { |
| "epoch": 1.4894488834255275, |
| "grad_norm": 0.4479576647281647, |
| "learning_rate": 1.6996410388760465e-05, |
| "loss": 0.0133, |
| "step": 14540 |
| }, |
| { |
| "epoch": 1.4904732636754763, |
| "grad_norm": 0.44661545753479004, |
| "learning_rate": 1.698002158332809e-05, |
| "loss": 0.0124, |
| "step": 14550 |
| }, |
| { |
| "epoch": 1.491497643925425, |
| "grad_norm": 0.30754074454307556, |
| "learning_rate": 1.6963630371982783e-05, |
| "loss": 0.0157, |
| "step": 14560 |
| }, |
| { |
| "epoch": 1.4925220241753738, |
| "grad_norm": 0.2589346766471863, |
| "learning_rate": 1.6947236774641416e-05, |
| "loss": 0.0147, |
| "step": 14570 |
| }, |
| { |
| "epoch": 1.4935464044253226, |
| "grad_norm": 0.1713082492351532, |
| "learning_rate": 1.6930840811223755e-05, |
| "loss": 0.0124, |
| "step": 14580 |
| }, |
| { |
| "epoch": 1.4945707846752714, |
| "grad_norm": 0.3472796678543091, |
| "learning_rate": 1.6914442501652448e-05, |
| "loss": 0.0157, |
| "step": 14590 |
| }, |
| { |
| "epoch": 1.4955951649252204, |
| "grad_norm": 0.4575025141239166, |
| "learning_rate": 1.6898041865852974e-05, |
| "loss": 0.0143, |
| "step": 14600 |
| }, |
| { |
| "epoch": 1.496619545175169, |
| "grad_norm": 0.2858721613883972, |
| "learning_rate": 1.6881638923753663e-05, |
| "loss": 0.0145, |
| "step": 14610 |
| }, |
| { |
| "epoch": 1.497643925425118, |
| "grad_norm": 1.7555986642837524, |
| "learning_rate": 1.6865233695285632e-05, |
| "loss": 0.0142, |
| "step": 14620 |
| }, |
| { |
| "epoch": 1.4986683056750665, |
| "grad_norm": 0.2419038712978363, |
| "learning_rate": 1.6848826200382782e-05, |
| "loss": 0.0137, |
| "step": 14630 |
| }, |
| { |
| "epoch": 1.4996926859250155, |
| "grad_norm": 0.41063398122787476, |
| "learning_rate": 1.6832416458981762e-05, |
| "loss": 0.0127, |
| "step": 14640 |
| }, |
| { |
| "epoch": 1.500717066174964, |
| "grad_norm": 0.2533039152622223, |
| "learning_rate": 1.681600449102196e-05, |
| "loss": 0.0147, |
| "step": 14650 |
| }, |
| { |
| "epoch": 1.501741446424913, |
| "grad_norm": 0.26742658019065857, |
| "learning_rate": 1.6799590316445468e-05, |
| "loss": 0.0138, |
| "step": 14660 |
| }, |
| { |
| "epoch": 1.5027658266748616, |
| "grad_norm": 0.2679586112499237, |
| "learning_rate": 1.678317395519705e-05, |
| "loss": 0.0142, |
| "step": 14670 |
| }, |
| { |
| "epoch": 1.5037902069248106, |
| "grad_norm": 0.32000815868377686, |
| "learning_rate": 1.676675542722413e-05, |
| "loss": 0.013, |
| "step": 14680 |
| }, |
| { |
| "epoch": 1.5048145871747591, |
| "grad_norm": 0.5830944180488586, |
| "learning_rate": 1.6750334752476776e-05, |
| "loss": 0.011, |
| "step": 14690 |
| }, |
| { |
| "epoch": 1.5058389674247081, |
| "grad_norm": 0.3938954770565033, |
| "learning_rate": 1.673391195090765e-05, |
| "loss": 0.0125, |
| "step": 14700 |
| }, |
| { |
| "epoch": 1.5068633476746567, |
| "grad_norm": 0.1556713581085205, |
| "learning_rate": 1.6717487042472006e-05, |
| "loss": 0.012, |
| "step": 14710 |
| }, |
| { |
| "epoch": 1.5078877279246057, |
| "grad_norm": 0.2426462024450302, |
| "learning_rate": 1.6701060047127655e-05, |
| "loss": 0.0103, |
| "step": 14720 |
| }, |
| { |
| "epoch": 1.5089121081745542, |
| "grad_norm": 0.1990267038345337, |
| "learning_rate": 1.6684630984834947e-05, |
| "loss": 0.0128, |
| "step": 14730 |
| }, |
| { |
| "epoch": 1.5099364884245032, |
| "grad_norm": 0.4909380078315735, |
| "learning_rate": 1.666819987555675e-05, |
| "loss": 0.013, |
| "step": 14740 |
| }, |
| { |
| "epoch": 1.510960868674452, |
| "grad_norm": 0.7860000729560852, |
| "learning_rate": 1.6651766739258392e-05, |
| "loss": 0.0157, |
| "step": 14750 |
| }, |
| { |
| "epoch": 1.5119852489244008, |
| "grad_norm": 0.2559044361114502, |
| "learning_rate": 1.6635331595907704e-05, |
| "loss": 0.0143, |
| "step": 14760 |
| }, |
| { |
| "epoch": 1.5130096291743496, |
| "grad_norm": 0.16307435929775238, |
| "learning_rate": 1.6618894465474923e-05, |
| "loss": 0.0118, |
| "step": 14770 |
| }, |
| { |
| "epoch": 1.5140340094242983, |
| "grad_norm": 0.400304913520813, |
| "learning_rate": 1.660245536793271e-05, |
| "loss": 0.0116, |
| "step": 14780 |
| }, |
| { |
| "epoch": 1.515058389674247, |
| "grad_norm": 0.1297372430562973, |
| "learning_rate": 1.6586014323256126e-05, |
| "loss": 0.013, |
| "step": 14790 |
| }, |
| { |
| "epoch": 1.5160827699241959, |
| "grad_norm": 0.2916795015335083, |
| "learning_rate": 1.656957135142259e-05, |
| "loss": 0.0136, |
| "step": 14800 |
| }, |
| { |
| "epoch": 1.5171071501741447, |
| "grad_norm": 0.8888742923736572, |
| "learning_rate": 1.655312647241186e-05, |
| "loss": 0.0136, |
| "step": 14810 |
| }, |
| { |
| "epoch": 1.5181315304240934, |
| "grad_norm": 0.22017106413841248, |
| "learning_rate": 1.6536679706206015e-05, |
| "loss": 0.0112, |
| "step": 14820 |
| }, |
| { |
| "epoch": 1.5191559106740422, |
| "grad_norm": 0.2081090211868286, |
| "learning_rate": 1.6520231072789427e-05, |
| "loss": 0.0145, |
| "step": 14830 |
| }, |
| { |
| "epoch": 1.520180290923991, |
| "grad_norm": 0.347054660320282, |
| "learning_rate": 1.6503780592148737e-05, |
| "loss": 0.0129, |
| "step": 14840 |
| }, |
| { |
| "epoch": 1.5212046711739398, |
| "grad_norm": 0.14401641488075256, |
| "learning_rate": 1.648732828427283e-05, |
| "loss": 0.0123, |
| "step": 14850 |
| }, |
| { |
| "epoch": 1.5222290514238885, |
| "grad_norm": 0.1614667773246765, |
| "learning_rate": 1.647087416915281e-05, |
| "loss": 0.0111, |
| "step": 14860 |
| }, |
| { |
| "epoch": 1.5232534316738373, |
| "grad_norm": 0.36733609437942505, |
| "learning_rate": 1.6454418266781987e-05, |
| "loss": 0.0121, |
| "step": 14870 |
| }, |
| { |
| "epoch": 1.524277811923786, |
| "grad_norm": 0.08113733679056168, |
| "learning_rate": 1.6437960597155824e-05, |
| "loss": 0.0119, |
| "step": 14880 |
| }, |
| { |
| "epoch": 1.5253021921737349, |
| "grad_norm": 0.38818246126174927, |
| "learning_rate": 1.642150118027194e-05, |
| "loss": 0.0138, |
| "step": 14890 |
| }, |
| { |
| "epoch": 1.5263265724236836, |
| "grad_norm": 0.799624502658844, |
| "learning_rate": 1.640504003613009e-05, |
| "loss": 0.0155, |
| "step": 14900 |
| }, |
| { |
| "epoch": 1.5273509526736324, |
| "grad_norm": 0.24060015380382538, |
| "learning_rate": 1.6388577184732107e-05, |
| "loss": 0.0135, |
| "step": 14910 |
| }, |
| { |
| "epoch": 1.5283753329235812, |
| "grad_norm": 0.27020570635795593, |
| "learning_rate": 1.6372112646081907e-05, |
| "loss": 0.011, |
| "step": 14920 |
| }, |
| { |
| "epoch": 1.52939971317353, |
| "grad_norm": 0.2329522669315338, |
| "learning_rate": 1.6355646440185464e-05, |
| "loss": 0.0107, |
| "step": 14930 |
| }, |
| { |
| "epoch": 1.5304240934234787, |
| "grad_norm": 0.3512904644012451, |
| "learning_rate": 1.6339178587050764e-05, |
| "loss": 0.0113, |
| "step": 14940 |
| }, |
| { |
| "epoch": 1.5314484736734277, |
| "grad_norm": 0.8161691427230835, |
| "learning_rate": 1.632270910668781e-05, |
| "loss": 0.0149, |
| "step": 14950 |
| }, |
| { |
| "epoch": 1.5324728539233763, |
| "grad_norm": 0.23326362669467926, |
| "learning_rate": 1.6306238019108562e-05, |
| "loss": 0.0142, |
| "step": 14960 |
| }, |
| { |
| "epoch": 1.5334972341733253, |
| "grad_norm": 0.3596373200416565, |
| "learning_rate": 1.6289765344326952e-05, |
| "loss": 0.0121, |
| "step": 14970 |
| }, |
| { |
| "epoch": 1.5345216144232738, |
| "grad_norm": 0.2724621593952179, |
| "learning_rate": 1.6273291102358833e-05, |
| "loss": 0.0129, |
| "step": 14980 |
| }, |
| { |
| "epoch": 1.5355459946732228, |
| "grad_norm": 0.14456290006637573, |
| "learning_rate": 1.6256815313221963e-05, |
| "loss": 0.0149, |
| "step": 14990 |
| }, |
| { |
| "epoch": 1.5365703749231714, |
| "grad_norm": 0.26871898770332336, |
| "learning_rate": 1.6240337996935972e-05, |
| "loss": 0.0117, |
| "step": 15000 |
| }, |
| { |
| "epoch": 1.5375947551731204, |
| "grad_norm": 0.3085518777370453, |
| "learning_rate": 1.6223859173522367e-05, |
| "loss": 0.01, |
| "step": 15010 |
| }, |
| { |
| "epoch": 1.538619135423069, |
| "grad_norm": 0.48268231749534607, |
| "learning_rate": 1.6207378863004455e-05, |
| "loss": 0.012, |
| "step": 15020 |
| }, |
| { |
| "epoch": 1.539643515673018, |
| "grad_norm": 0.1946774572134018, |
| "learning_rate": 1.6190897085407382e-05, |
| "loss": 0.0116, |
| "step": 15030 |
| }, |
| { |
| "epoch": 1.5406678959229665, |
| "grad_norm": 0.1722993701696396, |
| "learning_rate": 1.617441386075805e-05, |
| "loss": 0.0126, |
| "step": 15040 |
| }, |
| { |
| "epoch": 1.5416922761729155, |
| "grad_norm": 0.15520521998405457, |
| "learning_rate": 1.6157929209085142e-05, |
| "loss": 0.0129, |
| "step": 15050 |
| }, |
| { |
| "epoch": 1.542716656422864, |
| "grad_norm": 0.1566459834575653, |
| "learning_rate": 1.614144315041906e-05, |
| "loss": 0.0146, |
| "step": 15060 |
| }, |
| { |
| "epoch": 1.543741036672813, |
| "grad_norm": 0.16129659116268158, |
| "learning_rate": 1.612495570479192e-05, |
| "loss": 0.0132, |
| "step": 15070 |
| }, |
| { |
| "epoch": 1.5447654169227616, |
| "grad_norm": 0.622917652130127, |
| "learning_rate": 1.6108466892237517e-05, |
| "loss": 0.0124, |
| "step": 15080 |
| }, |
| { |
| "epoch": 1.5457897971727106, |
| "grad_norm": 0.36943677067756653, |
| "learning_rate": 1.609197673279132e-05, |
| "loss": 0.0119, |
| "step": 15090 |
| }, |
| { |
| "epoch": 1.5468141774226591, |
| "grad_norm": 0.4039340317249298, |
| "learning_rate": 1.6075485246490428e-05, |
| "loss": 0.011, |
| "step": 15100 |
| }, |
| { |
| "epoch": 1.5478385576726081, |
| "grad_norm": 0.4576469659805298, |
| "learning_rate": 1.605899245337355e-05, |
| "loss": 0.0146, |
| "step": 15110 |
| }, |
| { |
| "epoch": 1.5488629379225567, |
| "grad_norm": 0.30469411611557007, |
| "learning_rate": 1.6042498373480983e-05, |
| "loss": 0.0119, |
| "step": 15120 |
| }, |
| { |
| "epoch": 1.5498873181725057, |
| "grad_norm": 0.2016952633857727, |
| "learning_rate": 1.602600302685459e-05, |
| "loss": 0.0125, |
| "step": 15130 |
| }, |
| { |
| "epoch": 1.5509116984224545, |
| "grad_norm": 0.23838074505329132, |
| "learning_rate": 1.6009506433537774e-05, |
| "loss": 0.0124, |
| "step": 15140 |
| }, |
| { |
| "epoch": 1.5519360786724032, |
| "grad_norm": 0.15262597799301147, |
| "learning_rate": 1.599300861357545e-05, |
| "loss": 0.014, |
| "step": 15150 |
| }, |
| { |
| "epoch": 1.552960458922352, |
| "grad_norm": 0.8459177017211914, |
| "learning_rate": 1.5976509587014032e-05, |
| "loss": 0.0129, |
| "step": 15160 |
| }, |
| { |
| "epoch": 1.5539848391723008, |
| "grad_norm": 0.45141810178756714, |
| "learning_rate": 1.5960009373901384e-05, |
| "loss": 0.0142, |
| "step": 15170 |
| }, |
| { |
| "epoch": 1.5550092194222496, |
| "grad_norm": 0.5091553330421448, |
| "learning_rate": 1.5943507994286823e-05, |
| "loss": 0.0142, |
| "step": 15180 |
| }, |
| { |
| "epoch": 1.5560335996721983, |
| "grad_norm": 0.36023563146591187, |
| "learning_rate": 1.5927005468221082e-05, |
| "loss": 0.0123, |
| "step": 15190 |
| }, |
| { |
| "epoch": 1.5570579799221471, |
| "grad_norm": 0.23594684898853302, |
| "learning_rate": 1.5910501815756285e-05, |
| "loss": 0.0136, |
| "step": 15200 |
| }, |
| { |
| "epoch": 1.558082360172096, |
| "grad_norm": 0.22232870757579803, |
| "learning_rate": 1.589399705694593e-05, |
| "loss": 0.0116, |
| "step": 15210 |
| }, |
| { |
| "epoch": 1.5591067404220447, |
| "grad_norm": 0.479950875043869, |
| "learning_rate": 1.587749121184485e-05, |
| "loss": 0.0131, |
| "step": 15220 |
| }, |
| { |
| "epoch": 1.5601311206719934, |
| "grad_norm": 0.3703489303588867, |
| "learning_rate": 1.5860984300509206e-05, |
| "loss": 0.0126, |
| "step": 15230 |
| }, |
| { |
| "epoch": 1.5611555009219422, |
| "grad_norm": 0.49232810735702515, |
| "learning_rate": 1.5844476342996453e-05, |
| "loss": 0.0107, |
| "step": 15240 |
| }, |
| { |
| "epoch": 1.562179881171891, |
| "grad_norm": 0.15372860431671143, |
| "learning_rate": 1.5827967359365306e-05, |
| "loss": 0.0127, |
| "step": 15250 |
| }, |
| { |
| "epoch": 1.5632042614218398, |
| "grad_norm": 0.1859782487154007, |
| "learning_rate": 1.5811457369675758e-05, |
| "loss": 0.0128, |
| "step": 15260 |
| }, |
| { |
| "epoch": 1.5642286416717885, |
| "grad_norm": 0.23249293863773346, |
| "learning_rate": 1.5794946393988982e-05, |
| "loss": 0.0159, |
| "step": 15270 |
| }, |
| { |
| "epoch": 1.5652530219217373, |
| "grad_norm": 0.28603556752204895, |
| "learning_rate": 1.5778434452367382e-05, |
| "loss": 0.0121, |
| "step": 15280 |
| }, |
| { |
| "epoch": 1.566277402171686, |
| "grad_norm": 0.17670315504074097, |
| "learning_rate": 1.576192156487452e-05, |
| "loss": 0.0118, |
| "step": 15290 |
| }, |
| { |
| "epoch": 1.5673017824216349, |
| "grad_norm": 0.20689330995082855, |
| "learning_rate": 1.574540775157511e-05, |
| "loss": 0.0146, |
| "step": 15300 |
| }, |
| { |
| "epoch": 1.5683261626715836, |
| "grad_norm": 0.11415048688650131, |
| "learning_rate": 1.5728893032534998e-05, |
| "loss": 0.0107, |
| "step": 15310 |
| }, |
| { |
| "epoch": 1.5693505429215324, |
| "grad_norm": 0.2701919674873352, |
| "learning_rate": 1.571237742782112e-05, |
| "loss": 0.0129, |
| "step": 15320 |
| }, |
| { |
| "epoch": 1.5703749231714812, |
| "grad_norm": 0.7738894820213318, |
| "learning_rate": 1.569586095750149e-05, |
| "loss": 0.0103, |
| "step": 15330 |
| }, |
| { |
| "epoch": 1.5713993034214302, |
| "grad_norm": 0.5723729133605957, |
| "learning_rate": 1.5679343641645193e-05, |
| "loss": 0.0141, |
| "step": 15340 |
| }, |
| { |
| "epoch": 1.5724236836713787, |
| "grad_norm": 0.2242802083492279, |
| "learning_rate": 1.5662825500322304e-05, |
| "loss": 0.0131, |
| "step": 15350 |
| }, |
| { |
| "epoch": 1.5734480639213277, |
| "grad_norm": 0.42409548163414, |
| "learning_rate": 1.5646306553603932e-05, |
| "loss": 0.0111, |
| "step": 15360 |
| }, |
| { |
| "epoch": 1.5744724441712763, |
| "grad_norm": 0.2120344042778015, |
| "learning_rate": 1.5629786821562154e-05, |
| "loss": 0.0102, |
| "step": 15370 |
| }, |
| { |
| "epoch": 1.5754968244212253, |
| "grad_norm": 0.1669318825006485, |
| "learning_rate": 1.561326632427001e-05, |
| "loss": 0.0098, |
| "step": 15380 |
| }, |
| { |
| "epoch": 1.5765212046711738, |
| "grad_norm": 0.21599072217941284, |
| "learning_rate": 1.5596745081801447e-05, |
| "loss": 0.0124, |
| "step": 15390 |
| }, |
| { |
| "epoch": 1.5775455849211228, |
| "grad_norm": 0.17248526215553284, |
| "learning_rate": 1.5580223114231348e-05, |
| "loss": 0.0136, |
| "step": 15400 |
| }, |
| { |
| "epoch": 1.5785699651710714, |
| "grad_norm": 0.2344321310520172, |
| "learning_rate": 1.5563700441635445e-05, |
| "loss": 0.0106, |
| "step": 15410 |
| }, |
| { |
| "epoch": 1.5795943454210204, |
| "grad_norm": 0.45455455780029297, |
| "learning_rate": 1.5547177084090358e-05, |
| "loss": 0.0128, |
| "step": 15420 |
| }, |
| { |
| "epoch": 1.580618725670969, |
| "grad_norm": 0.1927558034658432, |
| "learning_rate": 1.5530653061673514e-05, |
| "loss": 0.0146, |
| "step": 15430 |
| }, |
| { |
| "epoch": 1.581643105920918, |
| "grad_norm": 0.2750268280506134, |
| "learning_rate": 1.5514128394463167e-05, |
| "loss": 0.0106, |
| "step": 15440 |
| }, |
| { |
| "epoch": 1.5826674861708665, |
| "grad_norm": 0.15731939673423767, |
| "learning_rate": 1.549760310253834e-05, |
| "loss": 0.0113, |
| "step": 15450 |
| }, |
| { |
| "epoch": 1.5836918664208155, |
| "grad_norm": 0.3140643835067749, |
| "learning_rate": 1.5481077205978816e-05, |
| "loss": 0.0144, |
| "step": 15460 |
| }, |
| { |
| "epoch": 1.584716246670764, |
| "grad_norm": 0.2282572239637375, |
| "learning_rate": 1.5464550724865135e-05, |
| "loss": 0.0154, |
| "step": 15470 |
| }, |
| { |
| "epoch": 1.585740626920713, |
| "grad_norm": 0.3891045153141022, |
| "learning_rate": 1.5448023679278517e-05, |
| "loss": 0.0122, |
| "step": 15480 |
| }, |
| { |
| "epoch": 1.5867650071706616, |
| "grad_norm": 0.24993914365768433, |
| "learning_rate": 1.543149608930088e-05, |
| "loss": 0.0114, |
| "step": 15490 |
| }, |
| { |
| "epoch": 1.5877893874206106, |
| "grad_norm": 0.5398283004760742, |
| "learning_rate": 1.5414967975014805e-05, |
| "loss": 0.0135, |
| "step": 15500 |
| }, |
| { |
| "epoch": 1.5888137676705592, |
| "grad_norm": 0.3381696045398712, |
| "learning_rate": 1.539843935650351e-05, |
| "loss": 0.0112, |
| "step": 15510 |
| }, |
| { |
| "epoch": 1.5898381479205081, |
| "grad_norm": 0.42225590348243713, |
| "learning_rate": 1.5381910253850834e-05, |
| "loss": 0.0143, |
| "step": 15520 |
| }, |
| { |
| "epoch": 1.590862528170457, |
| "grad_norm": 0.13190264999866486, |
| "learning_rate": 1.5365380687141183e-05, |
| "loss": 0.0136, |
| "step": 15530 |
| }, |
| { |
| "epoch": 1.5918869084204057, |
| "grad_norm": 1.9489840269088745, |
| "learning_rate": 1.5348850676459544e-05, |
| "loss": 0.011, |
| "step": 15540 |
| }, |
| { |
| "epoch": 1.5929112886703545, |
| "grad_norm": 0.16822892427444458, |
| "learning_rate": 1.5332320241891438e-05, |
| "loss": 0.011, |
| "step": 15550 |
| }, |
| { |
| "epoch": 1.5939356689203033, |
| "grad_norm": 0.16203664243221283, |
| "learning_rate": 1.5315789403522908e-05, |
| "loss": 0.0112, |
| "step": 15560 |
| }, |
| { |
| "epoch": 1.594960049170252, |
| "grad_norm": 0.3673192858695984, |
| "learning_rate": 1.529925818144047e-05, |
| "loss": 0.0126, |
| "step": 15570 |
| }, |
| { |
| "epoch": 1.5959844294202008, |
| "grad_norm": 0.1719881147146225, |
| "learning_rate": 1.5282726595731126e-05, |
| "loss": 0.0127, |
| "step": 15580 |
| }, |
| { |
| "epoch": 1.5970088096701496, |
| "grad_norm": 0.19995316863059998, |
| "learning_rate": 1.5266194666482304e-05, |
| "loss": 0.0133, |
| "step": 15590 |
| }, |
| { |
| "epoch": 1.5980331899200984, |
| "grad_norm": 0.520680844783783, |
| "learning_rate": 1.5249662413781866e-05, |
| "loss": 0.0105, |
| "step": 15600 |
| }, |
| { |
| "epoch": 1.5990575701700471, |
| "grad_norm": 0.6654216051101685, |
| "learning_rate": 1.5233129857718054e-05, |
| "loss": 0.0137, |
| "step": 15610 |
| }, |
| { |
| "epoch": 1.600081950419996, |
| "grad_norm": 0.3126027286052704, |
| "learning_rate": 1.521659701837948e-05, |
| "loss": 0.0116, |
| "step": 15620 |
| }, |
| { |
| "epoch": 1.6011063306699447, |
| "grad_norm": 0.5229927897453308, |
| "learning_rate": 1.5200063915855107e-05, |
| "loss": 0.0112, |
| "step": 15630 |
| }, |
| { |
| "epoch": 1.6021307109198935, |
| "grad_norm": 0.2965322434902191, |
| "learning_rate": 1.5183530570234212e-05, |
| "loss": 0.0127, |
| "step": 15640 |
| }, |
| { |
| "epoch": 1.6031550911698422, |
| "grad_norm": 0.18564768135547638, |
| "learning_rate": 1.5166997001606361e-05, |
| "loss": 0.0119, |
| "step": 15650 |
| }, |
| { |
| "epoch": 1.604179471419791, |
| "grad_norm": 0.37351834774017334, |
| "learning_rate": 1.5150463230061417e-05, |
| "loss": 0.0124, |
| "step": 15660 |
| }, |
| { |
| "epoch": 1.6052038516697398, |
| "grad_norm": 0.37598884105682373, |
| "learning_rate": 1.5133929275689455e-05, |
| "loss": 0.0114, |
| "step": 15670 |
| }, |
| { |
| "epoch": 1.6062282319196886, |
| "grad_norm": 0.4020622670650482, |
| "learning_rate": 1.5117395158580795e-05, |
| "loss": 0.0151, |
| "step": 15680 |
| }, |
| { |
| "epoch": 1.6072526121696373, |
| "grad_norm": 0.4920317530632019, |
| "learning_rate": 1.510086089882595e-05, |
| "loss": 0.012, |
| "step": 15690 |
| }, |
| { |
| "epoch": 1.608276992419586, |
| "grad_norm": 0.3606903553009033, |
| "learning_rate": 1.5084326516515599e-05, |
| "loss": 0.0123, |
| "step": 15700 |
| }, |
| { |
| "epoch": 1.6093013726695349, |
| "grad_norm": 0.27939411997795105, |
| "learning_rate": 1.5067792031740582e-05, |
| "loss": 0.0143, |
| "step": 15710 |
| }, |
| { |
| "epoch": 1.6103257529194837, |
| "grad_norm": 0.2548691928386688, |
| "learning_rate": 1.505125746459185e-05, |
| "loss": 0.0117, |
| "step": 15720 |
| }, |
| { |
| "epoch": 1.6113501331694327, |
| "grad_norm": 0.3508896231651306, |
| "learning_rate": 1.5034722835160466e-05, |
| "loss": 0.0127, |
| "step": 15730 |
| }, |
| { |
| "epoch": 1.6123745134193812, |
| "grad_norm": 0.2370970994234085, |
| "learning_rate": 1.5018188163537566e-05, |
| "loss": 0.0125, |
| "step": 15740 |
| }, |
| { |
| "epoch": 1.6133988936693302, |
| "grad_norm": 0.7799522876739502, |
| "learning_rate": 1.5001653469814331e-05, |
| "loss": 0.0125, |
| "step": 15750 |
| }, |
| { |
| "epoch": 1.6144232739192788, |
| "grad_norm": 0.16358549892902374, |
| "learning_rate": 1.4985118774081978e-05, |
| "loss": 0.011, |
| "step": 15760 |
| }, |
| { |
| "epoch": 1.6154476541692278, |
| "grad_norm": 0.33201149106025696, |
| "learning_rate": 1.4968584096431708e-05, |
| "loss": 0.0134, |
| "step": 15770 |
| }, |
| { |
| "epoch": 1.6164720344191763, |
| "grad_norm": 0.43521931767463684, |
| "learning_rate": 1.4952049456954731e-05, |
| "loss": 0.0132, |
| "step": 15780 |
| }, |
| { |
| "epoch": 1.6174964146691253, |
| "grad_norm": 0.20080403983592987, |
| "learning_rate": 1.4935514875742182e-05, |
| "loss": 0.0115, |
| "step": 15790 |
| }, |
| { |
| "epoch": 1.6185207949190739, |
| "grad_norm": 0.23654168844223022, |
| "learning_rate": 1.4918980372885143e-05, |
| "loss": 0.0125, |
| "step": 15800 |
| }, |
| { |
| "epoch": 1.6195451751690229, |
| "grad_norm": 0.448486328125, |
| "learning_rate": 1.4902445968474591e-05, |
| "loss": 0.0097, |
| "step": 15810 |
| }, |
| { |
| "epoch": 1.6205695554189714, |
| "grad_norm": 0.24987168610095978, |
| "learning_rate": 1.4885911682601378e-05, |
| "loss": 0.0146, |
| "step": 15820 |
| }, |
| { |
| "epoch": 1.6215939356689204, |
| "grad_norm": 0.3721388578414917, |
| "learning_rate": 1.4869377535356235e-05, |
| "loss": 0.0128, |
| "step": 15830 |
| }, |
| { |
| "epoch": 1.622618315918869, |
| "grad_norm": 1.1354315280914307, |
| "learning_rate": 1.4852843546829704e-05, |
| "loss": 0.0112, |
| "step": 15840 |
| }, |
| { |
| "epoch": 1.623642696168818, |
| "grad_norm": 0.17065320909023285, |
| "learning_rate": 1.483630973711214e-05, |
| "loss": 0.0132, |
| "step": 15850 |
| }, |
| { |
| "epoch": 1.6246670764187665, |
| "grad_norm": 0.16755881905555725, |
| "learning_rate": 1.4819776126293677e-05, |
| "loss": 0.0122, |
| "step": 15860 |
| }, |
| { |
| "epoch": 1.6256914566687155, |
| "grad_norm": 0.2911642789840698, |
| "learning_rate": 1.4803242734464214e-05, |
| "loss": 0.0144, |
| "step": 15870 |
| }, |
| { |
| "epoch": 1.626715836918664, |
| "grad_norm": 0.16902896761894226, |
| "learning_rate": 1.4786709581713383e-05, |
| "loss": 0.0105, |
| "step": 15880 |
| }, |
| { |
| "epoch": 1.627740217168613, |
| "grad_norm": 0.3070961534976959, |
| "learning_rate": 1.4770176688130524e-05, |
| "loss": 0.0159, |
| "step": 15890 |
| }, |
| { |
| "epoch": 1.6287645974185616, |
| "grad_norm": 0.25144413113594055, |
| "learning_rate": 1.475364407380466e-05, |
| "loss": 0.0135, |
| "step": 15900 |
| }, |
| { |
| "epoch": 1.6297889776685106, |
| "grad_norm": 0.5798361301422119, |
| "learning_rate": 1.4737111758824476e-05, |
| "loss": 0.012, |
| "step": 15910 |
| }, |
| { |
| "epoch": 1.6308133579184594, |
| "grad_norm": 0.16372306644916534, |
| "learning_rate": 1.4720579763278289e-05, |
| "loss": 0.0137, |
| "step": 15920 |
| }, |
| { |
| "epoch": 1.6318377381684082, |
| "grad_norm": 0.399537593126297, |
| "learning_rate": 1.4704048107254042e-05, |
| "loss": 0.0166, |
| "step": 15930 |
| }, |
| { |
| "epoch": 1.632862118418357, |
| "grad_norm": 0.1206987127661705, |
| "learning_rate": 1.4687516810839246e-05, |
| "loss": 0.0108, |
| "step": 15940 |
| }, |
| { |
| "epoch": 1.6338864986683057, |
| "grad_norm": 0.1505211740732193, |
| "learning_rate": 1.4670985894120997e-05, |
| "loss": 0.0119, |
| "step": 15950 |
| }, |
| { |
| "epoch": 1.6349108789182545, |
| "grad_norm": 0.6216564178466797, |
| "learning_rate": 1.465445537718591e-05, |
| "loss": 0.0126, |
| "step": 15960 |
| }, |
| { |
| "epoch": 1.6359352591682033, |
| "grad_norm": 0.9141365885734558, |
| "learning_rate": 1.463792528012012e-05, |
| "loss": 0.0129, |
| "step": 15970 |
| }, |
| { |
| "epoch": 1.636959639418152, |
| "grad_norm": 0.2426273375749588, |
| "learning_rate": 1.4621395623009252e-05, |
| "loss": 0.0129, |
| "step": 15980 |
| }, |
| { |
| "epoch": 1.6379840196681008, |
| "grad_norm": 0.21901562809944153, |
| "learning_rate": 1.4604866425938407e-05, |
| "loss": 0.0116, |
| "step": 15990 |
| }, |
| { |
| "epoch": 1.6390083999180496, |
| "grad_norm": 0.3899953067302704, |
| "learning_rate": 1.4588337708992115e-05, |
| "loss": 0.0123, |
| "step": 16000 |
| }, |
| { |
| "epoch": 1.6400327801679984, |
| "grad_norm": 0.18610766530036926, |
| "learning_rate": 1.4571809492254318e-05, |
| "loss": 0.0143, |
| "step": 16010 |
| }, |
| { |
| "epoch": 1.6410571604179471, |
| "grad_norm": 0.1480393260717392, |
| "learning_rate": 1.4555281795808371e-05, |
| "loss": 0.0133, |
| "step": 16020 |
| }, |
| { |
| "epoch": 1.642081540667896, |
| "grad_norm": 0.22870606184005737, |
| "learning_rate": 1.4538754639736966e-05, |
| "loss": 0.011, |
| "step": 16030 |
| }, |
| { |
| "epoch": 1.6431059209178447, |
| "grad_norm": 0.18412669003009796, |
| "learning_rate": 1.4522228044122174e-05, |
| "loss": 0.0117, |
| "step": 16040 |
| }, |
| { |
| "epoch": 1.6441303011677935, |
| "grad_norm": 0.5071516633033752, |
| "learning_rate": 1.4505702029045359e-05, |
| "loss": 0.012, |
| "step": 16050 |
| }, |
| { |
| "epoch": 1.6451546814177422, |
| "grad_norm": 0.278516560792923, |
| "learning_rate": 1.4489176614587189e-05, |
| "loss": 0.0132, |
| "step": 16060 |
| }, |
| { |
| "epoch": 1.646179061667691, |
| "grad_norm": 0.1435951143503189, |
| "learning_rate": 1.4472651820827595e-05, |
| "loss": 0.0135, |
| "step": 16070 |
| }, |
| { |
| "epoch": 1.6472034419176398, |
| "grad_norm": 0.15499387681484222, |
| "learning_rate": 1.4456127667845766e-05, |
| "loss": 0.0146, |
| "step": 16080 |
| }, |
| { |
| "epoch": 1.6482278221675886, |
| "grad_norm": 0.1704794466495514, |
| "learning_rate": 1.4439604175720104e-05, |
| "loss": 0.0149, |
| "step": 16090 |
| }, |
| { |
| "epoch": 1.6492522024175373, |
| "grad_norm": 0.3329339921474457, |
| "learning_rate": 1.4423081364528212e-05, |
| "loss": 0.0157, |
| "step": 16100 |
| }, |
| { |
| "epoch": 1.6502765826674861, |
| "grad_norm": 0.5079371333122253, |
| "learning_rate": 1.4406559254346863e-05, |
| "loss": 0.0127, |
| "step": 16110 |
| }, |
| { |
| "epoch": 1.6513009629174351, |
| "grad_norm": 1.0479086637496948, |
| "learning_rate": 1.4390037865251978e-05, |
| "loss": 0.0152, |
| "step": 16120 |
| }, |
| { |
| "epoch": 1.6523253431673837, |
| "grad_norm": 0.6414211988449097, |
| "learning_rate": 1.4373517217318597e-05, |
| "loss": 0.0111, |
| "step": 16130 |
| }, |
| { |
| "epoch": 1.6533497234173327, |
| "grad_norm": 0.18504156172275543, |
| "learning_rate": 1.4356997330620875e-05, |
| "loss": 0.0115, |
| "step": 16140 |
| }, |
| { |
| "epoch": 1.6543741036672812, |
| "grad_norm": 0.6075134873390198, |
| "learning_rate": 1.434047822523203e-05, |
| "loss": 0.0139, |
| "step": 16150 |
| }, |
| { |
| "epoch": 1.6553984839172302, |
| "grad_norm": 0.3212584853172302, |
| "learning_rate": 1.4323959921224322e-05, |
| "loss": 0.0115, |
| "step": 16160 |
| }, |
| { |
| "epoch": 1.6564228641671788, |
| "grad_norm": 0.16684280335903168, |
| "learning_rate": 1.4307442438669066e-05, |
| "loss": 0.0103, |
| "step": 16170 |
| }, |
| { |
| "epoch": 1.6574472444171278, |
| "grad_norm": 0.12114604562520981, |
| "learning_rate": 1.4290925797636546e-05, |
| "loss": 0.0147, |
| "step": 16180 |
| }, |
| { |
| "epoch": 1.6584716246670763, |
| "grad_norm": 0.16698206961154938, |
| "learning_rate": 1.4274410018196039e-05, |
| "loss": 0.0107, |
| "step": 16190 |
| }, |
| { |
| "epoch": 1.6594960049170253, |
| "grad_norm": 0.362398624420166, |
| "learning_rate": 1.4257895120415783e-05, |
| "loss": 0.014, |
| "step": 16200 |
| }, |
| { |
| "epoch": 1.6605203851669739, |
| "grad_norm": 0.6945683360099792, |
| "learning_rate": 1.4241381124362933e-05, |
| "loss": 0.0136, |
| "step": 16210 |
| }, |
| { |
| "epoch": 1.6615447654169229, |
| "grad_norm": 0.3858531415462494, |
| "learning_rate": 1.4224868050103546e-05, |
| "loss": 0.0114, |
| "step": 16220 |
| }, |
| { |
| "epoch": 1.6625691456668714, |
| "grad_norm": 0.7119501233100891, |
| "learning_rate": 1.4208355917702573e-05, |
| "loss": 0.0127, |
| "step": 16230 |
| }, |
| { |
| "epoch": 1.6635935259168204, |
| "grad_norm": 0.1909143030643463, |
| "learning_rate": 1.41918447472238e-05, |
| "loss": 0.0121, |
| "step": 16240 |
| }, |
| { |
| "epoch": 1.664617906166769, |
| "grad_norm": 0.5542840957641602, |
| "learning_rate": 1.4175334558729869e-05, |
| "loss": 0.0154, |
| "step": 16250 |
| }, |
| { |
| "epoch": 1.665642286416718, |
| "grad_norm": 0.4423328638076782, |
| "learning_rate": 1.4158825372282211e-05, |
| "loss": 0.0124, |
| "step": 16260 |
| }, |
| { |
| "epoch": 1.6666666666666665, |
| "grad_norm": 0.3055688738822937, |
| "learning_rate": 1.4142317207941042e-05, |
| "loss": 0.0123, |
| "step": 16270 |
| }, |
| { |
| "epoch": 1.6676910469166155, |
| "grad_norm": 0.13187655806541443, |
| "learning_rate": 1.4125810085765339e-05, |
| "loss": 0.0124, |
| "step": 16280 |
| }, |
| { |
| "epoch": 1.668715427166564, |
| "grad_norm": 0.44667479395866394, |
| "learning_rate": 1.410930402581281e-05, |
| "loss": 0.0132, |
| "step": 16290 |
| }, |
| { |
| "epoch": 1.669739807416513, |
| "grad_norm": 0.7828208804130554, |
| "learning_rate": 1.4092799048139876e-05, |
| "loss": 0.0125, |
| "step": 16300 |
| }, |
| { |
| "epoch": 1.6707641876664618, |
| "grad_norm": 0.4062930941581726, |
| "learning_rate": 1.4076295172801647e-05, |
| "loss": 0.0154, |
| "step": 16310 |
| }, |
| { |
| "epoch": 1.6717885679164106, |
| "grad_norm": 0.32807251811027527, |
| "learning_rate": 1.4059792419851884e-05, |
| "loss": 0.0143, |
| "step": 16320 |
| }, |
| { |
| "epoch": 1.6728129481663594, |
| "grad_norm": 0.5853361487388611, |
| "learning_rate": 1.4043290809342984e-05, |
| "loss": 0.0109, |
| "step": 16330 |
| }, |
| { |
| "epoch": 1.6738373284163082, |
| "grad_norm": 0.3566567897796631, |
| "learning_rate": 1.4026790361325958e-05, |
| "loss": 0.0085, |
| "step": 16340 |
| }, |
| { |
| "epoch": 1.674861708666257, |
| "grad_norm": 0.2144250124692917, |
| "learning_rate": 1.4010291095850415e-05, |
| "loss": 0.0103, |
| "step": 16350 |
| }, |
| { |
| "epoch": 1.6758860889162057, |
| "grad_norm": 0.3882652819156647, |
| "learning_rate": 1.3993793032964515e-05, |
| "loss": 0.0117, |
| "step": 16360 |
| }, |
| { |
| "epoch": 1.6769104691661545, |
| "grad_norm": 0.2607526481151581, |
| "learning_rate": 1.3977296192714956e-05, |
| "loss": 0.0127, |
| "step": 16370 |
| }, |
| { |
| "epoch": 1.6779348494161033, |
| "grad_norm": 0.2961130440235138, |
| "learning_rate": 1.396080059514696e-05, |
| "loss": 0.0135, |
| "step": 16380 |
| }, |
| { |
| "epoch": 1.678959229666052, |
| "grad_norm": 0.186335951089859, |
| "learning_rate": 1.3944306260304232e-05, |
| "loss": 0.0131, |
| "step": 16390 |
| }, |
| { |
| "epoch": 1.6799836099160008, |
| "grad_norm": 0.2621561586856842, |
| "learning_rate": 1.3927813208228936e-05, |
| "loss": 0.0129, |
| "step": 16400 |
| }, |
| { |
| "epoch": 1.6810079901659496, |
| "grad_norm": 0.15196587145328522, |
| "learning_rate": 1.39113214589617e-05, |
| "loss": 0.0112, |
| "step": 16410 |
| }, |
| { |
| "epoch": 1.6820323704158984, |
| "grad_norm": 0.27515655755996704, |
| "learning_rate": 1.3894831032541545e-05, |
| "loss": 0.0155, |
| "step": 16420 |
| }, |
| { |
| "epoch": 1.6830567506658471, |
| "grad_norm": 0.3887760043144226, |
| "learning_rate": 1.3878341949005895e-05, |
| "loss": 0.0113, |
| "step": 16430 |
| }, |
| { |
| "epoch": 1.684081130915796, |
| "grad_norm": 0.1677476465702057, |
| "learning_rate": 1.3861854228390546e-05, |
| "loss": 0.0114, |
| "step": 16440 |
| }, |
| { |
| "epoch": 1.6851055111657447, |
| "grad_norm": 0.1832284927368164, |
| "learning_rate": 1.3845367890729621e-05, |
| "loss": 0.0128, |
| "step": 16450 |
| }, |
| { |
| "epoch": 1.6861298914156935, |
| "grad_norm": 0.2938048243522644, |
| "learning_rate": 1.3828882956055591e-05, |
| "loss": 0.0116, |
| "step": 16460 |
| }, |
| { |
| "epoch": 1.6871542716656422, |
| "grad_norm": 0.311798095703125, |
| "learning_rate": 1.38123994443992e-05, |
| "loss": 0.015, |
| "step": 16470 |
| }, |
| { |
| "epoch": 1.688178651915591, |
| "grad_norm": 0.35156646370887756, |
| "learning_rate": 1.3795917375789465e-05, |
| "loss": 0.0111, |
| "step": 16480 |
| }, |
| { |
| "epoch": 1.6892030321655398, |
| "grad_norm": 0.12406232953071594, |
| "learning_rate": 1.3779436770253658e-05, |
| "loss": 0.0102, |
| "step": 16490 |
| }, |
| { |
| "epoch": 1.6902274124154886, |
| "grad_norm": 0.47371745109558105, |
| "learning_rate": 1.3762957647817258e-05, |
| "loss": 0.0107, |
| "step": 16500 |
| }, |
| { |
| "epoch": 1.6912517926654376, |
| "grad_norm": 0.3539314270019531, |
| "learning_rate": 1.374648002850397e-05, |
| "loss": 0.0145, |
| "step": 16510 |
| }, |
| { |
| "epoch": 1.6922761729153861, |
| "grad_norm": 0.1341642588376999, |
| "learning_rate": 1.3730003932335642e-05, |
| "loss": 0.0124, |
| "step": 16520 |
| }, |
| { |
| "epoch": 1.6933005531653351, |
| "grad_norm": 0.2647896707057953, |
| "learning_rate": 1.3713529379332292e-05, |
| "loss": 0.0142, |
| "step": 16530 |
| }, |
| { |
| "epoch": 1.6943249334152837, |
| "grad_norm": 0.2546844482421875, |
| "learning_rate": 1.3697056389512054e-05, |
| "loss": 0.0119, |
| "step": 16540 |
| }, |
| { |
| "epoch": 1.6953493136652327, |
| "grad_norm": 0.45358991622924805, |
| "learning_rate": 1.3680584982891162e-05, |
| "loss": 0.0139, |
| "step": 16550 |
| }, |
| { |
| "epoch": 1.6963736939151812, |
| "grad_norm": 0.24719421565532684, |
| "learning_rate": 1.3664115179483921e-05, |
| "loss": 0.0122, |
| "step": 16560 |
| }, |
| { |
| "epoch": 1.6973980741651302, |
| "grad_norm": 0.3718450367450714, |
| "learning_rate": 1.364764699930271e-05, |
| "loss": 0.0118, |
| "step": 16570 |
| }, |
| { |
| "epoch": 1.6984224544150788, |
| "grad_norm": 0.5170800685882568, |
| "learning_rate": 1.363118046235791e-05, |
| "loss": 0.0156, |
| "step": 16580 |
| }, |
| { |
| "epoch": 1.6994468346650278, |
| "grad_norm": 0.19894546270370483, |
| "learning_rate": 1.3614715588657921e-05, |
| "loss": 0.0148, |
| "step": 16590 |
| }, |
| { |
| "epoch": 1.7004712149149763, |
| "grad_norm": 0.18417096138000488, |
| "learning_rate": 1.3598252398209118e-05, |
| "loss": 0.0123, |
| "step": 16600 |
| }, |
| { |
| "epoch": 1.7014955951649253, |
| "grad_norm": 0.664962887763977, |
| "learning_rate": 1.358179091101582e-05, |
| "loss": 0.0133, |
| "step": 16610 |
| }, |
| { |
| "epoch": 1.7025199754148739, |
| "grad_norm": 0.484524130821228, |
| "learning_rate": 1.3565331147080299e-05, |
| "loss": 0.0135, |
| "step": 16620 |
| }, |
| { |
| "epoch": 1.7035443556648229, |
| "grad_norm": 0.4392278790473938, |
| "learning_rate": 1.3548873126402716e-05, |
| "loss": 0.0112, |
| "step": 16630 |
| }, |
| { |
| "epoch": 1.7045687359147714, |
| "grad_norm": 0.5205717086791992, |
| "learning_rate": 1.3532416868981117e-05, |
| "loss": 0.0127, |
| "step": 16640 |
| }, |
| { |
| "epoch": 1.7055931161647204, |
| "grad_norm": 0.5060121417045593, |
| "learning_rate": 1.3515962394811401e-05, |
| "loss": 0.012, |
| "step": 16650 |
| }, |
| { |
| "epoch": 1.706617496414669, |
| "grad_norm": 0.9905937314033508, |
| "learning_rate": 1.3499509723887313e-05, |
| "loss": 0.01, |
| "step": 16660 |
| }, |
| { |
| "epoch": 1.707641876664618, |
| "grad_norm": 0.23239454627037048, |
| "learning_rate": 1.3483058876200396e-05, |
| "loss": 0.0124, |
| "step": 16670 |
| }, |
| { |
| "epoch": 1.7086662569145665, |
| "grad_norm": 0.3469693660736084, |
| "learning_rate": 1.3466609871739987e-05, |
| "loss": 0.0109, |
| "step": 16680 |
| }, |
| { |
| "epoch": 1.7096906371645155, |
| "grad_norm": 0.16700613498687744, |
| "learning_rate": 1.3450162730493172e-05, |
| "loss": 0.0136, |
| "step": 16690 |
| }, |
| { |
| "epoch": 1.7107150174144643, |
| "grad_norm": 0.21758684515953064, |
| "learning_rate": 1.3433717472444781e-05, |
| "loss": 0.0114, |
| "step": 16700 |
| }, |
| { |
| "epoch": 1.711739397664413, |
| "grad_norm": 0.2410518378019333, |
| "learning_rate": 1.3417274117577343e-05, |
| "loss": 0.0129, |
| "step": 16710 |
| }, |
| { |
| "epoch": 1.7127637779143619, |
| "grad_norm": 0.6778503060340881, |
| "learning_rate": 1.3400832685871102e-05, |
| "loss": 0.009, |
| "step": 16720 |
| }, |
| { |
| "epoch": 1.7137881581643106, |
| "grad_norm": 0.15354952216148376, |
| "learning_rate": 1.3384393197303936e-05, |
| "loss": 0.0137, |
| "step": 16730 |
| }, |
| { |
| "epoch": 1.7148125384142594, |
| "grad_norm": 0.31132760643959045, |
| "learning_rate": 1.336795567185138e-05, |
| "loss": 0.0104, |
| "step": 16740 |
| }, |
| { |
| "epoch": 1.7158369186642082, |
| "grad_norm": 0.23697757720947266, |
| "learning_rate": 1.3351520129486573e-05, |
| "loss": 0.0144, |
| "step": 16750 |
| }, |
| { |
| "epoch": 1.716861298914157, |
| "grad_norm": 0.25235486030578613, |
| "learning_rate": 1.3335086590180252e-05, |
| "loss": 0.0108, |
| "step": 16760 |
| }, |
| { |
| "epoch": 1.7178856791641057, |
| "grad_norm": 0.3697109818458557, |
| "learning_rate": 1.3318655073900708e-05, |
| "loss": 0.0126, |
| "step": 16770 |
| }, |
| { |
| "epoch": 1.7189100594140545, |
| "grad_norm": 0.2317681610584259, |
| "learning_rate": 1.3302225600613796e-05, |
| "loss": 0.011, |
| "step": 16780 |
| }, |
| { |
| "epoch": 1.7199344396640033, |
| "grad_norm": 0.15237580239772797, |
| "learning_rate": 1.3285798190282869e-05, |
| "loss": 0.0113, |
| "step": 16790 |
| }, |
| { |
| "epoch": 1.720958819913952, |
| "grad_norm": 0.260105699300766, |
| "learning_rate": 1.3269372862868776e-05, |
| "loss": 0.0142, |
| "step": 16800 |
| }, |
| { |
| "epoch": 1.7219832001639008, |
| "grad_norm": 0.35300225019454956, |
| "learning_rate": 1.3252949638329845e-05, |
| "loss": 0.0108, |
| "step": 16810 |
| }, |
| { |
| "epoch": 1.7230075804138496, |
| "grad_norm": 0.24469733238220215, |
| "learning_rate": 1.3236528536621834e-05, |
| "loss": 0.0143, |
| "step": 16820 |
| }, |
| { |
| "epoch": 1.7240319606637984, |
| "grad_norm": 0.12540049850940704, |
| "learning_rate": 1.3220109577697943e-05, |
| "loss": 0.0115, |
| "step": 16830 |
| }, |
| { |
| "epoch": 1.7250563409137472, |
| "grad_norm": 0.9491755366325378, |
| "learning_rate": 1.3203692781508745e-05, |
| "loss": 0.0139, |
| "step": 16840 |
| }, |
| { |
| "epoch": 1.726080721163696, |
| "grad_norm": 0.13651156425476074, |
| "learning_rate": 1.3187278168002198e-05, |
| "loss": 0.0114, |
| "step": 16850 |
| }, |
| { |
| "epoch": 1.7271051014136447, |
| "grad_norm": 0.1758773922920227, |
| "learning_rate": 1.3170865757123606e-05, |
| "loss": 0.0146, |
| "step": 16860 |
| }, |
| { |
| "epoch": 1.7281294816635935, |
| "grad_norm": 0.16502247750759125, |
| "learning_rate": 1.3154455568815592e-05, |
| "loss": 0.0109, |
| "step": 16870 |
| }, |
| { |
| "epoch": 1.7291538619135423, |
| "grad_norm": 0.3135172426700592, |
| "learning_rate": 1.3138047623018087e-05, |
| "loss": 0.0122, |
| "step": 16880 |
| }, |
| { |
| "epoch": 1.730178242163491, |
| "grad_norm": 0.39172297716140747, |
| "learning_rate": 1.3121641939668293e-05, |
| "loss": 0.0135, |
| "step": 16890 |
| }, |
| { |
| "epoch": 1.73120262241344, |
| "grad_norm": 0.21979132294654846, |
| "learning_rate": 1.3105238538700659e-05, |
| "loss": 0.0123, |
| "step": 16900 |
| }, |
| { |
| "epoch": 1.7322270026633886, |
| "grad_norm": 0.20279595255851746, |
| "learning_rate": 1.3088837440046866e-05, |
| "loss": 0.0116, |
| "step": 16910 |
| }, |
| { |
| "epoch": 1.7332513829133376, |
| "grad_norm": 0.2845816910266876, |
| "learning_rate": 1.3072438663635787e-05, |
| "loss": 0.0154, |
| "step": 16920 |
| }, |
| { |
| "epoch": 1.7342757631632861, |
| "grad_norm": 0.3355707824230194, |
| "learning_rate": 1.3056042229393497e-05, |
| "loss": 0.0127, |
| "step": 16930 |
| }, |
| { |
| "epoch": 1.7353001434132351, |
| "grad_norm": 0.4383613169193268, |
| "learning_rate": 1.30396481572432e-05, |
| "loss": 0.0119, |
| "step": 16940 |
| }, |
| { |
| "epoch": 1.7363245236631837, |
| "grad_norm": 0.3221171498298645, |
| "learning_rate": 1.3023256467105237e-05, |
| "loss": 0.0143, |
| "step": 16950 |
| }, |
| { |
| "epoch": 1.7373489039131327, |
| "grad_norm": 0.2653118669986725, |
| "learning_rate": 1.3006867178897067e-05, |
| "loss": 0.0117, |
| "step": 16960 |
| }, |
| { |
| "epoch": 1.7383732841630812, |
| "grad_norm": 0.655436098575592, |
| "learning_rate": 1.2990480312533213e-05, |
| "loss": 0.0122, |
| "step": 16970 |
| }, |
| { |
| "epoch": 1.7393976644130302, |
| "grad_norm": 0.430197149515152, |
| "learning_rate": 1.2974095887925258e-05, |
| "loss": 0.0122, |
| "step": 16980 |
| }, |
| { |
| "epoch": 1.7404220446629788, |
| "grad_norm": 0.19856366515159607, |
| "learning_rate": 1.2957713924981837e-05, |
| "loss": 0.0125, |
| "step": 16990 |
| }, |
| { |
| "epoch": 1.7414464249129278, |
| "grad_norm": 0.4852939546108246, |
| "learning_rate": 1.2941334443608572e-05, |
| "loss": 0.0118, |
| "step": 17000 |
| }, |
| { |
| "epoch": 1.7424708051628763, |
| "grad_norm": 0.9122607707977295, |
| "learning_rate": 1.2924957463708073e-05, |
| "loss": 0.0128, |
| "step": 17010 |
| }, |
| { |
| "epoch": 1.7434951854128253, |
| "grad_norm": 0.334989458322525, |
| "learning_rate": 1.2908583005179924e-05, |
| "loss": 0.011, |
| "step": 17020 |
| }, |
| { |
| "epoch": 1.7445195656627739, |
| "grad_norm": 0.6880782246589661, |
| "learning_rate": 1.2892211087920624e-05, |
| "loss": 0.0119, |
| "step": 17030 |
| }, |
| { |
| "epoch": 1.7455439459127229, |
| "grad_norm": 0.22520025074481964, |
| "learning_rate": 1.287584173182361e-05, |
| "loss": 0.0125, |
| "step": 17040 |
| }, |
| { |
| "epoch": 1.7465683261626714, |
| "grad_norm": 0.26832300424575806, |
| "learning_rate": 1.2859474956779188e-05, |
| "loss": 0.0123, |
| "step": 17050 |
| }, |
| { |
| "epoch": 1.7475927064126204, |
| "grad_norm": 0.21992184221744537, |
| "learning_rate": 1.2843110782674532e-05, |
| "loss": 0.0123, |
| "step": 17060 |
| }, |
| { |
| "epoch": 1.748617086662569, |
| "grad_norm": 0.2512039244174957, |
| "learning_rate": 1.2826749229393655e-05, |
| "loss": 0.0109, |
| "step": 17070 |
| }, |
| { |
| "epoch": 1.749641466912518, |
| "grad_norm": 0.2234875112771988, |
| "learning_rate": 1.2810390316817381e-05, |
| "loss": 0.0116, |
| "step": 17080 |
| }, |
| { |
| "epoch": 1.7506658471624668, |
| "grad_norm": 0.7877574563026428, |
| "learning_rate": 1.2794034064823346e-05, |
| "loss": 0.0122, |
| "step": 17090 |
| }, |
| { |
| "epoch": 1.7516902274124155, |
| "grad_norm": 0.20786187052726746, |
| "learning_rate": 1.277768049328593e-05, |
| "loss": 0.0119, |
| "step": 17100 |
| }, |
| { |
| "epoch": 1.7527146076623643, |
| "grad_norm": 0.24137593805789948, |
| "learning_rate": 1.2761329622076269e-05, |
| "loss": 0.0116, |
| "step": 17110 |
| }, |
| { |
| "epoch": 1.753738987912313, |
| "grad_norm": 0.20107676088809967, |
| "learning_rate": 1.274498147106221e-05, |
| "loss": 0.0115, |
| "step": 17120 |
| }, |
| { |
| "epoch": 1.7547633681622619, |
| "grad_norm": 0.13584840297698975, |
| "learning_rate": 1.272863606010829e-05, |
| "loss": 0.0147, |
| "step": 17130 |
| }, |
| { |
| "epoch": 1.7557877484122106, |
| "grad_norm": 0.3015841841697693, |
| "learning_rate": 1.2712293409075739e-05, |
| "loss": 0.0102, |
| "step": 17140 |
| }, |
| { |
| "epoch": 1.7568121286621594, |
| "grad_norm": 0.3974423110485077, |
| "learning_rate": 1.2695953537822413e-05, |
| "loss": 0.0124, |
| "step": 17150 |
| }, |
| { |
| "epoch": 1.7578365089121082, |
| "grad_norm": 0.168575257062912, |
| "learning_rate": 1.2679616466202792e-05, |
| "loss": 0.0149, |
| "step": 17160 |
| }, |
| { |
| "epoch": 1.758860889162057, |
| "grad_norm": 0.4580287039279938, |
| "learning_rate": 1.2663282214067967e-05, |
| "loss": 0.0102, |
| "step": 17170 |
| }, |
| { |
| "epoch": 1.7598852694120057, |
| "grad_norm": 0.12398034334182739, |
| "learning_rate": 1.264695080126559e-05, |
| "loss": 0.0099, |
| "step": 17180 |
| }, |
| { |
| "epoch": 1.7609096496619545, |
| "grad_norm": 0.20169466733932495, |
| "learning_rate": 1.263062224763986e-05, |
| "loss": 0.0119, |
| "step": 17190 |
| }, |
| { |
| "epoch": 1.7619340299119033, |
| "grad_norm": 0.2577838599681854, |
| "learning_rate": 1.2614296573031522e-05, |
| "loss": 0.0126, |
| "step": 17200 |
| }, |
| { |
| "epoch": 1.762958410161852, |
| "grad_norm": 1.2920314073562622, |
| "learning_rate": 1.259797379727781e-05, |
| "loss": 0.0116, |
| "step": 17210 |
| }, |
| { |
| "epoch": 1.7639827904118008, |
| "grad_norm": 0.3633112609386444, |
| "learning_rate": 1.2581653940212423e-05, |
| "loss": 0.0136, |
| "step": 17220 |
| }, |
| { |
| "epoch": 1.7650071706617496, |
| "grad_norm": 0.33048486709594727, |
| "learning_rate": 1.2565337021665537e-05, |
| "loss": 0.0131, |
| "step": 17230 |
| }, |
| { |
| "epoch": 1.7660315509116984, |
| "grad_norm": 0.27000537514686584, |
| "learning_rate": 1.2549023061463739e-05, |
| "loss": 0.0109, |
| "step": 17240 |
| }, |
| { |
| "epoch": 1.7670559311616472, |
| "grad_norm": 2.9015114307403564, |
| "learning_rate": 1.2532712079430038e-05, |
| "loss": 0.0138, |
| "step": 17250 |
| }, |
| { |
| "epoch": 1.768080311411596, |
| "grad_norm": 0.315607488155365, |
| "learning_rate": 1.2516404095383813e-05, |
| "loss": 0.012, |
| "step": 17260 |
| }, |
| { |
| "epoch": 1.7691046916615447, |
| "grad_norm": 0.30724623799324036, |
| "learning_rate": 1.2500099129140796e-05, |
| "loss": 0.0123, |
| "step": 17270 |
| }, |
| { |
| "epoch": 1.7701290719114935, |
| "grad_norm": 0.145302414894104, |
| "learning_rate": 1.2483797200513065e-05, |
| "loss": 0.0153, |
| "step": 17280 |
| }, |
| { |
| "epoch": 1.7711534521614425, |
| "grad_norm": 0.1357731819152832, |
| "learning_rate": 1.2467498329308992e-05, |
| "loss": 0.0141, |
| "step": 17290 |
| }, |
| { |
| "epoch": 1.772177832411391, |
| "grad_norm": 0.40171942114830017, |
| "learning_rate": 1.2451202535333253e-05, |
| "loss": 0.0122, |
| "step": 17300 |
| }, |
| { |
| "epoch": 1.77320221266134, |
| "grad_norm": 0.1585506647825241, |
| "learning_rate": 1.2434909838386768e-05, |
| "loss": 0.0133, |
| "step": 17310 |
| }, |
| { |
| "epoch": 1.7742265929112886, |
| "grad_norm": 1.3782010078430176, |
| "learning_rate": 1.2418620258266703e-05, |
| "loss": 0.0099, |
| "step": 17320 |
| }, |
| { |
| "epoch": 1.7752509731612376, |
| "grad_norm": 0.2317962497472763, |
| "learning_rate": 1.2402333814766432e-05, |
| "loss": 0.0127, |
| "step": 17330 |
| }, |
| { |
| "epoch": 1.7762753534111861, |
| "grad_norm": 0.18060310184955597, |
| "learning_rate": 1.2386050527675515e-05, |
| "loss": 0.0136, |
| "step": 17340 |
| }, |
| { |
| "epoch": 1.7772997336611351, |
| "grad_norm": 0.2932301461696625, |
| "learning_rate": 1.2369770416779693e-05, |
| "loss": 0.0116, |
| "step": 17350 |
| }, |
| { |
| "epoch": 1.7783241139110837, |
| "grad_norm": 0.15665015578269958, |
| "learning_rate": 1.2353493501860828e-05, |
| "loss": 0.0127, |
| "step": 17360 |
| }, |
| { |
| "epoch": 1.7793484941610327, |
| "grad_norm": 0.37491491436958313, |
| "learning_rate": 1.2337219802696905e-05, |
| "loss": 0.0141, |
| "step": 17370 |
| }, |
| { |
| "epoch": 1.7803728744109812, |
| "grad_norm": 0.5742243528366089, |
| "learning_rate": 1.2320949339062008e-05, |
| "loss": 0.0157, |
| "step": 17380 |
| }, |
| { |
| "epoch": 1.7813972546609302, |
| "grad_norm": 0.181259423494339, |
| "learning_rate": 1.2304682130726281e-05, |
| "loss": 0.0137, |
| "step": 17390 |
| }, |
| { |
| "epoch": 1.7824216349108788, |
| "grad_norm": 0.2689659297466278, |
| "learning_rate": 1.2288418197455913e-05, |
| "loss": 0.0116, |
| "step": 17400 |
| }, |
| { |
| "epoch": 1.7834460151608278, |
| "grad_norm": 0.613187313079834, |
| "learning_rate": 1.2272157559013126e-05, |
| "loss": 0.0133, |
| "step": 17410 |
| }, |
| { |
| "epoch": 1.7844703954107763, |
| "grad_norm": 0.2752578556537628, |
| "learning_rate": 1.2255900235156123e-05, |
| "loss": 0.0116, |
| "step": 17420 |
| }, |
| { |
| "epoch": 1.7854947756607253, |
| "grad_norm": 0.4182765483856201, |
| "learning_rate": 1.2239646245639087e-05, |
| "loss": 0.0116, |
| "step": 17430 |
| }, |
| { |
| "epoch": 1.786519155910674, |
| "grad_norm": 0.48946669697761536, |
| "learning_rate": 1.2223395610212142e-05, |
| "loss": 0.0116, |
| "step": 17440 |
| }, |
| { |
| "epoch": 1.787543536160623, |
| "grad_norm": 0.2814163267612457, |
| "learning_rate": 1.2207148348621348e-05, |
| "loss": 0.0123, |
| "step": 17450 |
| }, |
| { |
| "epoch": 1.7885679164105714, |
| "grad_norm": 0.3896740674972534, |
| "learning_rate": 1.219090448060866e-05, |
| "loss": 0.0117, |
| "step": 17460 |
| }, |
| { |
| "epoch": 1.7895922966605204, |
| "grad_norm": 0.4878005087375641, |
| "learning_rate": 1.217466402591191e-05, |
| "loss": 0.012, |
| "step": 17470 |
| }, |
| { |
| "epoch": 1.7906166769104692, |
| "grad_norm": 0.13916684687137604, |
| "learning_rate": 1.2158427004264785e-05, |
| "loss": 0.0129, |
| "step": 17480 |
| }, |
| { |
| "epoch": 1.791641057160418, |
| "grad_norm": 0.108980193734169, |
| "learning_rate": 1.2142193435396788e-05, |
| "loss": 0.0105, |
| "step": 17490 |
| }, |
| { |
| "epoch": 1.7926654374103668, |
| "grad_norm": 0.3070497512817383, |
| "learning_rate": 1.2125963339033234e-05, |
| "loss": 0.0135, |
| "step": 17500 |
| }, |
| { |
| "epoch": 1.7936898176603155, |
| "grad_norm": 0.11105567961931229, |
| "learning_rate": 1.2109736734895237e-05, |
| "loss": 0.0092, |
| "step": 17510 |
| }, |
| { |
| "epoch": 1.7947141979102643, |
| "grad_norm": 0.19401030242443085, |
| "learning_rate": 1.2093513642699634e-05, |
| "loss": 0.0146, |
| "step": 17520 |
| }, |
| { |
| "epoch": 1.795738578160213, |
| "grad_norm": 0.10553690046072006, |
| "learning_rate": 1.2077294082159024e-05, |
| "loss": 0.0104, |
| "step": 17530 |
| }, |
| { |
| "epoch": 1.7967629584101619, |
| "grad_norm": 0.12308189272880554, |
| "learning_rate": 1.2061078072981697e-05, |
| "loss": 0.0104, |
| "step": 17540 |
| }, |
| { |
| "epoch": 1.7977873386601106, |
| "grad_norm": 0.6914670467376709, |
| "learning_rate": 1.2044865634871626e-05, |
| "loss": 0.0131, |
| "step": 17550 |
| }, |
| { |
| "epoch": 1.7988117189100594, |
| "grad_norm": 0.19085097312927246, |
| "learning_rate": 1.2028656787528468e-05, |
| "loss": 0.0111, |
| "step": 17560 |
| }, |
| { |
| "epoch": 1.7998360991600082, |
| "grad_norm": 0.14124077558517456, |
| "learning_rate": 1.2012451550647491e-05, |
| "loss": 0.0126, |
| "step": 17570 |
| }, |
| { |
| "epoch": 1.800860479409957, |
| "grad_norm": 0.2104957401752472, |
| "learning_rate": 1.1996249943919589e-05, |
| "loss": 0.014, |
| "step": 17580 |
| }, |
| { |
| "epoch": 1.8018848596599057, |
| "grad_norm": 0.3127448856830597, |
| "learning_rate": 1.1980051987031237e-05, |
| "loss": 0.0137, |
| "step": 17590 |
| }, |
| { |
| "epoch": 1.8029092399098545, |
| "grad_norm": 0.4436160922050476, |
| "learning_rate": 1.1963857699664483e-05, |
| "loss": 0.0097, |
| "step": 17600 |
| }, |
| { |
| "epoch": 1.8039336201598033, |
| "grad_norm": 0.17081530392169952, |
| "learning_rate": 1.1947667101496911e-05, |
| "loss": 0.0111, |
| "step": 17610 |
| }, |
| { |
| "epoch": 1.804958000409752, |
| "grad_norm": 0.1459273397922516, |
| "learning_rate": 1.1931480212201633e-05, |
| "loss": 0.0132, |
| "step": 17620 |
| }, |
| { |
| "epoch": 1.8059823806597008, |
| "grad_norm": 0.23161429166793823, |
| "learning_rate": 1.1915297051447237e-05, |
| "loss": 0.0104, |
| "step": 17630 |
| }, |
| { |
| "epoch": 1.8070067609096496, |
| "grad_norm": 0.37272289395332336, |
| "learning_rate": 1.1899117638897793e-05, |
| "loss": 0.0113, |
| "step": 17640 |
| }, |
| { |
| "epoch": 1.8080311411595984, |
| "grad_norm": 0.4716450273990631, |
| "learning_rate": 1.1882941994212811e-05, |
| "loss": 0.0095, |
| "step": 17650 |
| }, |
| { |
| "epoch": 1.8090555214095472, |
| "grad_norm": 0.6288555264472961, |
| "learning_rate": 1.1866770137047221e-05, |
| "loss": 0.0124, |
| "step": 17660 |
| }, |
| { |
| "epoch": 1.810079901659496, |
| "grad_norm": 0.28994685411453247, |
| "learning_rate": 1.1850602087051363e-05, |
| "loss": 0.0098, |
| "step": 17670 |
| }, |
| { |
| "epoch": 1.811104281909445, |
| "grad_norm": 0.48982322216033936, |
| "learning_rate": 1.183443786387094e-05, |
| "loss": 0.0145, |
| "step": 17680 |
| }, |
| { |
| "epoch": 1.8121286621593935, |
| "grad_norm": 0.19510626792907715, |
| "learning_rate": 1.1818277487147005e-05, |
| "loss": 0.0118, |
| "step": 17690 |
| }, |
| { |
| "epoch": 1.8131530424093425, |
| "grad_norm": 0.9195302128791809, |
| "learning_rate": 1.180212097651594e-05, |
| "loss": 0.0119, |
| "step": 17700 |
| }, |
| { |
| "epoch": 1.814177422659291, |
| "grad_norm": 0.5732307434082031, |
| "learning_rate": 1.1785968351609422e-05, |
| "loss": 0.0156, |
| "step": 17710 |
| }, |
| { |
| "epoch": 1.81520180290924, |
| "grad_norm": 0.4677578806877136, |
| "learning_rate": 1.1769819632054427e-05, |
| "loss": 0.0107, |
| "step": 17720 |
| }, |
| { |
| "epoch": 1.8162261831591886, |
| "grad_norm": 0.6228772401809692, |
| "learning_rate": 1.1753674837473165e-05, |
| "loss": 0.0122, |
| "step": 17730 |
| }, |
| { |
| "epoch": 1.8172505634091376, |
| "grad_norm": 0.28368785977363586, |
| "learning_rate": 1.1737533987483081e-05, |
| "loss": 0.0114, |
| "step": 17740 |
| }, |
| { |
| "epoch": 1.8182749436590862, |
| "grad_norm": 0.5946877598762512, |
| "learning_rate": 1.1721397101696835e-05, |
| "loss": 0.0133, |
| "step": 17750 |
| }, |
| { |
| "epoch": 1.8192993239090351, |
| "grad_norm": 0.2884722352027893, |
| "learning_rate": 1.1705264199722257e-05, |
| "loss": 0.0114, |
| "step": 17760 |
| }, |
| { |
| "epoch": 1.8203237041589837, |
| "grad_norm": 0.4753044843673706, |
| "learning_rate": 1.1689135301162356e-05, |
| "loss": 0.0124, |
| "step": 17770 |
| }, |
| { |
| "epoch": 1.8213480844089327, |
| "grad_norm": 0.25160688161849976, |
| "learning_rate": 1.1673010425615258e-05, |
| "loss": 0.0131, |
| "step": 17780 |
| }, |
| { |
| "epoch": 1.8223724646588813, |
| "grad_norm": 0.37275123596191406, |
| "learning_rate": 1.1656889592674208e-05, |
| "loss": 0.0114, |
| "step": 17790 |
| }, |
| { |
| "epoch": 1.8233968449088302, |
| "grad_norm": 0.18329623341560364, |
| "learning_rate": 1.1640772821927533e-05, |
| "loss": 0.016, |
| "step": 17800 |
| }, |
| { |
| "epoch": 1.8244212251587788, |
| "grad_norm": 0.34593334794044495, |
| "learning_rate": 1.162466013295864e-05, |
| "loss": 0.0123, |
| "step": 17810 |
| }, |
| { |
| "epoch": 1.8254456054087278, |
| "grad_norm": 0.4751342236995697, |
| "learning_rate": 1.160855154534595e-05, |
| "loss": 0.011, |
| "step": 17820 |
| }, |
| { |
| "epoch": 1.8264699856586764, |
| "grad_norm": 0.13010437786579132, |
| "learning_rate": 1.1592447078662931e-05, |
| "loss": 0.0116, |
| "step": 17830 |
| }, |
| { |
| "epoch": 1.8274943659086254, |
| "grad_norm": 0.22769075632095337, |
| "learning_rate": 1.1576346752478025e-05, |
| "loss": 0.0103, |
| "step": 17840 |
| }, |
| { |
| "epoch": 1.828518746158574, |
| "grad_norm": 0.15802736580371857, |
| "learning_rate": 1.1560250586354643e-05, |
| "loss": 0.0115, |
| "step": 17850 |
| }, |
| { |
| "epoch": 1.829543126408523, |
| "grad_norm": 0.563073992729187, |
| "learning_rate": 1.1544158599851146e-05, |
| "loss": 0.0133, |
| "step": 17860 |
| }, |
| { |
| "epoch": 1.8305675066584717, |
| "grad_norm": 0.3027478754520416, |
| "learning_rate": 1.1528070812520809e-05, |
| "loss": 0.0138, |
| "step": 17870 |
| }, |
| { |
| "epoch": 1.8315918869084205, |
| "grad_norm": 0.37770262360572815, |
| "learning_rate": 1.1511987243911821e-05, |
| "loss": 0.0126, |
| "step": 17880 |
| }, |
| { |
| "epoch": 1.8326162671583692, |
| "grad_norm": 0.6384933590888977, |
| "learning_rate": 1.1495907913567228e-05, |
| "loss": 0.0111, |
| "step": 17890 |
| }, |
| { |
| "epoch": 1.833640647408318, |
| "grad_norm": 0.12204177677631378, |
| "learning_rate": 1.1479832841024936e-05, |
| "loss": 0.0124, |
| "step": 17900 |
| }, |
| { |
| "epoch": 1.8346650276582668, |
| "grad_norm": 0.9139361381530762, |
| "learning_rate": 1.1463762045817675e-05, |
| "loss": 0.0125, |
| "step": 17910 |
| }, |
| { |
| "epoch": 1.8356894079082156, |
| "grad_norm": 0.524914026260376, |
| "learning_rate": 1.1447695547472964e-05, |
| "loss": 0.0147, |
| "step": 17920 |
| }, |
| { |
| "epoch": 1.8367137881581643, |
| "grad_norm": 0.1673925518989563, |
| "learning_rate": 1.1431633365513132e-05, |
| "loss": 0.0109, |
| "step": 17930 |
| }, |
| { |
| "epoch": 1.837738168408113, |
| "grad_norm": 0.5542842149734497, |
| "learning_rate": 1.1415575519455235e-05, |
| "loss": 0.0147, |
| "step": 17940 |
| }, |
| { |
| "epoch": 1.8387625486580619, |
| "grad_norm": 0.5920175313949585, |
| "learning_rate": 1.1399522028811069e-05, |
| "loss": 0.014, |
| "step": 17950 |
| }, |
| { |
| "epoch": 1.8397869289080107, |
| "grad_norm": 0.2451794296503067, |
| "learning_rate": 1.1383472913087143e-05, |
| "loss": 0.0122, |
| "step": 17960 |
| }, |
| { |
| "epoch": 1.8408113091579594, |
| "grad_norm": 0.1467130333185196, |
| "learning_rate": 1.1367428191784643e-05, |
| "loss": 0.0114, |
| "step": 17970 |
| }, |
| { |
| "epoch": 1.8418356894079082, |
| "grad_norm": 0.19405756890773773, |
| "learning_rate": 1.1351387884399413e-05, |
| "loss": 0.0115, |
| "step": 17980 |
| }, |
| { |
| "epoch": 1.842860069657857, |
| "grad_norm": 0.3612029254436493, |
| "learning_rate": 1.1335352010421953e-05, |
| "loss": 0.0113, |
| "step": 17990 |
| }, |
| { |
| "epoch": 1.8438844499078058, |
| "grad_norm": 0.2280980795621872, |
| "learning_rate": 1.1319320589337352e-05, |
| "loss": 0.0124, |
| "step": 18000 |
| }, |
| { |
| "epoch": 1.8449088301577545, |
| "grad_norm": 0.7842258810997009, |
| "learning_rate": 1.1303293640625299e-05, |
| "loss": 0.0125, |
| "step": 18010 |
| }, |
| { |
| "epoch": 1.8459332104077033, |
| "grad_norm": 0.6953174471855164, |
| "learning_rate": 1.1287271183760052e-05, |
| "loss": 0.0099, |
| "step": 18020 |
| }, |
| { |
| "epoch": 1.846957590657652, |
| "grad_norm": 0.27951809763908386, |
| "learning_rate": 1.1271253238210398e-05, |
| "loss": 0.0142, |
| "step": 18030 |
| }, |
| { |
| "epoch": 1.8479819709076009, |
| "grad_norm": 0.45048537850379944, |
| "learning_rate": 1.1255239823439663e-05, |
| "loss": 0.0101, |
| "step": 18040 |
| }, |
| { |
| "epoch": 1.8490063511575496, |
| "grad_norm": 0.2814275324344635, |
| "learning_rate": 1.1239230958905652e-05, |
| "loss": 0.0124, |
| "step": 18050 |
| }, |
| { |
| "epoch": 1.8500307314074984, |
| "grad_norm": 0.4940418303012848, |
| "learning_rate": 1.1223226664060641e-05, |
| "loss": 0.0097, |
| "step": 18060 |
| }, |
| { |
| "epoch": 1.8510551116574474, |
| "grad_norm": 0.2801018953323364, |
| "learning_rate": 1.1207226958351362e-05, |
| "loss": 0.0097, |
| "step": 18070 |
| }, |
| { |
| "epoch": 1.852079491907396, |
| "grad_norm": 0.3635845482349396, |
| "learning_rate": 1.1191231861218955e-05, |
| "loss": 0.0118, |
| "step": 18080 |
| }, |
| { |
| "epoch": 1.853103872157345, |
| "grad_norm": 0.2087579369544983, |
| "learning_rate": 1.1175241392098988e-05, |
| "loss": 0.0108, |
| "step": 18090 |
| }, |
| { |
| "epoch": 1.8541282524072935, |
| "grad_norm": 0.39255040884017944, |
| "learning_rate": 1.1159255570421376e-05, |
| "loss": 0.0096, |
| "step": 18100 |
| }, |
| { |
| "epoch": 1.8551526326572425, |
| "grad_norm": 0.2322632372379303, |
| "learning_rate": 1.1143274415610408e-05, |
| "loss": 0.0108, |
| "step": 18110 |
| }, |
| { |
| "epoch": 1.856177012907191, |
| "grad_norm": 0.23586155474185944, |
| "learning_rate": 1.1127297947084692e-05, |
| "loss": 0.0124, |
| "step": 18120 |
| }, |
| { |
| "epoch": 1.85720139315714, |
| "grad_norm": 0.587321400642395, |
| "learning_rate": 1.1111326184257135e-05, |
| "loss": 0.0118, |
| "step": 18130 |
| }, |
| { |
| "epoch": 1.8582257734070886, |
| "grad_norm": 0.6931122541427612, |
| "learning_rate": 1.109535914653495e-05, |
| "loss": 0.0136, |
| "step": 18140 |
| }, |
| { |
| "epoch": 1.8592501536570376, |
| "grad_norm": 0.874700129032135, |
| "learning_rate": 1.1079396853319582e-05, |
| "loss": 0.0137, |
| "step": 18150 |
| }, |
| { |
| "epoch": 1.8602745339069862, |
| "grad_norm": 0.45225510001182556, |
| "learning_rate": 1.1063439324006724e-05, |
| "loss": 0.0108, |
| "step": 18160 |
| }, |
| { |
| "epoch": 1.8612989141569352, |
| "grad_norm": 0.12028156220912933, |
| "learning_rate": 1.1047486577986285e-05, |
| "loss": 0.0115, |
| "step": 18170 |
| }, |
| { |
| "epoch": 1.8623232944068837, |
| "grad_norm": 0.15535667538642883, |
| "learning_rate": 1.1031538634642353e-05, |
| "loss": 0.0112, |
| "step": 18180 |
| }, |
| { |
| "epoch": 1.8633476746568327, |
| "grad_norm": 0.893221378326416, |
| "learning_rate": 1.1015595513353172e-05, |
| "loss": 0.0111, |
| "step": 18190 |
| }, |
| { |
| "epoch": 1.8643720549067813, |
| "grad_norm": 0.6724375486373901, |
| "learning_rate": 1.0999657233491156e-05, |
| "loss": 0.0122, |
| "step": 18200 |
| }, |
| { |
| "epoch": 1.8653964351567303, |
| "grad_norm": 0.27230575680732727, |
| "learning_rate": 1.0983723814422808e-05, |
| "loss": 0.0113, |
| "step": 18210 |
| }, |
| { |
| "epoch": 1.8664208154066788, |
| "grad_norm": 0.3997829556465149, |
| "learning_rate": 1.0967795275508735e-05, |
| "loss": 0.0128, |
| "step": 18220 |
| }, |
| { |
| "epoch": 1.8674451956566278, |
| "grad_norm": 0.18954512476921082, |
| "learning_rate": 1.095187163610361e-05, |
| "loss": 0.0111, |
| "step": 18230 |
| }, |
| { |
| "epoch": 1.8684695759065764, |
| "grad_norm": 0.2385961264371872, |
| "learning_rate": 1.0935952915556163e-05, |
| "loss": 0.0114, |
| "step": 18240 |
| }, |
| { |
| "epoch": 1.8694939561565254, |
| "grad_norm": 0.4808025062084198, |
| "learning_rate": 1.0920039133209137e-05, |
| "loss": 0.0146, |
| "step": 18250 |
| }, |
| { |
| "epoch": 1.8705183364064741, |
| "grad_norm": 0.15356230735778809, |
| "learning_rate": 1.0904130308399283e-05, |
| "loss": 0.01, |
| "step": 18260 |
| }, |
| { |
| "epoch": 1.871542716656423, |
| "grad_norm": 0.14623777568340302, |
| "learning_rate": 1.0888226460457318e-05, |
| "loss": 0.0125, |
| "step": 18270 |
| }, |
| { |
| "epoch": 1.8725670969063717, |
| "grad_norm": 0.4730057716369629, |
| "learning_rate": 1.087232760870792e-05, |
| "loss": 0.0123, |
| "step": 18280 |
| }, |
| { |
| "epoch": 1.8735914771563205, |
| "grad_norm": 1.2261250019073486, |
| "learning_rate": 1.0856433772469686e-05, |
| "loss": 0.0122, |
| "step": 18290 |
| }, |
| { |
| "epoch": 1.8746158574062692, |
| "grad_norm": 0.26899483799934387, |
| "learning_rate": 1.0840544971055135e-05, |
| "loss": 0.0112, |
| "step": 18300 |
| }, |
| { |
| "epoch": 1.875640237656218, |
| "grad_norm": 0.8467230796813965, |
| "learning_rate": 1.0824661223770655e-05, |
| "loss": 0.013, |
| "step": 18310 |
| }, |
| { |
| "epoch": 1.8766646179061668, |
| "grad_norm": 0.2070353478193283, |
| "learning_rate": 1.0808782549916498e-05, |
| "loss": 0.0117, |
| "step": 18320 |
| }, |
| { |
| "epoch": 1.8776889981561156, |
| "grad_norm": 0.17462734878063202, |
| "learning_rate": 1.079290896878675e-05, |
| "loss": 0.0132, |
| "step": 18330 |
| }, |
| { |
| "epoch": 1.8787133784060643, |
| "grad_norm": 0.4248207211494446, |
| "learning_rate": 1.07770404996693e-05, |
| "loss": 0.0114, |
| "step": 18340 |
| }, |
| { |
| "epoch": 1.8797377586560131, |
| "grad_norm": 0.732032835483551, |
| "learning_rate": 1.0761177161845848e-05, |
| "loss": 0.0113, |
| "step": 18350 |
| }, |
| { |
| "epoch": 1.8807621389059619, |
| "grad_norm": 0.23866704106330872, |
| "learning_rate": 1.0745318974591841e-05, |
| "loss": 0.013, |
| "step": 18360 |
| }, |
| { |
| "epoch": 1.8817865191559107, |
| "grad_norm": 0.43293970823287964, |
| "learning_rate": 1.0729465957176469e-05, |
| "loss": 0.0128, |
| "step": 18370 |
| }, |
| { |
| "epoch": 1.8828108994058594, |
| "grad_norm": 0.255631685256958, |
| "learning_rate": 1.0713618128862642e-05, |
| "loss": 0.0127, |
| "step": 18380 |
| }, |
| { |
| "epoch": 1.8838352796558082, |
| "grad_norm": 0.14825154840946198, |
| "learning_rate": 1.0697775508906969e-05, |
| "loss": 0.0091, |
| "step": 18390 |
| }, |
| { |
| "epoch": 1.884859659905757, |
| "grad_norm": 0.2974323034286499, |
| "learning_rate": 1.068193811655972e-05, |
| "loss": 0.0118, |
| "step": 18400 |
| }, |
| { |
| "epoch": 1.8858840401557058, |
| "grad_norm": 0.09230872988700867, |
| "learning_rate": 1.0666105971064835e-05, |
| "loss": 0.011, |
| "step": 18410 |
| }, |
| { |
| "epoch": 1.8869084204056545, |
| "grad_norm": 0.1551532745361328, |
| "learning_rate": 1.0650279091659852e-05, |
| "loss": 0.0102, |
| "step": 18420 |
| }, |
| { |
| "epoch": 1.8879328006556033, |
| "grad_norm": 0.21479369699954987, |
| "learning_rate": 1.0634457497575926e-05, |
| "loss": 0.0115, |
| "step": 18430 |
| }, |
| { |
| "epoch": 1.888957180905552, |
| "grad_norm": 0.5313194394111633, |
| "learning_rate": 1.061864120803778e-05, |
| "loss": 0.0103, |
| "step": 18440 |
| }, |
| { |
| "epoch": 1.8899815611555009, |
| "grad_norm": 0.5225009322166443, |
| "learning_rate": 1.0602830242263701e-05, |
| "loss": 0.013, |
| "step": 18450 |
| }, |
| { |
| "epoch": 1.8910059414054499, |
| "grad_norm": 0.6374237537384033, |
| "learning_rate": 1.0587024619465506e-05, |
| "loss": 0.0099, |
| "step": 18460 |
| }, |
| { |
| "epoch": 1.8920303216553984, |
| "grad_norm": 0.12096835672855377, |
| "learning_rate": 1.0571224358848517e-05, |
| "loss": 0.0139, |
| "step": 18470 |
| }, |
| { |
| "epoch": 1.8930547019053474, |
| "grad_norm": 0.13517457246780396, |
| "learning_rate": 1.0555429479611537e-05, |
| "loss": 0.0117, |
| "step": 18480 |
| }, |
| { |
| "epoch": 1.894079082155296, |
| "grad_norm": 0.12858080863952637, |
| "learning_rate": 1.0539640000946835e-05, |
| "loss": 0.0103, |
| "step": 18490 |
| }, |
| { |
| "epoch": 1.895103462405245, |
| "grad_norm": 0.383672833442688, |
| "learning_rate": 1.0523855942040111e-05, |
| "loss": 0.0107, |
| "step": 18500 |
| }, |
| { |
| "epoch": 1.8961278426551935, |
| "grad_norm": 0.436855673789978, |
| "learning_rate": 1.0508077322070496e-05, |
| "loss": 0.0111, |
| "step": 18510 |
| }, |
| { |
| "epoch": 1.8971522229051425, |
| "grad_norm": 0.28863710165023804, |
| "learning_rate": 1.0492304160210495e-05, |
| "loss": 0.0111, |
| "step": 18520 |
| }, |
| { |
| "epoch": 1.898176603155091, |
| "grad_norm": 0.24150292575359344, |
| "learning_rate": 1.0476536475625985e-05, |
| "loss": 0.0135, |
| "step": 18530 |
| }, |
| { |
| "epoch": 1.89920098340504, |
| "grad_norm": 0.11909355223178864, |
| "learning_rate": 1.0460774287476195e-05, |
| "loss": 0.0097, |
| "step": 18540 |
| }, |
| { |
| "epoch": 1.9002253636549886, |
| "grad_norm": 0.29321151971817017, |
| "learning_rate": 1.044501761491366e-05, |
| "loss": 0.0112, |
| "step": 18550 |
| }, |
| { |
| "epoch": 1.9012497439049376, |
| "grad_norm": 0.16932713985443115, |
| "learning_rate": 1.0429266477084235e-05, |
| "loss": 0.0149, |
| "step": 18560 |
| }, |
| { |
| "epoch": 1.9022741241548862, |
| "grad_norm": 0.22183865308761597, |
| "learning_rate": 1.0413520893127034e-05, |
| "loss": 0.011, |
| "step": 18570 |
| }, |
| { |
| "epoch": 1.9032985044048352, |
| "grad_norm": 0.3859527111053467, |
| "learning_rate": 1.0397780882174424e-05, |
| "loss": 0.0099, |
| "step": 18580 |
| }, |
| { |
| "epoch": 1.9043228846547837, |
| "grad_norm": 0.3806331157684326, |
| "learning_rate": 1.0382046463351998e-05, |
| "loss": 0.0119, |
| "step": 18590 |
| }, |
| { |
| "epoch": 1.9053472649047327, |
| "grad_norm": 0.3770027756690979, |
| "learning_rate": 1.0366317655778568e-05, |
| "loss": 0.0092, |
| "step": 18600 |
| }, |
| { |
| "epoch": 1.9063716451546813, |
| "grad_norm": 0.4361119270324707, |
| "learning_rate": 1.035059447856611e-05, |
| "loss": 0.0101, |
| "step": 18610 |
| }, |
| { |
| "epoch": 1.9073960254046303, |
| "grad_norm": 0.7225149273872375, |
| "learning_rate": 1.0334876950819777e-05, |
| "loss": 0.0106, |
| "step": 18620 |
| }, |
| { |
| "epoch": 1.9084204056545788, |
| "grad_norm": 0.35265228152275085, |
| "learning_rate": 1.0319165091637837e-05, |
| "loss": 0.0131, |
| "step": 18630 |
| }, |
| { |
| "epoch": 1.9094447859045278, |
| "grad_norm": 0.13477647304534912, |
| "learning_rate": 1.0303458920111686e-05, |
| "loss": 0.0105, |
| "step": 18640 |
| }, |
| { |
| "epoch": 1.9104691661544766, |
| "grad_norm": 0.6580484509468079, |
| "learning_rate": 1.0287758455325803e-05, |
| "loss": 0.0119, |
| "step": 18650 |
| }, |
| { |
| "epoch": 1.9114935464044254, |
| "grad_norm": 0.10426221042871475, |
| "learning_rate": 1.0272063716357729e-05, |
| "loss": 0.0106, |
| "step": 18660 |
| }, |
| { |
| "epoch": 1.9125179266543741, |
| "grad_norm": 0.26669803261756897, |
| "learning_rate": 1.0256374722278057e-05, |
| "loss": 0.0103, |
| "step": 18670 |
| }, |
| { |
| "epoch": 1.913542306904323, |
| "grad_norm": 0.6429899334907532, |
| "learning_rate": 1.0240691492150388e-05, |
| "loss": 0.0116, |
| "step": 18680 |
| }, |
| { |
| "epoch": 1.9145666871542717, |
| "grad_norm": 0.44394451379776, |
| "learning_rate": 1.0225014045031334e-05, |
| "loss": 0.0136, |
| "step": 18690 |
| }, |
| { |
| "epoch": 1.9155910674042205, |
| "grad_norm": 0.5241128206253052, |
| "learning_rate": 1.0209342399970468e-05, |
| "loss": 0.0126, |
| "step": 18700 |
| }, |
| { |
| "epoch": 1.9166154476541692, |
| "grad_norm": 0.14107242226600647, |
| "learning_rate": 1.019367657601031e-05, |
| "loss": 0.0124, |
| "step": 18710 |
| }, |
| { |
| "epoch": 1.917639827904118, |
| "grad_norm": 0.2960687577724457, |
| "learning_rate": 1.0178016592186323e-05, |
| "loss": 0.011, |
| "step": 18720 |
| }, |
| { |
| "epoch": 1.9186642081540668, |
| "grad_norm": 0.28569573163986206, |
| "learning_rate": 1.0162362467526861e-05, |
| "loss": 0.013, |
| "step": 18730 |
| }, |
| { |
| "epoch": 1.9196885884040156, |
| "grad_norm": 0.15021054446697235, |
| "learning_rate": 1.0146714221053159e-05, |
| "loss": 0.0135, |
| "step": 18740 |
| }, |
| { |
| "epoch": 1.9207129686539643, |
| "grad_norm": 0.3242014944553375, |
| "learning_rate": 1.0131071871779314e-05, |
| "loss": 0.0118, |
| "step": 18750 |
| }, |
| { |
| "epoch": 1.9217373489039131, |
| "grad_norm": 0.4146333634853363, |
| "learning_rate": 1.0115435438712249e-05, |
| "loss": 0.0146, |
| "step": 18760 |
| }, |
| { |
| "epoch": 1.922761729153862, |
| "grad_norm": 0.5535944700241089, |
| "learning_rate": 1.0099804940851716e-05, |
| "loss": 0.0122, |
| "step": 18770 |
| }, |
| { |
| "epoch": 1.9237861094038107, |
| "grad_norm": 1.1524237394332886, |
| "learning_rate": 1.0084180397190238e-05, |
| "loss": 0.0118, |
| "step": 18780 |
| }, |
| { |
| "epoch": 1.9248104896537594, |
| "grad_norm": 0.4808367192745209, |
| "learning_rate": 1.006856182671311e-05, |
| "loss": 0.0091, |
| "step": 18790 |
| }, |
| { |
| "epoch": 1.9258348699037082, |
| "grad_norm": 0.2144576460123062, |
| "learning_rate": 1.0052949248398367e-05, |
| "loss": 0.012, |
| "step": 18800 |
| }, |
| { |
| "epoch": 1.926859250153657, |
| "grad_norm": 0.22437527775764465, |
| "learning_rate": 1.003734268121676e-05, |
| "loss": 0.0141, |
| "step": 18810 |
| }, |
| { |
| "epoch": 1.9278836304036058, |
| "grad_norm": 0.16689532995224, |
| "learning_rate": 1.0021742144131744e-05, |
| "loss": 0.0119, |
| "step": 18820 |
| }, |
| { |
| "epoch": 1.9289080106535546, |
| "grad_norm": 0.8312996029853821, |
| "learning_rate": 1.000614765609944e-05, |
| "loss": 0.0145, |
| "step": 18830 |
| }, |
| { |
| "epoch": 1.9299323909035033, |
| "grad_norm": 0.6707962155342102, |
| "learning_rate": 9.990559236068628e-06, |
| "loss": 0.0121, |
| "step": 18840 |
| }, |
| { |
| "epoch": 1.9309567711534523, |
| "grad_norm": 0.178877592086792, |
| "learning_rate": 9.974976902980703e-06, |
| "loss": 0.0098, |
| "step": 18850 |
| }, |
| { |
| "epoch": 1.9319811514034009, |
| "grad_norm": 0.3031393587589264, |
| "learning_rate": 9.959400675769668e-06, |
| "loss": 0.012, |
| "step": 18860 |
| }, |
| { |
| "epoch": 1.9330055316533499, |
| "grad_norm": 0.37016579508781433, |
| "learning_rate": 9.943830573362106e-06, |
| "loss": 0.0137, |
| "step": 18870 |
| }, |
| { |
| "epoch": 1.9340299119032984, |
| "grad_norm": 0.8534723520278931, |
| "learning_rate": 9.928266614677168e-06, |
| "loss": 0.0098, |
| "step": 18880 |
| }, |
| { |
| "epoch": 1.9350542921532474, |
| "grad_norm": 0.3111989200115204, |
| "learning_rate": 9.912708818626524e-06, |
| "loss": 0.0101, |
| "step": 18890 |
| }, |
| { |
| "epoch": 1.936078672403196, |
| "grad_norm": 0.7605631947517395, |
| "learning_rate": 9.897157204114369e-06, |
| "loss": 0.0092, |
| "step": 18900 |
| }, |
| { |
| "epoch": 1.937103052653145, |
| "grad_norm": 0.9483798146247864, |
| "learning_rate": 9.881611790037375e-06, |
| "loss": 0.0107, |
| "step": 18910 |
| }, |
| { |
| "epoch": 1.9381274329030935, |
| "grad_norm": 0.552249014377594, |
| "learning_rate": 9.866072595284684e-06, |
| "loss": 0.0105, |
| "step": 18920 |
| }, |
| { |
| "epoch": 1.9391518131530425, |
| "grad_norm": 0.6689223647117615, |
| "learning_rate": 9.850539638737894e-06, |
| "loss": 0.012, |
| "step": 18930 |
| }, |
| { |
| "epoch": 1.940176193402991, |
| "grad_norm": 0.24246972799301147, |
| "learning_rate": 9.835012939271006e-06, |
| "loss": 0.0104, |
| "step": 18940 |
| }, |
| { |
| "epoch": 1.94120057365294, |
| "grad_norm": 0.24735592305660248, |
| "learning_rate": 9.819492515750418e-06, |
| "loss": 0.0104, |
| "step": 18950 |
| }, |
| { |
| "epoch": 1.9422249539028886, |
| "grad_norm": 0.8990926146507263, |
| "learning_rate": 9.803978387034918e-06, |
| "loss": 0.0105, |
| "step": 18960 |
| }, |
| { |
| "epoch": 1.9432493341528376, |
| "grad_norm": 0.23809437453746796, |
| "learning_rate": 9.788470571975624e-06, |
| "loss": 0.0104, |
| "step": 18970 |
| }, |
| { |
| "epoch": 1.9442737144027862, |
| "grad_norm": 0.38396546244621277, |
| "learning_rate": 9.772969089416008e-06, |
| "loss": 0.0112, |
| "step": 18980 |
| }, |
| { |
| "epoch": 1.9452980946527352, |
| "grad_norm": 0.25134187936782837, |
| "learning_rate": 9.757473958191822e-06, |
| "loss": 0.0107, |
| "step": 18990 |
| }, |
| { |
| "epoch": 1.9463224749026837, |
| "grad_norm": 0.3143649101257324, |
| "learning_rate": 9.741985197131118e-06, |
| "loss": 0.0123, |
| "step": 19000 |
| }, |
| { |
| "epoch": 1.9473468551526327, |
| "grad_norm": 0.16194580495357513, |
| "learning_rate": 9.726502825054195e-06, |
| "loss": 0.0127, |
| "step": 19010 |
| }, |
| { |
| "epoch": 1.9483712354025813, |
| "grad_norm": 0.16495661437511444, |
| "learning_rate": 9.711026860773595e-06, |
| "loss": 0.0113, |
| "step": 19020 |
| }, |
| { |
| "epoch": 1.9493956156525303, |
| "grad_norm": 0.129180446267128, |
| "learning_rate": 9.69555732309408e-06, |
| "loss": 0.0111, |
| "step": 19030 |
| }, |
| { |
| "epoch": 1.950419995902479, |
| "grad_norm": 0.30498477816581726, |
| "learning_rate": 9.68009423081259e-06, |
| "loss": 0.0128, |
| "step": 19040 |
| }, |
| { |
| "epoch": 1.9514443761524278, |
| "grad_norm": 0.7881457805633545, |
| "learning_rate": 9.664637602718249e-06, |
| "loss": 0.0113, |
| "step": 19050 |
| }, |
| { |
| "epoch": 1.9524687564023766, |
| "grad_norm": 0.2032298594713211, |
| "learning_rate": 9.649187457592313e-06, |
| "loss": 0.0121, |
| "step": 19060 |
| }, |
| { |
| "epoch": 1.9534931366523254, |
| "grad_norm": 0.1237609013915062, |
| "learning_rate": 9.63374381420816e-06, |
| "loss": 0.0098, |
| "step": 19070 |
| }, |
| { |
| "epoch": 1.9545175169022742, |
| "grad_norm": 0.3659161329269409, |
| "learning_rate": 9.618306691331279e-06, |
| "loss": 0.0121, |
| "step": 19080 |
| }, |
| { |
| "epoch": 1.955541897152223, |
| "grad_norm": 0.21692106127738953, |
| "learning_rate": 9.602876107719228e-06, |
| "loss": 0.0089, |
| "step": 19090 |
| }, |
| { |
| "epoch": 1.9565662774021717, |
| "grad_norm": 0.13327930867671967, |
| "learning_rate": 9.587452082121621e-06, |
| "loss": 0.0114, |
| "step": 19100 |
| }, |
| { |
| "epoch": 1.9575906576521205, |
| "grad_norm": 2.4722342491149902, |
| "learning_rate": 9.57203463328011e-06, |
| "loss": 0.0126, |
| "step": 19110 |
| }, |
| { |
| "epoch": 1.9586150379020693, |
| "grad_norm": 0.25132232904434204, |
| "learning_rate": 9.556623779928341e-06, |
| "loss": 0.0119, |
| "step": 19120 |
| }, |
| { |
| "epoch": 1.959639418152018, |
| "grad_norm": 0.1803523451089859, |
| "learning_rate": 9.541219540791949e-06, |
| "loss": 0.011, |
| "step": 19130 |
| }, |
| { |
| "epoch": 1.9606637984019668, |
| "grad_norm": 0.4660748243331909, |
| "learning_rate": 9.525821934588551e-06, |
| "loss": 0.0106, |
| "step": 19140 |
| }, |
| { |
| "epoch": 1.9616881786519156, |
| "grad_norm": 0.5150839686393738, |
| "learning_rate": 9.510430980027684e-06, |
| "loss": 0.0103, |
| "step": 19150 |
| }, |
| { |
| "epoch": 1.9627125589018644, |
| "grad_norm": 0.2933011054992676, |
| "learning_rate": 9.495046695810809e-06, |
| "loss": 0.0111, |
| "step": 19160 |
| }, |
| { |
| "epoch": 1.9637369391518131, |
| "grad_norm": 0.22308166325092316, |
| "learning_rate": 9.47966910063128e-06, |
| "loss": 0.0118, |
| "step": 19170 |
| }, |
| { |
| "epoch": 1.964761319401762, |
| "grad_norm": 0.38911253213882446, |
| "learning_rate": 9.464298213174325e-06, |
| "loss": 0.01, |
| "step": 19180 |
| }, |
| { |
| "epoch": 1.9657856996517107, |
| "grad_norm": 0.5436503291130066, |
| "learning_rate": 9.448934052117024e-06, |
| "loss": 0.0102, |
| "step": 19190 |
| }, |
| { |
| "epoch": 1.9668100799016595, |
| "grad_norm": 0.20788440108299255, |
| "learning_rate": 9.433576636128284e-06, |
| "loss": 0.0119, |
| "step": 19200 |
| }, |
| { |
| "epoch": 1.9678344601516082, |
| "grad_norm": 0.15353049337863922, |
| "learning_rate": 9.41822598386881e-06, |
| "loss": 0.0117, |
| "step": 19210 |
| }, |
| { |
| "epoch": 1.968858840401557, |
| "grad_norm": 0.42655694484710693, |
| "learning_rate": 9.402882113991094e-06, |
| "loss": 0.0118, |
| "step": 19220 |
| }, |
| { |
| "epoch": 1.9698832206515058, |
| "grad_norm": 0.38925090432167053, |
| "learning_rate": 9.387545045139381e-06, |
| "loss": 0.0107, |
| "step": 19230 |
| }, |
| { |
| "epoch": 1.9709076009014548, |
| "grad_norm": 0.14344170689582825, |
| "learning_rate": 9.372214795949655e-06, |
| "loss": 0.0127, |
| "step": 19240 |
| }, |
| { |
| "epoch": 1.9719319811514033, |
| "grad_norm": 0.22055144608020782, |
| "learning_rate": 9.356891385049618e-06, |
| "loss": 0.0134, |
| "step": 19250 |
| }, |
| { |
| "epoch": 1.9729563614013523, |
| "grad_norm": 0.5907231569290161, |
| "learning_rate": 9.341574831058664e-06, |
| "loss": 0.0091, |
| "step": 19260 |
| }, |
| { |
| "epoch": 1.9739807416513009, |
| "grad_norm": 0.4834750294685364, |
| "learning_rate": 9.326265152587844e-06, |
| "loss": 0.0112, |
| "step": 19270 |
| }, |
| { |
| "epoch": 1.9750051219012499, |
| "grad_norm": 0.6110846400260925, |
| "learning_rate": 9.310962368239861e-06, |
| "loss": 0.0113, |
| "step": 19280 |
| }, |
| { |
| "epoch": 1.9760295021511984, |
| "grad_norm": 0.3214181959629059, |
| "learning_rate": 9.295666496609036e-06, |
| "loss": 0.0107, |
| "step": 19290 |
| }, |
| { |
| "epoch": 1.9770538824011474, |
| "grad_norm": 0.139206200838089, |
| "learning_rate": 9.280377556281306e-06, |
| "loss": 0.0123, |
| "step": 19300 |
| }, |
| { |
| "epoch": 1.978078262651096, |
| "grad_norm": 0.4006515443325043, |
| "learning_rate": 9.265095565834172e-06, |
| "loss": 0.0111, |
| "step": 19310 |
| }, |
| { |
| "epoch": 1.979102642901045, |
| "grad_norm": 0.1318816989660263, |
| "learning_rate": 9.249820543836688e-06, |
| "loss": 0.0106, |
| "step": 19320 |
| }, |
| { |
| "epoch": 1.9801270231509935, |
| "grad_norm": 0.8921860456466675, |
| "learning_rate": 9.234552508849454e-06, |
| "loss": 0.0109, |
| "step": 19330 |
| }, |
| { |
| "epoch": 1.9811514034009425, |
| "grad_norm": 1.0229899883270264, |
| "learning_rate": 9.21929147942456e-06, |
| "loss": 0.0102, |
| "step": 19340 |
| }, |
| { |
| "epoch": 1.982175783650891, |
| "grad_norm": 0.24664360284805298, |
| "learning_rate": 9.204037474105611e-06, |
| "loss": 0.016, |
| "step": 19350 |
| }, |
| { |
| "epoch": 1.98320016390084, |
| "grad_norm": 0.20093755424022675, |
| "learning_rate": 9.188790511427653e-06, |
| "loss": 0.0103, |
| "step": 19360 |
| }, |
| { |
| "epoch": 1.9842245441507886, |
| "grad_norm": 0.23338110744953156, |
| "learning_rate": 9.173550609917186e-06, |
| "loss": 0.0111, |
| "step": 19370 |
| }, |
| { |
| "epoch": 1.9852489244007376, |
| "grad_norm": 0.1440836787223816, |
| "learning_rate": 9.158317788092124e-06, |
| "loss": 0.0151, |
| "step": 19380 |
| }, |
| { |
| "epoch": 1.9862733046506862, |
| "grad_norm": 0.566990077495575, |
| "learning_rate": 9.143092064461783e-06, |
| "loss": 0.0129, |
| "step": 19390 |
| }, |
| { |
| "epoch": 1.9872976849006352, |
| "grad_norm": 0.3577151298522949, |
| "learning_rate": 9.127873457526855e-06, |
| "loss": 0.011, |
| "step": 19400 |
| }, |
| { |
| "epoch": 1.9883220651505837, |
| "grad_norm": 0.5068030953407288, |
| "learning_rate": 9.112661985779382e-06, |
| "loss": 0.0103, |
| "step": 19410 |
| }, |
| { |
| "epoch": 1.9893464454005327, |
| "grad_norm": 0.359560489654541, |
| "learning_rate": 9.09745766770274e-06, |
| "loss": 0.0109, |
| "step": 19420 |
| }, |
| { |
| "epoch": 1.9903708256504815, |
| "grad_norm": 0.4114885628223419, |
| "learning_rate": 9.082260521771601e-06, |
| "loss": 0.0136, |
| "step": 19430 |
| }, |
| { |
| "epoch": 1.9913952059004303, |
| "grad_norm": 0.3066050708293915, |
| "learning_rate": 9.067070566451936e-06, |
| "loss": 0.0118, |
| "step": 19440 |
| }, |
| { |
| "epoch": 1.992419586150379, |
| "grad_norm": 0.3649365305900574, |
| "learning_rate": 9.051887820200966e-06, |
| "loss": 0.0109, |
| "step": 19450 |
| }, |
| { |
| "epoch": 1.9934439664003278, |
| "grad_norm": 0.14579759538173676, |
| "learning_rate": 9.03671230146717e-06, |
| "loss": 0.0102, |
| "step": 19460 |
| }, |
| { |
| "epoch": 1.9944683466502766, |
| "grad_norm": 0.13388313353061676, |
| "learning_rate": 9.021544028690228e-06, |
| "loss": 0.0116, |
| "step": 19470 |
| }, |
| { |
| "epoch": 1.9954927269002254, |
| "grad_norm": 0.5732167363166809, |
| "learning_rate": 9.006383020301024e-06, |
| "loss": 0.0109, |
| "step": 19480 |
| }, |
| { |
| "epoch": 1.9965171071501742, |
| "grad_norm": 0.2357679307460785, |
| "learning_rate": 8.991229294721613e-06, |
| "loss": 0.0131, |
| "step": 19490 |
| }, |
| { |
| "epoch": 1.997541487400123, |
| "grad_norm": 1.0520175695419312, |
| "learning_rate": 8.976082870365191e-06, |
| "loss": 0.0127, |
| "step": 19500 |
| }, |
| { |
| "epoch": 1.9985658676500717, |
| "grad_norm": 0.2682754397392273, |
| "learning_rate": 8.960943765636105e-06, |
| "loss": 0.0107, |
| "step": 19510 |
| }, |
| { |
| "epoch": 1.9995902479000205, |
| "grad_norm": 0.20680128037929535, |
| "learning_rate": 8.945811998929792e-06, |
| "loss": 0.0119, |
| "step": 19520 |
| }, |
| { |
| "epoch": 2.0006146281499695, |
| "grad_norm": 0.5246221423149109, |
| "learning_rate": 8.93068758863277e-06, |
| "loss": 0.0129, |
| "step": 19530 |
| }, |
| { |
| "epoch": 2.001639008399918, |
| "grad_norm": 0.4808241128921509, |
| "learning_rate": 8.915570553122628e-06, |
| "loss": 0.0115, |
| "step": 19540 |
| }, |
| { |
| "epoch": 2.002663388649867, |
| "grad_norm": 0.8199883103370667, |
| "learning_rate": 8.900460910767982e-06, |
| "loss": 0.0125, |
| "step": 19550 |
| }, |
| { |
| "epoch": 2.0036877688998156, |
| "grad_norm": 0.18950200080871582, |
| "learning_rate": 8.885358679928486e-06, |
| "loss": 0.0128, |
| "step": 19560 |
| }, |
| { |
| "epoch": 2.0047121491497646, |
| "grad_norm": 0.6444079279899597, |
| "learning_rate": 8.870263878954767e-06, |
| "loss": 0.0111, |
| "step": 19570 |
| }, |
| { |
| "epoch": 2.005736529399713, |
| "grad_norm": 0.43850311636924744, |
| "learning_rate": 8.855176526188428e-06, |
| "loss": 0.0106, |
| "step": 19580 |
| }, |
| { |
| "epoch": 2.006760909649662, |
| "grad_norm": 0.27670779824256897, |
| "learning_rate": 8.840096639962033e-06, |
| "loss": 0.0122, |
| "step": 19590 |
| }, |
| { |
| "epoch": 2.0077852898996107, |
| "grad_norm": 0.457205206155777, |
| "learning_rate": 8.825024238599056e-06, |
| "loss": 0.013, |
| "step": 19600 |
| }, |
| { |
| "epoch": 2.0088096701495597, |
| "grad_norm": 0.19198937714099884, |
| "learning_rate": 8.809959340413891e-06, |
| "loss": 0.0113, |
| "step": 19610 |
| }, |
| { |
| "epoch": 2.0098340503995082, |
| "grad_norm": 0.4897865653038025, |
| "learning_rate": 8.79490196371181e-06, |
| "loss": 0.0122, |
| "step": 19620 |
| }, |
| { |
| "epoch": 2.0108584306494572, |
| "grad_norm": 0.21301206946372986, |
| "learning_rate": 8.779852126788949e-06, |
| "loss": 0.0131, |
| "step": 19630 |
| }, |
| { |
| "epoch": 2.011882810899406, |
| "grad_norm": 0.25893697142601013, |
| "learning_rate": 8.764809847932277e-06, |
| "loss": 0.0121, |
| "step": 19640 |
| }, |
| { |
| "epoch": 2.012907191149355, |
| "grad_norm": 0.7507102489471436, |
| "learning_rate": 8.749775145419577e-06, |
| "loss": 0.0116, |
| "step": 19650 |
| }, |
| { |
| "epoch": 2.0139315713993033, |
| "grad_norm": 0.15337583422660828, |
| "learning_rate": 8.734748037519427e-06, |
| "loss": 0.0126, |
| "step": 19660 |
| }, |
| { |
| "epoch": 2.0149559516492523, |
| "grad_norm": 0.23958542943000793, |
| "learning_rate": 8.719728542491195e-06, |
| "loss": 0.0092, |
| "step": 19670 |
| }, |
| { |
| "epoch": 2.015980331899201, |
| "grad_norm": 0.24345184862613678, |
| "learning_rate": 8.704716678584976e-06, |
| "loss": 0.0107, |
| "step": 19680 |
| }, |
| { |
| "epoch": 2.01700471214915, |
| "grad_norm": 0.37324249744415283, |
| "learning_rate": 8.689712464041596e-06, |
| "loss": 0.0132, |
| "step": 19690 |
| }, |
| { |
| "epoch": 2.0180290923990984, |
| "grad_norm": 0.1278141289949417, |
| "learning_rate": 8.674715917092592e-06, |
| "loss": 0.0111, |
| "step": 19700 |
| }, |
| { |
| "epoch": 2.0190534726490474, |
| "grad_norm": 0.22775138914585114, |
| "learning_rate": 8.65972705596018e-06, |
| "loss": 0.01, |
| "step": 19710 |
| }, |
| { |
| "epoch": 2.020077852898996, |
| "grad_norm": 0.23256318271160126, |
| "learning_rate": 8.64474589885725e-06, |
| "loss": 0.0104, |
| "step": 19720 |
| }, |
| { |
| "epoch": 2.021102233148945, |
| "grad_norm": 0.15660132467746735, |
| "learning_rate": 8.629772463987315e-06, |
| "loss": 0.0112, |
| "step": 19730 |
| }, |
| { |
| "epoch": 2.0221266133988935, |
| "grad_norm": 0.1920231580734253, |
| "learning_rate": 8.614806769544507e-06, |
| "loss": 0.011, |
| "step": 19740 |
| }, |
| { |
| "epoch": 2.0231509936488425, |
| "grad_norm": 0.9880847930908203, |
| "learning_rate": 8.59984883371356e-06, |
| "loss": 0.0127, |
| "step": 19750 |
| }, |
| { |
| "epoch": 2.024175373898791, |
| "grad_norm": 0.2476850152015686, |
| "learning_rate": 8.584898674669772e-06, |
| "loss": 0.0108, |
| "step": 19760 |
| }, |
| { |
| "epoch": 2.02519975414874, |
| "grad_norm": 0.10527423769235611, |
| "learning_rate": 8.569956310579001e-06, |
| "loss": 0.0099, |
| "step": 19770 |
| }, |
| { |
| "epoch": 2.0262241343986886, |
| "grad_norm": 0.45273977518081665, |
| "learning_rate": 8.555021759597627e-06, |
| "loss": 0.0127, |
| "step": 19780 |
| }, |
| { |
| "epoch": 2.0272485146486376, |
| "grad_norm": 0.3690817952156067, |
| "learning_rate": 8.54009503987253e-06, |
| "loss": 0.0118, |
| "step": 19790 |
| }, |
| { |
| "epoch": 2.028272894898586, |
| "grad_norm": 0.4257254898548126, |
| "learning_rate": 8.525176169541095e-06, |
| "loss": 0.0115, |
| "step": 19800 |
| }, |
| { |
| "epoch": 2.029297275148535, |
| "grad_norm": 0.16916978359222412, |
| "learning_rate": 8.510265166731151e-06, |
| "loss": 0.0152, |
| "step": 19810 |
| }, |
| { |
| "epoch": 2.0303216553984837, |
| "grad_norm": 0.23369191586971283, |
| "learning_rate": 8.495362049560964e-06, |
| "loss": 0.0087, |
| "step": 19820 |
| }, |
| { |
| "epoch": 2.0313460356484327, |
| "grad_norm": 0.16445417702198029, |
| "learning_rate": 8.480466836139241e-06, |
| "loss": 0.0128, |
| "step": 19830 |
| }, |
| { |
| "epoch": 2.0323704158983813, |
| "grad_norm": 0.28569164872169495, |
| "learning_rate": 8.465579544565066e-06, |
| "loss": 0.0116, |
| "step": 19840 |
| }, |
| { |
| "epoch": 2.0333947961483303, |
| "grad_norm": 0.634433388710022, |
| "learning_rate": 8.450700192927897e-06, |
| "loss": 0.0123, |
| "step": 19850 |
| }, |
| { |
| "epoch": 2.034419176398279, |
| "grad_norm": 0.35075703263282776, |
| "learning_rate": 8.435828799307551e-06, |
| "loss": 0.0097, |
| "step": 19860 |
| }, |
| { |
| "epoch": 2.035443556648228, |
| "grad_norm": 0.15156176686286926, |
| "learning_rate": 8.420965381774172e-06, |
| "loss": 0.0105, |
| "step": 19870 |
| }, |
| { |
| "epoch": 2.0364679368981764, |
| "grad_norm": 0.2092888504266739, |
| "learning_rate": 8.406109958388222e-06, |
| "loss": 0.0115, |
| "step": 19880 |
| }, |
| { |
| "epoch": 2.0374923171481254, |
| "grad_norm": 0.17040668427944183, |
| "learning_rate": 8.391262547200434e-06, |
| "loss": 0.009, |
| "step": 19890 |
| }, |
| { |
| "epoch": 2.038516697398074, |
| "grad_norm": 0.6679179668426514, |
| "learning_rate": 8.376423166251816e-06, |
| "loss": 0.0105, |
| "step": 19900 |
| }, |
| { |
| "epoch": 2.039541077648023, |
| "grad_norm": 0.605216383934021, |
| "learning_rate": 8.361591833573612e-06, |
| "loss": 0.0085, |
| "step": 19910 |
| }, |
| { |
| "epoch": 2.040565457897972, |
| "grad_norm": 0.257565975189209, |
| "learning_rate": 8.346768567187287e-06, |
| "loss": 0.0112, |
| "step": 19920 |
| }, |
| { |
| "epoch": 2.0415898381479205, |
| "grad_norm": 0.9964429140090942, |
| "learning_rate": 8.331953385104516e-06, |
| "loss": 0.011, |
| "step": 19930 |
| }, |
| { |
| "epoch": 2.0426142183978695, |
| "grad_norm": 0.5869243144989014, |
| "learning_rate": 8.317146305327134e-06, |
| "loss": 0.0119, |
| "step": 19940 |
| }, |
| { |
| "epoch": 2.043638598647818, |
| "grad_norm": 0.1279764324426651, |
| "learning_rate": 8.302347345847146e-06, |
| "loss": 0.0107, |
| "step": 19950 |
| }, |
| { |
| "epoch": 2.044662978897767, |
| "grad_norm": 0.3344166874885559, |
| "learning_rate": 8.287556524646683e-06, |
| "loss": 0.0117, |
| "step": 19960 |
| }, |
| { |
| "epoch": 2.0456873591477156, |
| "grad_norm": 0.46051591634750366, |
| "learning_rate": 8.272773859697978e-06, |
| "loss": 0.0113, |
| "step": 19970 |
| }, |
| { |
| "epoch": 2.0467117393976646, |
| "grad_norm": 0.2736915647983551, |
| "learning_rate": 8.257999368963374e-06, |
| "loss": 0.0093, |
| "step": 19980 |
| }, |
| { |
| "epoch": 2.047736119647613, |
| "grad_norm": 0.13827525079250336, |
| "learning_rate": 8.243233070395266e-06, |
| "loss": 0.0138, |
| "step": 19990 |
| }, |
| { |
| "epoch": 2.048760499897562, |
| "grad_norm": 0.24435526132583618, |
| "learning_rate": 8.2284749819361e-06, |
| "loss": 0.0116, |
| "step": 20000 |
| } |
| ], |
| "logging_steps": 10, |
| "max_steps": 30000, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 4, |
| "save_steps": 1000, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": false |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 0.0, |
| "train_batch_size": 60, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|