| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 41.328125, | |
| "eval_steps": 500, | |
| "global_step": 2645, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0, | |
| "eval_runtime": 2.9591, | |
| "eval_samples_per_second": 0.338, | |
| "eval_steps_per_second": 0.338, | |
| "step": 0 | |
| }, | |
| { | |
| "epoch": 0.15625, | |
| "grad_norm": 8.098007202148438, | |
| "learning_rate": 4.000000000000001e-06, | |
| "loss": 8.5142, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.3125, | |
| "grad_norm": 6.856651306152344, | |
| "learning_rate": 9e-06, | |
| "loss": 7.9708, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.46875, | |
| "grad_norm": 5.80125093460083, | |
| "learning_rate": 1.4000000000000001e-05, | |
| "loss": 8.01, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.625, | |
| "grad_norm": 5.687861442565918, | |
| "learning_rate": 1.9e-05, | |
| "loss": 7.6212, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.78125, | |
| "grad_norm": 6.117136478424072, | |
| "learning_rate": 2.4e-05, | |
| "loss": 7.3311, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.9375, | |
| "grad_norm": 5.861077785491943, | |
| "learning_rate": 2.9e-05, | |
| "loss": 7.5553, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 1.09375, | |
| "grad_norm": 6.021880149841309, | |
| "learning_rate": 3.4000000000000007e-05, | |
| "loss": 7.05, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 1.25, | |
| "grad_norm": 7.620683670043945, | |
| "learning_rate": 3.9000000000000006e-05, | |
| "loss": 6.7186, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.40625, | |
| "grad_norm": 7.098018169403076, | |
| "learning_rate": 4.4000000000000006e-05, | |
| "loss": 6.5668, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.5625, | |
| "grad_norm": 7.157684326171875, | |
| "learning_rate": 4.9e-05, | |
| "loss": 6.6148, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.71875, | |
| "grad_norm": 7.009696006774902, | |
| "learning_rate": 4.9936507936507936e-05, | |
| "loss": 6.5017, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.875, | |
| "grad_norm": 7.6622467041015625, | |
| "learning_rate": 4.985714285714286e-05, | |
| "loss": 6.53, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 2.03125, | |
| "grad_norm": 8.221333503723145, | |
| "learning_rate": 4.977777777777778e-05, | |
| "loss": 6.306, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 2.1875, | |
| "grad_norm": 8.832132339477539, | |
| "learning_rate": 4.96984126984127e-05, | |
| "loss": 4.8787, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 2.34375, | |
| "grad_norm": 9.245682716369629, | |
| "learning_rate": 4.961904761904762e-05, | |
| "loss": 4.7796, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 2.5, | |
| "grad_norm": 9.249746322631836, | |
| "learning_rate": 4.953968253968254e-05, | |
| "loss": 4.79, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 2.65625, | |
| "grad_norm": 8.098647117614746, | |
| "learning_rate": 4.9460317460317465e-05, | |
| "loss": 4.8157, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 2.8125, | |
| "grad_norm": 9.594801902770996, | |
| "learning_rate": 4.9380952380952386e-05, | |
| "loss": 4.5135, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 2.96875, | |
| "grad_norm": 9.500739097595215, | |
| "learning_rate": 4.930158730158731e-05, | |
| "loss": 4.9346, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 3.125, | |
| "grad_norm": 10.759628295898438, | |
| "learning_rate": 4.922222222222222e-05, | |
| "loss": 3.4195, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 3.28125, | |
| "grad_norm": 12.300085067749023, | |
| "learning_rate": 4.9142857142857144e-05, | |
| "loss": 3.136, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 3.4375, | |
| "grad_norm": 9.825932502746582, | |
| "learning_rate": 4.9063492063492065e-05, | |
| "loss": 3.0981, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 3.59375, | |
| "grad_norm": 10.764175415039062, | |
| "learning_rate": 4.898412698412699e-05, | |
| "loss": 3.0372, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 3.75, | |
| "grad_norm": 10.330010414123535, | |
| "learning_rate": 4.890476190476191e-05, | |
| "loss": 3.0934, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 3.90625, | |
| "grad_norm": 10.405282020568848, | |
| "learning_rate": 4.882539682539683e-05, | |
| "loss": 3.135, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 4.0625, | |
| "grad_norm": 9.842012405395508, | |
| "learning_rate": 4.874603174603175e-05, | |
| "loss": 2.7457, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 4.21875, | |
| "grad_norm": 10.659170150756836, | |
| "learning_rate": 4.866666666666667e-05, | |
| "loss": 1.981, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 4.375, | |
| "grad_norm": 9.820462226867676, | |
| "learning_rate": 4.858730158730159e-05, | |
| "loss": 2.0881, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 4.53125, | |
| "grad_norm": 9.645284652709961, | |
| "learning_rate": 4.850793650793651e-05, | |
| "loss": 1.9438, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 4.6875, | |
| "grad_norm": 9.881282806396484, | |
| "learning_rate": 4.842857142857143e-05, | |
| "loss": 1.8556, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 4.84375, | |
| "grad_norm": 10.5478515625, | |
| "learning_rate": 4.834920634920635e-05, | |
| "loss": 2.0494, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 19.77553939819336, | |
| "learning_rate": 4.8269841269841274e-05, | |
| "loss": 2.0695, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 5.15625, | |
| "grad_norm": 8.124504089355469, | |
| "learning_rate": 4.819047619047619e-05, | |
| "loss": 1.1424, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 5.3125, | |
| "grad_norm": 9.63291072845459, | |
| "learning_rate": 4.811111111111111e-05, | |
| "loss": 1.1688, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 5.46875, | |
| "grad_norm": 8.620952606201172, | |
| "learning_rate": 4.803174603174603e-05, | |
| "loss": 1.2346, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 5.625, | |
| "grad_norm": 8.060315132141113, | |
| "learning_rate": 4.795238095238095e-05, | |
| "loss": 1.176, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 5.78125, | |
| "grad_norm": 7.583358287811279, | |
| "learning_rate": 4.7873015873015874e-05, | |
| "loss": 1.2118, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 5.9375, | |
| "grad_norm": 10.535112380981445, | |
| "learning_rate": 4.7793650793650796e-05, | |
| "loss": 1.2446, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 6.09375, | |
| "grad_norm": 7.170854568481445, | |
| "learning_rate": 4.771428571428572e-05, | |
| "loss": 1.0445, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 6.25, | |
| "grad_norm": 6.0907392501831055, | |
| "learning_rate": 4.763492063492064e-05, | |
| "loss": 0.7688, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 6.40625, | |
| "grad_norm": 7.953549385070801, | |
| "learning_rate": 4.755555555555556e-05, | |
| "loss": 0.7333, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 6.5625, | |
| "grad_norm": 10.276406288146973, | |
| "learning_rate": 4.747619047619048e-05, | |
| "loss": 0.8645, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 6.71875, | |
| "grad_norm": 7.596552848815918, | |
| "learning_rate": 4.73968253968254e-05, | |
| "loss": 0.7799, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 6.875, | |
| "grad_norm": 6.478920936584473, | |
| "learning_rate": 4.7317460317460325e-05, | |
| "loss": 0.8262, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 7.03125, | |
| "grad_norm": 4.837390422821045, | |
| "learning_rate": 4.723809523809524e-05, | |
| "loss": 0.7218, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 7.1875, | |
| "grad_norm": 5.8777289390563965, | |
| "learning_rate": 4.715873015873016e-05, | |
| "loss": 0.5238, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 7.34375, | |
| "grad_norm": 4.988452434539795, | |
| "learning_rate": 4.707936507936508e-05, | |
| "loss": 0.5285, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 7.5, | |
| "grad_norm": 5.605051517486572, | |
| "learning_rate": 4.7e-05, | |
| "loss": 0.5368, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 7.65625, | |
| "grad_norm": 7.234042644500732, | |
| "learning_rate": 4.692063492063492e-05, | |
| "loss": 0.5803, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 7.8125, | |
| "grad_norm": 6.273004055023193, | |
| "learning_rate": 4.684126984126984e-05, | |
| "loss": 0.519, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 7.8125, | |
| "eval_runtime": 2.3062, | |
| "eval_samples_per_second": 0.434, | |
| "eval_steps_per_second": 0.434, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 7.96875, | |
| "grad_norm": 4.803182125091553, | |
| "learning_rate": 4.676190476190476e-05, | |
| "loss": 0.5611, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 8.125, | |
| "grad_norm": 4.977607727050781, | |
| "learning_rate": 4.668253968253968e-05, | |
| "loss": 0.3616, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 8.28125, | |
| "grad_norm": 4.4240336418151855, | |
| "learning_rate": 4.6603174603174605e-05, | |
| "loss": 0.4303, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 8.4375, | |
| "grad_norm": 5.406126022338867, | |
| "learning_rate": 4.6523809523809526e-05, | |
| "loss": 0.384, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 8.59375, | |
| "grad_norm": 4.34033203125, | |
| "learning_rate": 4.644444444444445e-05, | |
| "loss": 0.3784, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 8.75, | |
| "grad_norm": 5.256556034088135, | |
| "learning_rate": 4.636507936507937e-05, | |
| "loss": 0.4345, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 8.90625, | |
| "grad_norm": 5.054710388183594, | |
| "learning_rate": 4.628571428571429e-05, | |
| "loss": 0.4257, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 9.0625, | |
| "grad_norm": 5.220774173736572, | |
| "learning_rate": 4.6206349206349205e-05, | |
| "loss": 0.3146, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 9.21875, | |
| "grad_norm": 4.420045375823975, | |
| "learning_rate": 4.612698412698413e-05, | |
| "loss": 0.298, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 9.375, | |
| "grad_norm": 4.474491596221924, | |
| "learning_rate": 4.604761904761905e-05, | |
| "loss": 0.3023, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 9.53125, | |
| "grad_norm": 4.131107807159424, | |
| "learning_rate": 4.596825396825397e-05, | |
| "loss": 0.2846, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 9.6875, | |
| "grad_norm": 4.535623550415039, | |
| "learning_rate": 4.588888888888889e-05, | |
| "loss": 0.3244, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 9.84375, | |
| "grad_norm": 4.365594387054443, | |
| "learning_rate": 4.580952380952381e-05, | |
| "loss": 0.2925, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 10.0, | |
| "grad_norm": 7.354922294616699, | |
| "learning_rate": 4.5730158730158734e-05, | |
| "loss": 0.3234, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 10.15625, | |
| "grad_norm": 3.916835308074951, | |
| "learning_rate": 4.5650793650793656e-05, | |
| "loss": 0.236, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 10.3125, | |
| "grad_norm": 4.322419166564941, | |
| "learning_rate": 4.557142857142858e-05, | |
| "loss": 0.2353, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 10.46875, | |
| "grad_norm": 5.4846038818359375, | |
| "learning_rate": 4.54920634920635e-05, | |
| "loss": 0.2422, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 10.625, | |
| "grad_norm": 4.341657638549805, | |
| "learning_rate": 4.5412698412698414e-05, | |
| "loss": 0.2592, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 10.78125, | |
| "grad_norm": 4.786070346832275, | |
| "learning_rate": 4.5333333333333335e-05, | |
| "loss": 0.2501, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 10.9375, | |
| "grad_norm": 3.9264745712280273, | |
| "learning_rate": 4.525396825396826e-05, | |
| "loss": 0.2683, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 11.09375, | |
| "grad_norm": 3.851856231689453, | |
| "learning_rate": 4.517460317460318e-05, | |
| "loss": 0.1978, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 11.25, | |
| "grad_norm": 4.643581390380859, | |
| "learning_rate": 4.509523809523809e-05, | |
| "loss": 0.1841, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 11.40625, | |
| "grad_norm": 4.105888843536377, | |
| "learning_rate": 4.5015873015873014e-05, | |
| "loss": 0.1976, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 11.5625, | |
| "grad_norm": 3.809528350830078, | |
| "learning_rate": 4.4936507936507936e-05, | |
| "loss": 0.1896, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 11.71875, | |
| "grad_norm": 4.457248687744141, | |
| "learning_rate": 4.485714285714286e-05, | |
| "loss": 0.2125, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 11.875, | |
| "grad_norm": 3.7699806690216064, | |
| "learning_rate": 4.477777777777778e-05, | |
| "loss": 0.2212, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 12.03125, | |
| "grad_norm": 4.643237590789795, | |
| "learning_rate": 4.46984126984127e-05, | |
| "loss": 0.2235, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 12.1875, | |
| "grad_norm": 4.064443588256836, | |
| "learning_rate": 4.461904761904762e-05, | |
| "loss": 0.1503, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 12.34375, | |
| "grad_norm": 4.605109691619873, | |
| "learning_rate": 4.4539682539682543e-05, | |
| "loss": 0.172, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 12.5, | |
| "grad_norm": 3.8005576133728027, | |
| "learning_rate": 4.4460317460317465e-05, | |
| "loss": 0.1713, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 12.65625, | |
| "grad_norm": 4.829875469207764, | |
| "learning_rate": 4.4380952380952386e-05, | |
| "loss": 0.1783, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 12.8125, | |
| "grad_norm": 4.428802490234375, | |
| "learning_rate": 4.430158730158731e-05, | |
| "loss": 0.1742, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 12.96875, | |
| "grad_norm": 5.068906307220459, | |
| "learning_rate": 4.422222222222222e-05, | |
| "loss": 0.1797, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 13.125, | |
| "grad_norm": 3.3268795013427734, | |
| "learning_rate": 4.4142857142857144e-05, | |
| "loss": 0.1389, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 13.28125, | |
| "grad_norm": 3.312582015991211, | |
| "learning_rate": 4.4063492063492066e-05, | |
| "loss": 0.1191, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 13.4375, | |
| "grad_norm": 3.7299697399139404, | |
| "learning_rate": 4.398412698412699e-05, | |
| "loss": 0.1432, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 13.59375, | |
| "grad_norm": 4.627827167510986, | |
| "learning_rate": 4.39047619047619e-05, | |
| "loss": 0.1672, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 13.75, | |
| "grad_norm": 4.645421981811523, | |
| "learning_rate": 4.3825396825396823e-05, | |
| "loss": 0.1503, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 13.90625, | |
| "grad_norm": 5.321810245513916, | |
| "learning_rate": 4.3746031746031745e-05, | |
| "loss": 0.1486, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 14.0625, | |
| "grad_norm": 2.487105369567871, | |
| "learning_rate": 4.3666666666666666e-05, | |
| "loss": 0.1193, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 14.21875, | |
| "grad_norm": 4.098151683807373, | |
| "learning_rate": 4.358730158730159e-05, | |
| "loss": 0.1182, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 14.375, | |
| "grad_norm": 3.2538163661956787, | |
| "learning_rate": 4.350793650793651e-05, | |
| "loss": 0.1145, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 14.53125, | |
| "grad_norm": 3.7881181240081787, | |
| "learning_rate": 4.342857142857143e-05, | |
| "loss": 0.1255, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 14.6875, | |
| "grad_norm": 2.861689567565918, | |
| "learning_rate": 4.334920634920635e-05, | |
| "loss": 0.144, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 14.84375, | |
| "grad_norm": 3.1621310710906982, | |
| "learning_rate": 4.3269841269841274e-05, | |
| "loss": 0.1293, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 15.0, | |
| "grad_norm": 11.439355850219727, | |
| "learning_rate": 4.3190476190476195e-05, | |
| "loss": 0.1553, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 15.15625, | |
| "grad_norm": 3.6791841983795166, | |
| "learning_rate": 4.311111111111111e-05, | |
| "loss": 0.0949, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 15.3125, | |
| "grad_norm": 2.71702241897583, | |
| "learning_rate": 4.303174603174603e-05, | |
| "loss": 0.1125, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 15.46875, | |
| "grad_norm": 2.713571786880493, | |
| "learning_rate": 4.295238095238095e-05, | |
| "loss": 0.137, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 15.625, | |
| "grad_norm": 2.975768804550171, | |
| "learning_rate": 4.2873015873015875e-05, | |
| "loss": 0.108, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 15.625, | |
| "eval_runtime": 2.2622, | |
| "eval_samples_per_second": 0.442, | |
| "eval_steps_per_second": 0.442, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 15.78125, | |
| "grad_norm": 3.5744853019714355, | |
| "learning_rate": 4.2793650793650796e-05, | |
| "loss": 0.1207, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 15.9375, | |
| "grad_norm": 1.8845595121383667, | |
| "learning_rate": 4.271428571428572e-05, | |
| "loss": 0.1013, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 16.09375, | |
| "grad_norm": 4.227961540222168, | |
| "learning_rate": 4.263492063492064e-05, | |
| "loss": 0.112, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 16.25, | |
| "grad_norm": 3.289071798324585, | |
| "learning_rate": 4.255555555555556e-05, | |
| "loss": 0.1054, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 16.40625, | |
| "grad_norm": 1.9135278463363647, | |
| "learning_rate": 4.247619047619048e-05, | |
| "loss": 0.0933, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 16.5625, | |
| "grad_norm": 2.586151599884033, | |
| "learning_rate": 4.2396825396825404e-05, | |
| "loss": 0.1205, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 16.71875, | |
| "grad_norm": 3.5681819915771484, | |
| "learning_rate": 4.231746031746032e-05, | |
| "loss": 0.1015, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 16.875, | |
| "grad_norm": 3.7675039768218994, | |
| "learning_rate": 4.223809523809524e-05, | |
| "loss": 0.128, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 17.03125, | |
| "grad_norm": 2.631277322769165, | |
| "learning_rate": 4.215873015873016e-05, | |
| "loss": 0.107, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 17.1875, | |
| "grad_norm": 3.59251070022583, | |
| "learning_rate": 4.2079365079365076e-05, | |
| "loss": 0.0918, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 17.34375, | |
| "grad_norm": 3.785374641418457, | |
| "learning_rate": 4.2e-05, | |
| "loss": 0.1031, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 17.5, | |
| "grad_norm": 3.6201488971710205, | |
| "learning_rate": 4.192063492063492e-05, | |
| "loss": 0.0965, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 17.65625, | |
| "grad_norm": 3.608516216278076, | |
| "learning_rate": 4.184126984126984e-05, | |
| "loss": 0.1046, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 17.8125, | |
| "grad_norm": 3.3631608486175537, | |
| "learning_rate": 4.176190476190476e-05, | |
| "loss": 0.1047, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 17.96875, | |
| "grad_norm": 2.2599599361419678, | |
| "learning_rate": 4.1682539682539684e-05, | |
| "loss": 0.0976, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 18.125, | |
| "grad_norm": 3.3925111293792725, | |
| "learning_rate": 4.1603174603174605e-05, | |
| "loss": 0.081, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 18.28125, | |
| "grad_norm": 2.5443062782287598, | |
| "learning_rate": 4.152380952380953e-05, | |
| "loss": 0.0917, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 18.4375, | |
| "grad_norm": 4.180214881896973, | |
| "learning_rate": 4.144444444444445e-05, | |
| "loss": 0.0913, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 18.59375, | |
| "grad_norm": 2.3229408264160156, | |
| "learning_rate": 4.136507936507937e-05, | |
| "loss": 0.081, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 18.75, | |
| "grad_norm": 3.885908603668213, | |
| "learning_rate": 4.128571428571429e-05, | |
| "loss": 0.1078, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 18.90625, | |
| "grad_norm": 3.1222493648529053, | |
| "learning_rate": 4.120634920634921e-05, | |
| "loss": 0.1177, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 19.0625, | |
| "grad_norm": 3.3118879795074463, | |
| "learning_rate": 4.112698412698413e-05, | |
| "loss": 0.0775, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 19.21875, | |
| "grad_norm": 1.8756282329559326, | |
| "learning_rate": 4.104761904761905e-05, | |
| "loss": 0.0906, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 19.375, | |
| "grad_norm": 3.6675071716308594, | |
| "learning_rate": 4.096825396825397e-05, | |
| "loss": 0.0938, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 19.53125, | |
| "grad_norm": 2.8696744441986084, | |
| "learning_rate": 4.088888888888889e-05, | |
| "loss": 0.0821, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 19.6875, | |
| "grad_norm": 3.246438980102539, | |
| "learning_rate": 4.0809523809523813e-05, | |
| "loss": 0.1216, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 19.84375, | |
| "grad_norm": 2.579602003097534, | |
| "learning_rate": 4.073015873015873e-05, | |
| "loss": 0.0705, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 20.0, | |
| "grad_norm": 14.062989234924316, | |
| "learning_rate": 4.065079365079365e-05, | |
| "loss": 0.1125, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 20.15625, | |
| "grad_norm": 2.5907585620880127, | |
| "learning_rate": 4.057142857142857e-05, | |
| "loss": 0.0764, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 20.3125, | |
| "grad_norm": 4.850607872009277, | |
| "learning_rate": 4.049206349206349e-05, | |
| "loss": 0.0882, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 20.46875, | |
| "grad_norm": 2.3619115352630615, | |
| "learning_rate": 4.0412698412698414e-05, | |
| "loss": 0.0833, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 20.625, | |
| "grad_norm": 3.1803815364837646, | |
| "learning_rate": 4.0333333333333336e-05, | |
| "loss": 0.0894, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 20.78125, | |
| "grad_norm": 15.570907592773438, | |
| "learning_rate": 4.025396825396826e-05, | |
| "loss": 0.2225, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 20.9375, | |
| "grad_norm": 2.721440076828003, | |
| "learning_rate": 4.018253968253968e-05, | |
| "loss": 0.1233, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 21.09375, | |
| "grad_norm": 1.798134446144104, | |
| "learning_rate": 4.01031746031746e-05, | |
| "loss": 0.067, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 21.25, | |
| "grad_norm": 3.387782335281372, | |
| "learning_rate": 4.0023809523809524e-05, | |
| "loss": 0.0692, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 21.40625, | |
| "grad_norm": 2.723196029663086, | |
| "learning_rate": 3.9944444444444446e-05, | |
| "loss": 0.0683, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 21.5625, | |
| "grad_norm": 3.666444778442383, | |
| "learning_rate": 3.986507936507937e-05, | |
| "loss": 0.0883, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 21.71875, | |
| "grad_norm": 1.9571526050567627, | |
| "learning_rate": 3.978571428571429e-05, | |
| "loss": 0.0802, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 21.875, | |
| "grad_norm": 2.5271904468536377, | |
| "learning_rate": 3.970634920634921e-05, | |
| "loss": 0.0737, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 22.03125, | |
| "grad_norm": 3.5426900386810303, | |
| "learning_rate": 3.962698412698413e-05, | |
| "loss": 0.0838, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 22.1875, | |
| "grad_norm": 1.7901580333709717, | |
| "learning_rate": 3.954761904761905e-05, | |
| "loss": 0.0671, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 22.34375, | |
| "grad_norm": 3.4260764122009277, | |
| "learning_rate": 3.946825396825397e-05, | |
| "loss": 0.0775, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 22.5, | |
| "grad_norm": 2.499107837677002, | |
| "learning_rate": 3.938888888888889e-05, | |
| "loss": 0.0826, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 22.65625, | |
| "grad_norm": 2.7331862449645996, | |
| "learning_rate": 3.930952380952381e-05, | |
| "loss": 0.0771, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 22.8125, | |
| "grad_norm": 3.2004685401916504, | |
| "learning_rate": 3.923015873015873e-05, | |
| "loss": 0.088, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 22.96875, | |
| "grad_norm": 3.175179958343506, | |
| "learning_rate": 3.9150793650793654e-05, | |
| "loss": 0.0861, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 23.125, | |
| "grad_norm": 2.032646417617798, | |
| "learning_rate": 3.9071428571428575e-05, | |
| "loss": 0.0682, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 23.28125, | |
| "grad_norm": 1.5635634660720825, | |
| "learning_rate": 3.89920634920635e-05, | |
| "loss": 0.0833, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 23.4375, | |
| "grad_norm": 1.8121321201324463, | |
| "learning_rate": 3.891269841269842e-05, | |
| "loss": 0.0965, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 23.4375, | |
| "eval_runtime": 2.0794, | |
| "eval_samples_per_second": 0.481, | |
| "eval_steps_per_second": 0.481, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 23.59375, | |
| "grad_norm": 1.8793394565582275, | |
| "learning_rate": 3.883333333333333e-05, | |
| "loss": 0.0887, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 23.75, | |
| "grad_norm": 2.1231632232666016, | |
| "learning_rate": 3.8753968253968255e-05, | |
| "loss": 0.0879, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 23.90625, | |
| "grad_norm": 3.5764803886413574, | |
| "learning_rate": 3.8674603174603176e-05, | |
| "loss": 0.0861, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 24.0625, | |
| "grad_norm": 2.105710029602051, | |
| "learning_rate": 3.85952380952381e-05, | |
| "loss": 0.07, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 24.21875, | |
| "grad_norm": 2.8722870349884033, | |
| "learning_rate": 3.851587301587302e-05, | |
| "loss": 0.076, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 24.375, | |
| "grad_norm": 2.6126277446746826, | |
| "learning_rate": 3.843650793650794e-05, | |
| "loss": 0.0983, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 24.53125, | |
| "grad_norm": 2.2635769844055176, | |
| "learning_rate": 3.8357142857142855e-05, | |
| "loss": 0.0864, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 24.6875, | |
| "grad_norm": 2.11098313331604, | |
| "learning_rate": 3.827777777777778e-05, | |
| "loss": 0.0923, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 24.84375, | |
| "grad_norm": 1.6163533926010132, | |
| "learning_rate": 3.81984126984127e-05, | |
| "loss": 0.0711, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 25.0, | |
| "grad_norm": 2.54638671875, | |
| "learning_rate": 3.811904761904762e-05, | |
| "loss": 0.0705, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 25.15625, | |
| "grad_norm": 2.592470407485962, | |
| "learning_rate": 3.803968253968254e-05, | |
| "loss": 0.0765, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 25.3125, | |
| "grad_norm": 2.1981208324432373, | |
| "learning_rate": 3.796031746031746e-05, | |
| "loss": 0.0856, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 25.46875, | |
| "grad_norm": 1.384098768234253, | |
| "learning_rate": 3.7880952380952384e-05, | |
| "loss": 0.0951, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 25.625, | |
| "grad_norm": 1.502350091934204, | |
| "learning_rate": 3.7801587301587306e-05, | |
| "loss": 0.0858, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 25.78125, | |
| "grad_norm": 1.4763522148132324, | |
| "learning_rate": 3.772222222222223e-05, | |
| "loss": 0.0647, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 25.9375, | |
| "grad_norm": 1.7555052042007446, | |
| "learning_rate": 3.764285714285715e-05, | |
| "loss": 0.0874, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 26.09375, | |
| "grad_norm": 1.0187015533447266, | |
| "learning_rate": 3.756349206349207e-05, | |
| "loss": 0.0635, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 26.25, | |
| "grad_norm": 1.7138936519622803, | |
| "learning_rate": 3.7484126984126985e-05, | |
| "loss": 0.0772, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 26.40625, | |
| "grad_norm": 2.3353724479675293, | |
| "learning_rate": 3.7404761904761907e-05, | |
| "loss": 0.0892, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 26.5625, | |
| "grad_norm": 2.6141700744628906, | |
| "learning_rate": 3.732539682539682e-05, | |
| "loss": 0.0897, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 26.71875, | |
| "grad_norm": 1.7785848379135132, | |
| "learning_rate": 3.724603174603174e-05, | |
| "loss": 0.0899, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 26.875, | |
| "grad_norm": 2.6693010330200195, | |
| "learning_rate": 3.7166666666666664e-05, | |
| "loss": 0.0901, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 27.03125, | |
| "grad_norm": 2.059981346130371, | |
| "learning_rate": 3.7087301587301586e-05, | |
| "loss": 0.0822, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 27.1875, | |
| "grad_norm": 1.6238901615142822, | |
| "learning_rate": 3.700793650793651e-05, | |
| "loss": 0.0777, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 27.34375, | |
| "grad_norm": 2.782425880432129, | |
| "learning_rate": 3.692857142857143e-05, | |
| "loss": 0.0782, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 27.5, | |
| "grad_norm": 1.8468166589736938, | |
| "learning_rate": 3.684920634920635e-05, | |
| "loss": 0.0835, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 27.65625, | |
| "grad_norm": 1.3156135082244873, | |
| "learning_rate": 3.676984126984127e-05, | |
| "loss": 0.1072, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 27.8125, | |
| "grad_norm": 3.260084390640259, | |
| "learning_rate": 3.669047619047619e-05, | |
| "loss": 0.0939, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 27.96875, | |
| "grad_norm": 2.518204689025879, | |
| "learning_rate": 3.6611111111111115e-05, | |
| "loss": 0.0884, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 28.125, | |
| "grad_norm": 2.598057985305786, | |
| "learning_rate": 3.6531746031746036e-05, | |
| "loss": 0.0683, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 28.28125, | |
| "grad_norm": 1.8533433675765991, | |
| "learning_rate": 3.645238095238096e-05, | |
| "loss": 0.0833, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 28.4375, | |
| "grad_norm": 1.2828975915908813, | |
| "learning_rate": 3.637301587301587e-05, | |
| "loss": 0.0832, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 28.59375, | |
| "grad_norm": 1.7714905738830566, | |
| "learning_rate": 3.6293650793650794e-05, | |
| "loss": 0.0901, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 28.75, | |
| "grad_norm": 2.098923921585083, | |
| "learning_rate": 3.6214285714285716e-05, | |
| "loss": 0.0954, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 28.90625, | |
| "grad_norm": 2.298226833343506, | |
| "learning_rate": 3.613492063492064e-05, | |
| "loss": 0.0888, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 29.0625, | |
| "grad_norm": 1.5519624948501587, | |
| "learning_rate": 3.605555555555556e-05, | |
| "loss": 0.0664, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 29.21875, | |
| "grad_norm": 2.015573501586914, | |
| "learning_rate": 3.597619047619048e-05, | |
| "loss": 0.0637, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 29.375, | |
| "grad_norm": 1.925529956817627, | |
| "learning_rate": 3.58968253968254e-05, | |
| "loss": 0.0803, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 29.53125, | |
| "grad_norm": 2.4342522621154785, | |
| "learning_rate": 3.581746031746032e-05, | |
| "loss": 0.0911, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 29.6875, | |
| "grad_norm": 1.8124195337295532, | |
| "learning_rate": 3.573809523809524e-05, | |
| "loss": 0.0803, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 29.84375, | |
| "grad_norm": 2.3409860134124756, | |
| "learning_rate": 3.565873015873016e-05, | |
| "loss": 0.0943, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 30.0, | |
| "grad_norm": 1.2545162439346313, | |
| "learning_rate": 3.557936507936508e-05, | |
| "loss": 0.0853, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 30.15625, | |
| "grad_norm": 1.9356091022491455, | |
| "learning_rate": 3.55e-05, | |
| "loss": 0.0585, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 30.3125, | |
| "grad_norm": 1.6294385194778442, | |
| "learning_rate": 3.5420634920634924e-05, | |
| "loss": 0.0876, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 30.46875, | |
| "grad_norm": 2.081688165664673, | |
| "learning_rate": 3.534126984126984e-05, | |
| "loss": 0.0827, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 30.625, | |
| "grad_norm": 1.7468382120132446, | |
| "learning_rate": 3.526190476190476e-05, | |
| "loss": 0.0732, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 30.78125, | |
| "grad_norm": 2.1742124557495117, | |
| "learning_rate": 3.518253968253968e-05, | |
| "loss": 0.084, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 30.9375, | |
| "grad_norm": 2.147754430770874, | |
| "learning_rate": 3.51031746031746e-05, | |
| "loss": 0.0856, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 31.09375, | |
| "grad_norm": 2.1251964569091797, | |
| "learning_rate": 3.5023809523809525e-05, | |
| "loss": 0.0762, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 31.25, | |
| "grad_norm": 2.6378941535949707, | |
| "learning_rate": 3.4944444444444446e-05, | |
| "loss": 0.071, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 31.25, | |
| "eval_runtime": 2.0743, | |
| "eval_samples_per_second": 0.482, | |
| "eval_steps_per_second": 0.482, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 31.40625, | |
| "grad_norm": 2.126807689666748, | |
| "learning_rate": 3.486507936507937e-05, | |
| "loss": 0.0785, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 31.5625, | |
| "grad_norm": 2.4585835933685303, | |
| "learning_rate": 3.478571428571429e-05, | |
| "loss": 0.1053, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 31.71875, | |
| "grad_norm": 1.9649542570114136, | |
| "learning_rate": 3.470634920634921e-05, | |
| "loss": 0.0734, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 31.875, | |
| "grad_norm": 2.3006460666656494, | |
| "learning_rate": 3.462698412698413e-05, | |
| "loss": 0.0858, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 32.03125, | |
| "grad_norm": 2.163447856903076, | |
| "learning_rate": 3.4547619047619054e-05, | |
| "loss": 0.0826, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 32.1875, | |
| "grad_norm": 2.0441508293151855, | |
| "learning_rate": 3.4468253968253975e-05, | |
| "loss": 0.0712, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 32.34375, | |
| "grad_norm": 1.5795445442199707, | |
| "learning_rate": 3.438888888888889e-05, | |
| "loss": 0.056, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 32.5, | |
| "grad_norm": 3.5661540031433105, | |
| "learning_rate": 3.430952380952381e-05, | |
| "loss": 0.069, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 32.65625, | |
| "grad_norm": 1.6884055137634277, | |
| "learning_rate": 3.423015873015873e-05, | |
| "loss": 0.0899, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 32.8125, | |
| "grad_norm": 2.421724557876587, | |
| "learning_rate": 3.415079365079365e-05, | |
| "loss": 0.0768, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 32.96875, | |
| "grad_norm": 1.1584899425506592, | |
| "learning_rate": 3.407142857142857e-05, | |
| "loss": 0.0763, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 33.125, | |
| "grad_norm": 2.069801092147827, | |
| "learning_rate": 3.399206349206349e-05, | |
| "loss": 0.0636, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 33.28125, | |
| "grad_norm": 2.5826573371887207, | |
| "learning_rate": 3.391269841269841e-05, | |
| "loss": 0.0676, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 33.4375, | |
| "grad_norm": 1.119449496269226, | |
| "learning_rate": 3.3833333333333334e-05, | |
| "loss": 0.0613, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 33.59375, | |
| "grad_norm": 1.6656103134155273, | |
| "learning_rate": 3.3753968253968255e-05, | |
| "loss": 0.0598, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 33.75, | |
| "grad_norm": 1.7956265211105347, | |
| "learning_rate": 3.3674603174603177e-05, | |
| "loss": 0.0778, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 33.90625, | |
| "grad_norm": 2.2106685638427734, | |
| "learning_rate": 3.35952380952381e-05, | |
| "loss": 0.0737, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 34.0625, | |
| "grad_norm": 2.4478724002838135, | |
| "learning_rate": 3.351587301587302e-05, | |
| "loss": 0.0642, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 34.21875, | |
| "grad_norm": 2.3334341049194336, | |
| "learning_rate": 3.343650793650794e-05, | |
| "loss": 0.0697, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 34.375, | |
| "grad_norm": 1.870275855064392, | |
| "learning_rate": 3.3357142857142856e-05, | |
| "loss": 0.0523, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 34.53125, | |
| "grad_norm": 1.283444881439209, | |
| "learning_rate": 3.327777777777778e-05, | |
| "loss": 0.0619, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 34.6875, | |
| "grad_norm": 1.7918671369552612, | |
| "learning_rate": 3.31984126984127e-05, | |
| "loss": 0.061, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 34.84375, | |
| "grad_norm": 1.6546680927276611, | |
| "learning_rate": 3.311904761904762e-05, | |
| "loss": 0.058, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 35.0, | |
| "grad_norm": 3.8063647747039795, | |
| "learning_rate": 3.303968253968254e-05, | |
| "loss": 0.0605, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 35.15625, | |
| "grad_norm": 1.8701483011245728, | |
| "learning_rate": 3.296031746031746e-05, | |
| "loss": 0.0652, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 35.3125, | |
| "grad_norm": 0.48190346360206604, | |
| "learning_rate": 3.2880952380952385e-05, | |
| "loss": 0.0443, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 35.46875, | |
| "grad_norm": 1.0227997303009033, | |
| "learning_rate": 3.2801587301587306e-05, | |
| "loss": 0.0397, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 35.625, | |
| "grad_norm": 1.4256937503814697, | |
| "learning_rate": 3.272222222222223e-05, | |
| "loss": 0.0541, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 35.78125, | |
| "grad_norm": 2.3451199531555176, | |
| "learning_rate": 3.264285714285714e-05, | |
| "loss": 0.0568, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 35.9375, | |
| "grad_norm": 1.3683526515960693, | |
| "learning_rate": 3.2563492063492064e-05, | |
| "loss": 0.0587, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 36.09375, | |
| "grad_norm": 1.4543867111206055, | |
| "learning_rate": 3.2484126984126986e-05, | |
| "loss": 0.0394, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 36.25, | |
| "grad_norm": 1.9606877565383911, | |
| "learning_rate": 3.240476190476191e-05, | |
| "loss": 0.0545, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 36.40625, | |
| "grad_norm": 1.6910959482192993, | |
| "learning_rate": 3.232539682539683e-05, | |
| "loss": 0.0638, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 36.5625, | |
| "grad_norm": 1.735841155052185, | |
| "learning_rate": 3.224603174603174e-05, | |
| "loss": 0.0569, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 36.71875, | |
| "grad_norm": 1.5598944425582886, | |
| "learning_rate": 3.2166666666666665e-05, | |
| "loss": 0.0649, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 36.875, | |
| "grad_norm": 1.5700335502624512, | |
| "learning_rate": 3.2087301587301586e-05, | |
| "loss": 0.0689, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 37.03125, | |
| "grad_norm": 1.0767812728881836, | |
| "learning_rate": 3.200793650793651e-05, | |
| "loss": 0.0535, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 37.1875, | |
| "grad_norm": 2.505707263946533, | |
| "learning_rate": 3.192857142857143e-05, | |
| "loss": 0.0486, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 37.34375, | |
| "grad_norm": 2.1414477825164795, | |
| "learning_rate": 3.184920634920635e-05, | |
| "loss": 0.0462, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 37.5, | |
| "grad_norm": 1.0493581295013428, | |
| "learning_rate": 3.176984126984127e-05, | |
| "loss": 0.0515, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 37.65625, | |
| "grad_norm": 1.5255446434020996, | |
| "learning_rate": 3.1690476190476194e-05, | |
| "loss": 0.0485, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 37.8125, | |
| "grad_norm": 1.3334754705429077, | |
| "learning_rate": 3.1611111111111115e-05, | |
| "loss": 0.056, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 37.96875, | |
| "grad_norm": 1.745110273361206, | |
| "learning_rate": 3.153174603174604e-05, | |
| "loss": 0.0473, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 38.125, | |
| "grad_norm": 1.4413131475448608, | |
| "learning_rate": 3.145238095238096e-05, | |
| "loss": 0.0317, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 38.28125, | |
| "grad_norm": 1.6811962127685547, | |
| "learning_rate": 3.137301587301587e-05, | |
| "loss": 0.0496, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 38.4375, | |
| "grad_norm": 1.2820957899093628, | |
| "learning_rate": 3.1293650793650795e-05, | |
| "loss": 0.0369, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 38.59375, | |
| "grad_norm": 1.7850005626678467, | |
| "learning_rate": 3.1214285714285716e-05, | |
| "loss": 0.0478, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 38.75, | |
| "grad_norm": 2.456017255783081, | |
| "learning_rate": 3.113492063492064e-05, | |
| "loss": 0.0507, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 38.90625, | |
| "grad_norm": 2.1933865547180176, | |
| "learning_rate": 3.105555555555555e-05, | |
| "loss": 0.0436, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 39.0625, | |
| "grad_norm": 0.9716876745223999, | |
| "learning_rate": 3.0976190476190474e-05, | |
| "loss": 0.0439, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 39.0625, | |
| "eval_runtime": 2.0072, | |
| "eval_samples_per_second": 0.498, | |
| "eval_steps_per_second": 0.498, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 39.21875, | |
| "grad_norm": 2.0628113746643066, | |
| "learning_rate": 3.0896825396825395e-05, | |
| "loss": 0.0408, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 39.375, | |
| "grad_norm": 1.5137745141983032, | |
| "learning_rate": 3.081746031746032e-05, | |
| "loss": 0.0445, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 39.53125, | |
| "grad_norm": 1.6775884628295898, | |
| "learning_rate": 3.073809523809524e-05, | |
| "loss": 0.051, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 39.6875, | |
| "grad_norm": 1.440619707107544, | |
| "learning_rate": 3.065873015873016e-05, | |
| "loss": 0.0387, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 39.84375, | |
| "grad_norm": 2.1149260997772217, | |
| "learning_rate": 3.057936507936508e-05, | |
| "loss": 0.0387, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 40.0, | |
| "grad_norm": 2.163191318511963, | |
| "learning_rate": 3.05e-05, | |
| "loss": 0.0363, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 40.15625, | |
| "grad_norm": 3.246634006500244, | |
| "learning_rate": 3.042063492063492e-05, | |
| "loss": 0.0429, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 40.3125, | |
| "grad_norm": 1.188644289970398, | |
| "learning_rate": 3.0341269841269842e-05, | |
| "loss": 0.0316, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 40.46875, | |
| "grad_norm": 2.2617383003234863, | |
| "learning_rate": 3.0261904761904764e-05, | |
| "loss": 0.0406, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 40.625, | |
| "grad_norm": 0.6037064790725708, | |
| "learning_rate": 3.0182539682539685e-05, | |
| "loss": 0.031, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 40.78125, | |
| "grad_norm": 1.1936763525009155, | |
| "learning_rate": 3.0103174603174607e-05, | |
| "loss": 0.0401, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 40.9375, | |
| "grad_norm": 3.0310215950012207, | |
| "learning_rate": 3.0023809523809525e-05, | |
| "loss": 0.0339, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 41.09375, | |
| "grad_norm": 0.6716585755348206, | |
| "learning_rate": 2.9944444444444446e-05, | |
| "loss": 0.0338, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 41.25, | |
| "grad_norm": 0.5010519623756409, | |
| "learning_rate": 2.9865079365079368e-05, | |
| "loss": 0.0315, | |
| "step": 2640 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 6400, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 100, | |
| "save_steps": 115, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 0.0, | |
| "train_batch_size": 8, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |