| { | |
| "best_global_step": 5100, | |
| "best_metric": 1.591291904449463, | |
| "best_model_checkpoint": "./checkpoints/gpt2_test_chunks30/checkpoint-5100", | |
| "epoch": 0.51, | |
| "eval_steps": 50, | |
| "global_step": 5100, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.001, | |
| "grad_norm": 4.633359432220459, | |
| "learning_rate": 1.8e-06, | |
| "loss": 7.7108, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.002, | |
| "grad_norm": 2.3646275997161865, | |
| "learning_rate": 3.8e-06, | |
| "loss": 7.4581, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.003, | |
| "grad_norm": 2.262131929397583, | |
| "learning_rate": 5.8e-06, | |
| "loss": 7.2566, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.004, | |
| "grad_norm": 2.6301262378692627, | |
| "learning_rate": 7.8e-06, | |
| "loss": 7.0053, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.005, | |
| "grad_norm": 2.221440315246582, | |
| "learning_rate": 9.800000000000001e-06, | |
| "loss": 6.6106, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.005, | |
| "eval_loss": 6.111213684082031, | |
| "eval_runtime": 4.5453, | |
| "eval_samples_per_second": 440.015, | |
| "eval_steps_per_second": 27.501, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.006, | |
| "grad_norm": 1.959439992904663, | |
| "learning_rate": 1.18e-05, | |
| "loss": 6.1216, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.007, | |
| "grad_norm": 1.562995195388794, | |
| "learning_rate": 1.3800000000000002e-05, | |
| "loss": 5.6405, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.008, | |
| "grad_norm": 1.4331400394439697, | |
| "learning_rate": 1.58e-05, | |
| "loss": 5.1521, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 0.009, | |
| "grad_norm": 1.6445435285568237, | |
| "learning_rate": 1.78e-05, | |
| "loss": 4.7211, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "grad_norm": 1.3628312349319458, | |
| "learning_rate": 1.9800000000000004e-05, | |
| "loss": 4.4151, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.01, | |
| "eval_loss": 4.110222816467285, | |
| "eval_runtime": 4.5472, | |
| "eval_samples_per_second": 439.827, | |
| "eval_steps_per_second": 27.489, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 0.011, | |
| "grad_norm": 1.137888789176941, | |
| "learning_rate": 2.18e-05, | |
| "loss": 4.1114, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 0.012, | |
| "grad_norm": 1.4980727434158325, | |
| "learning_rate": 2.38e-05, | |
| "loss": 3.8488, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 0.013, | |
| "grad_norm": 1.1802358627319336, | |
| "learning_rate": 2.58e-05, | |
| "loss": 3.5541, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 0.014, | |
| "grad_norm": 1.0807380676269531, | |
| "learning_rate": 2.7800000000000005e-05, | |
| "loss": 3.4673, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 0.015, | |
| "grad_norm": 0.957696259021759, | |
| "learning_rate": 2.98e-05, | |
| "loss": 3.2401, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.015, | |
| "eval_loss": 3.2231221199035645, | |
| "eval_runtime": 4.5997, | |
| "eval_samples_per_second": 434.814, | |
| "eval_steps_per_second": 27.176, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 0.016, | |
| "grad_norm": 0.8120741248130798, | |
| "learning_rate": 3.18e-05, | |
| "loss": 3.1228, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.017, | |
| "grad_norm": 0.8813096284866333, | |
| "learning_rate": 3.38e-05, | |
| "loss": 3.1247, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 0.018, | |
| "grad_norm": 0.7591573596000671, | |
| "learning_rate": 3.58e-05, | |
| "loss": 2.9657, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 0.019, | |
| "grad_norm": 0.6872971057891846, | |
| "learning_rate": 3.7800000000000004e-05, | |
| "loss": 2.8526, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "grad_norm": 0.7402148246765137, | |
| "learning_rate": 3.9800000000000005e-05, | |
| "loss": 2.7961, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.02, | |
| "eval_loss": 2.8180038928985596, | |
| "eval_runtime": 4.579, | |
| "eval_samples_per_second": 436.777, | |
| "eval_steps_per_second": 27.299, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 0.021, | |
| "grad_norm": 0.7957950234413147, | |
| "learning_rate": 4.18e-05, | |
| "loss": 2.8335, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 0.022, | |
| "grad_norm": 0.6467453837394714, | |
| "learning_rate": 4.38e-05, | |
| "loss": 2.7491, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 0.023, | |
| "grad_norm": 0.7767886519432068, | |
| "learning_rate": 4.58e-05, | |
| "loss": 2.7367, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 0.024, | |
| "grad_norm": 0.7259961366653442, | |
| "learning_rate": 4.78e-05, | |
| "loss": 2.6567, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 0.025, | |
| "grad_norm": 0.6372230648994446, | |
| "learning_rate": 4.9800000000000004e-05, | |
| "loss": 2.6064, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.025, | |
| "eval_loss": 2.628188133239746, | |
| "eval_runtime": 4.6002, | |
| "eval_samples_per_second": 434.768, | |
| "eval_steps_per_second": 27.173, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 0.026, | |
| "grad_norm": 0.6601161956787109, | |
| "learning_rate": 5.1800000000000005e-05, | |
| "loss": 2.6328, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 0.027, | |
| "grad_norm": 0.5882264375686646, | |
| "learning_rate": 5.380000000000001e-05, | |
| "loss": 2.5728, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 0.028, | |
| "grad_norm": 0.601144552230835, | |
| "learning_rate": 5.580000000000001e-05, | |
| "loss": 2.5354, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 0.029, | |
| "grad_norm": 0.4997910261154175, | |
| "learning_rate": 5.7799999999999995e-05, | |
| "loss": 2.586, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "grad_norm": 0.5014902949333191, | |
| "learning_rate": 5.9800000000000003e-05, | |
| "loss": 2.5666, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.03, | |
| "eval_loss": 2.5160727500915527, | |
| "eval_runtime": 4.5957, | |
| "eval_samples_per_second": 435.193, | |
| "eval_steps_per_second": 27.2, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 0.031, | |
| "grad_norm": 0.5040959715843201, | |
| "learning_rate": 6.18e-05, | |
| "loss": 2.5207, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 0.032, | |
| "grad_norm": 0.5737811923027039, | |
| "learning_rate": 6.38e-05, | |
| "loss": 2.5029, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.033, | |
| "grad_norm": 0.5511871576309204, | |
| "learning_rate": 6.58e-05, | |
| "loss": 2.4785, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 0.034, | |
| "grad_norm": 0.5300912857055664, | |
| "learning_rate": 6.780000000000001e-05, | |
| "loss": 2.451, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 0.035, | |
| "grad_norm": 0.5354319214820862, | |
| "learning_rate": 6.98e-05, | |
| "loss": 2.4384, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.035, | |
| "eval_loss": 2.447114944458008, | |
| "eval_runtime": 4.6248, | |
| "eval_samples_per_second": 432.454, | |
| "eval_steps_per_second": 27.028, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 0.036, | |
| "grad_norm": 0.5186896324157715, | |
| "learning_rate": 7.18e-05, | |
| "loss": 2.3922, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 0.037, | |
| "grad_norm": 0.6687894463539124, | |
| "learning_rate": 7.38e-05, | |
| "loss": 2.3011, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 0.038, | |
| "grad_norm": 0.6902775764465332, | |
| "learning_rate": 7.58e-05, | |
| "loss": 2.3917, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 0.039, | |
| "grad_norm": 0.5241763591766357, | |
| "learning_rate": 7.780000000000001e-05, | |
| "loss": 2.4254, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "grad_norm": 0.5403384566307068, | |
| "learning_rate": 7.98e-05, | |
| "loss": 2.3213, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.04, | |
| "eval_loss": 2.4054787158966064, | |
| "eval_runtime": 4.6055, | |
| "eval_samples_per_second": 434.265, | |
| "eval_steps_per_second": 27.142, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 0.041, | |
| "grad_norm": 0.46202412247657776, | |
| "learning_rate": 8.18e-05, | |
| "loss": 2.3921, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 0.042, | |
| "grad_norm": 0.5079010725021362, | |
| "learning_rate": 8.38e-05, | |
| "loss": 2.3346, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 0.043, | |
| "grad_norm": 0.4820123612880707, | |
| "learning_rate": 8.58e-05, | |
| "loss": 2.2619, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 0.044, | |
| "grad_norm": 0.7787259817123413, | |
| "learning_rate": 8.78e-05, | |
| "loss": 2.2583, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 0.045, | |
| "grad_norm": 0.4925244152545929, | |
| "learning_rate": 8.98e-05, | |
| "loss": 2.3056, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.045, | |
| "eval_loss": 2.36269474029541, | |
| "eval_runtime": 4.6434, | |
| "eval_samples_per_second": 430.723, | |
| "eval_steps_per_second": 26.92, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 0.046, | |
| "grad_norm": 0.49801623821258545, | |
| "learning_rate": 9.180000000000001e-05, | |
| "loss": 2.3533, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 0.047, | |
| "grad_norm": 0.5879547595977783, | |
| "learning_rate": 9.38e-05, | |
| "loss": 2.3018, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 0.048, | |
| "grad_norm": 0.4898471236228943, | |
| "learning_rate": 9.58e-05, | |
| "loss": 2.3325, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.049, | |
| "grad_norm": 0.5592429041862488, | |
| "learning_rate": 9.78e-05, | |
| "loss": 2.2995, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "grad_norm": 0.6377447843551636, | |
| "learning_rate": 9.98e-05, | |
| "loss": 2.2615, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.05, | |
| "eval_loss": 2.3109450340270996, | |
| "eval_runtime": 4.6129, | |
| "eval_samples_per_second": 433.566, | |
| "eval_steps_per_second": 27.098, | |
| "step": 500 | |
| }, | |
| { | |
| "epoch": 0.051, | |
| "grad_norm": 0.4234420955181122, | |
| "learning_rate": 9.911197057469107e-05, | |
| "loss": 2.2316, | |
| "step": 510 | |
| }, | |
| { | |
| "epoch": 0.052, | |
| "grad_norm": 0.40216103196144104, | |
| "learning_rate": 9.815249038111776e-05, | |
| "loss": 2.2426, | |
| "step": 520 | |
| }, | |
| { | |
| "epoch": 0.053, | |
| "grad_norm": 0.46068865060806274, | |
| "learning_rate": 9.722034684781694e-05, | |
| "loss": 2.2844, | |
| "step": 530 | |
| }, | |
| { | |
| "epoch": 0.054, | |
| "grad_norm": 0.3902567923069, | |
| "learning_rate": 9.631426606617744e-05, | |
| "loss": 2.1736, | |
| "step": 540 | |
| }, | |
| { | |
| "epoch": 0.055, | |
| "grad_norm": 0.4389937222003937, | |
| "learning_rate": 9.543305571897804e-05, | |
| "loss": 2.2481, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.055, | |
| "eval_loss": 2.266875982284546, | |
| "eval_runtime": 4.6373, | |
| "eval_samples_per_second": 431.283, | |
| "eval_steps_per_second": 26.955, | |
| "step": 550 | |
| }, | |
| { | |
| "epoch": 0.056, | |
| "grad_norm": 0.4390330910682678, | |
| "learning_rate": 9.457559848219179e-05, | |
| "loss": 2.2735, | |
| "step": 560 | |
| }, | |
| { | |
| "epoch": 0.057, | |
| "grad_norm": 0.440461128950119, | |
| "learning_rate": 9.374084606744877e-05, | |
| "loss": 2.2337, | |
| "step": 570 | |
| }, | |
| { | |
| "epoch": 0.058, | |
| "grad_norm": 0.5304872393608093, | |
| "learning_rate": 9.292781383291611e-05, | |
| "loss": 2.2897, | |
| "step": 580 | |
| }, | |
| { | |
| "epoch": 0.059, | |
| "grad_norm": 0.4079591929912567, | |
| "learning_rate": 9.213557589959345e-05, | |
| "loss": 2.2064, | |
| "step": 590 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "grad_norm": 0.4388200640678406, | |
| "learning_rate": 9.136326071794409e-05, | |
| "loss": 2.1916, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.06, | |
| "eval_loss": 2.2343385219573975, | |
| "eval_runtime": 4.6153, | |
| "eval_samples_per_second": 433.343, | |
| "eval_steps_per_second": 27.084, | |
| "step": 600 | |
| }, | |
| { | |
| "epoch": 0.061, | |
| "grad_norm": 0.4618544280529022, | |
| "learning_rate": 9.061004703659374e-05, | |
| "loss": 2.2159, | |
| "step": 610 | |
| }, | |
| { | |
| "epoch": 0.062, | |
| "grad_norm": 0.42200735211372375, | |
| "learning_rate": 8.987516023070193e-05, | |
| "loss": 2.1533, | |
| "step": 620 | |
| }, | |
| { | |
| "epoch": 0.063, | |
| "grad_norm": 0.5431024432182312, | |
| "learning_rate": 8.915786895268651e-05, | |
| "loss": 2.194, | |
| "step": 630 | |
| }, | |
| { | |
| "epoch": 0.064, | |
| "grad_norm": 0.4802396595478058, | |
| "learning_rate": 8.84574820723792e-05, | |
| "loss": 2.1996, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.065, | |
| "grad_norm": 0.5224118828773499, | |
| "learning_rate": 8.777334587751072e-05, | |
| "loss": 2.1971, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.065, | |
| "eval_loss": 2.1872010231018066, | |
| "eval_runtime": 4.6406, | |
| "eval_samples_per_second": 430.98, | |
| "eval_steps_per_second": 26.936, | |
| "step": 650 | |
| }, | |
| { | |
| "epoch": 0.066, | |
| "grad_norm": 0.46909448504447937, | |
| "learning_rate": 8.710484150874758e-05, | |
| "loss": 2.1833, | |
| "step": 660 | |
| }, | |
| { | |
| "epoch": 0.067, | |
| "grad_norm": 0.43358996510505676, | |
| "learning_rate": 8.645138260640511e-05, | |
| "loss": 2.1338, | |
| "step": 670 | |
| }, | |
| { | |
| "epoch": 0.068, | |
| "grad_norm": 0.6461070775985718, | |
| "learning_rate": 8.581241314849611e-05, | |
| "loss": 2.0534, | |
| "step": 680 | |
| }, | |
| { | |
| "epoch": 0.069, | |
| "grad_norm": 0.5167747735977173, | |
| "learning_rate": 8.51874054619982e-05, | |
| "loss": 2.0476, | |
| "step": 690 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "grad_norm": 0.6826158165931702, | |
| "learning_rate": 8.457585839117285e-05, | |
| "loss": 2.1696, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.07, | |
| "eval_loss": 2.1201064586639404, | |
| "eval_runtime": 4.6341, | |
| "eval_samples_per_second": 431.585, | |
| "eval_steps_per_second": 26.974, | |
| "step": 700 | |
| }, | |
| { | |
| "epoch": 0.071, | |
| "grad_norm": 0.5946829319000244, | |
| "learning_rate": 8.397729560848631e-05, | |
| "loss": 2.1121, | |
| "step": 710 | |
| }, | |
| { | |
| "epoch": 0.072, | |
| "grad_norm": 0.6427505612373352, | |
| "learning_rate": 8.339126405519483e-05, | |
| "loss": 2.0529, | |
| "step": 720 | |
| }, | |
| { | |
| "epoch": 0.073, | |
| "grad_norm": 0.8502755761146545, | |
| "learning_rate": 8.281733249999222e-05, | |
| "loss": 2.0783, | |
| "step": 730 | |
| }, | |
| { | |
| "epoch": 0.074, | |
| "grad_norm": 0.5335196852684021, | |
| "learning_rate": 8.225509020529976e-05, | |
| "loss": 2.0196, | |
| "step": 740 | |
| }, | |
| { | |
| "epoch": 0.075, | |
| "grad_norm": 0.5112911462783813, | |
| "learning_rate": 8.170414569182506e-05, | |
| "loss": 1.9775, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.075, | |
| "eval_loss": 2.051039934158325, | |
| "eval_runtime": 4.6427, | |
| "eval_samples_per_second": 430.782, | |
| "eval_steps_per_second": 26.924, | |
| "step": 750 | |
| }, | |
| { | |
| "epoch": 0.076, | |
| "grad_norm": 0.6587328910827637, | |
| "learning_rate": 8.116412559294567e-05, | |
| "loss": 2.0011, | |
| "step": 760 | |
| }, | |
| { | |
| "epoch": 0.077, | |
| "grad_norm": 0.4874884784221649, | |
| "learning_rate": 8.063467359130037e-05, | |
| "loss": 2.0269, | |
| "step": 770 | |
| }, | |
| { | |
| "epoch": 0.078, | |
| "grad_norm": 0.5324885845184326, | |
| "learning_rate": 8.011544943070565e-05, | |
| "loss": 1.92, | |
| "step": 780 | |
| }, | |
| { | |
| "epoch": 0.079, | |
| "grad_norm": 1.048747181892395, | |
| "learning_rate": 7.960612799717214e-05, | |
| "loss": 2.0112, | |
| "step": 790 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "grad_norm": 0.5203439593315125, | |
| "learning_rate": 7.910639846338163e-05, | |
| "loss": 2.0567, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.08, | |
| "eval_loss": 2.01043438911438, | |
| "eval_runtime": 4.6186, | |
| "eval_samples_per_second": 433.034, | |
| "eval_steps_per_second": 27.065, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.081, | |
| "grad_norm": 0.526197075843811, | |
| "learning_rate": 7.861596349150974e-05, | |
| "loss": 2.0106, | |
| "step": 810 | |
| }, | |
| { | |
| "epoch": 0.082, | |
| "grad_norm": 0.5369076132774353, | |
| "learning_rate": 7.813453848974926e-05, | |
| "loss": 2.0475, | |
| "step": 820 | |
| }, | |
| { | |
| "epoch": 0.083, | |
| "grad_norm": 0.5480849742889404, | |
| "learning_rate": 7.766185091831061e-05, | |
| "loss": 1.9635, | |
| "step": 830 | |
| }, | |
| { | |
| "epoch": 0.084, | |
| "grad_norm": 0.5581432580947876, | |
| "learning_rate": 7.719763964105497e-05, | |
| "loss": 1.9332, | |
| "step": 840 | |
| }, | |
| { | |
| "epoch": 0.085, | |
| "grad_norm": 0.45510992407798767, | |
| "learning_rate": 7.674165431925523e-05, | |
| "loss": 1.9597, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.085, | |
| "eval_loss": 1.9628287553787231, | |
| "eval_runtime": 4.6152, | |
| "eval_samples_per_second": 433.353, | |
| "eval_steps_per_second": 27.085, | |
| "step": 850 | |
| }, | |
| { | |
| "epoch": 0.086, | |
| "grad_norm": 0.5276203155517578, | |
| "learning_rate": 7.629365484428845e-05, | |
| "loss": 1.9356, | |
| "step": 860 | |
| }, | |
| { | |
| "epoch": 0.087, | |
| "grad_norm": 0.5071442723274231, | |
| "learning_rate": 7.585341080633831e-05, | |
| "loss": 2.0046, | |
| "step": 870 | |
| }, | |
| { | |
| "epoch": 0.088, | |
| "grad_norm": 0.4676225483417511, | |
| "learning_rate": 7.542070099643788e-05, | |
| "loss": 1.9443, | |
| "step": 880 | |
| }, | |
| { | |
| "epoch": 0.089, | |
| "grad_norm": 0.5725594758987427, | |
| "learning_rate": 7.499531293940736e-05, | |
| "loss": 1.8964, | |
| "step": 890 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "grad_norm": 0.5482536554336548, | |
| "learning_rate": 7.457704245544709e-05, | |
| "loss": 1.9238, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.09, | |
| "eval_loss": 1.927959680557251, | |
| "eval_runtime": 4.609, | |
| "eval_samples_per_second": 433.932, | |
| "eval_steps_per_second": 27.121, | |
| "step": 900 | |
| }, | |
| { | |
| "epoch": 0.091, | |
| "grad_norm": 0.4957098960876465, | |
| "learning_rate": 7.41656932483308e-05, | |
| "loss": 1.937, | |
| "step": 910 | |
| }, | |
| { | |
| "epoch": 0.092, | |
| "grad_norm": 0.5255835652351379, | |
| "learning_rate": 7.376107651831263e-05, | |
| "loss": 1.955, | |
| "step": 920 | |
| }, | |
| { | |
| "epoch": 0.093, | |
| "grad_norm": 0.41037118434906006, | |
| "learning_rate": 7.336301059801394e-05, | |
| "loss": 1.8897, | |
| "step": 930 | |
| }, | |
| { | |
| "epoch": 0.094, | |
| "grad_norm": 0.6042472720146179, | |
| "learning_rate": 7.297132060969499e-05, | |
| "loss": 1.8852, | |
| "step": 940 | |
| }, | |
| { | |
| "epoch": 0.095, | |
| "grad_norm": 0.5588452219963074, | |
| "learning_rate": 7.258583814244268e-05, | |
| "loss": 1.8786, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.095, | |
| "eval_loss": 1.892114520072937, | |
| "eval_runtime": 4.6266, | |
| "eval_samples_per_second": 432.278, | |
| "eval_steps_per_second": 27.017, | |
| "step": 950 | |
| }, | |
| { | |
| "epoch": 0.096, | |
| "grad_norm": 0.5261280536651611, | |
| "learning_rate": 7.220640094792103e-05, | |
| "loss": 1.9327, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.097, | |
| "grad_norm": 0.5294102430343628, | |
| "learning_rate": 7.183285265343593e-05, | |
| "loss": 1.925, | |
| "step": 970 | |
| }, | |
| { | |
| "epoch": 0.098, | |
| "grad_norm": 0.4203556776046753, | |
| "learning_rate": 7.14650424911616e-05, | |
| "loss": 1.8371, | |
| "step": 980 | |
| }, | |
| { | |
| "epoch": 0.099, | |
| "grad_norm": 0.5418606996536255, | |
| "learning_rate": 7.110282504246376e-05, | |
| "loss": 1.8636, | |
| "step": 990 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "grad_norm": 0.5804753303527832, | |
| "learning_rate": 7.074605999633481e-05, | |
| "loss": 1.8812, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.1, | |
| "eval_loss": 1.8822709321975708, | |
| "eval_runtime": 4.6061, | |
| "eval_samples_per_second": 434.203, | |
| "eval_steps_per_second": 27.138, | |
| "step": 1000 | |
| }, | |
| { | |
| "epoch": 0.101, | |
| "grad_norm": 0.4180801510810852, | |
| "learning_rate": 7.03946119210298e-05, | |
| "loss": 1.9037, | |
| "step": 1010 | |
| }, | |
| { | |
| "epoch": 0.102, | |
| "grad_norm": 0.40715548396110535, | |
| "learning_rate": 7.004835004805859e-05, | |
| "loss": 1.8844, | |
| "step": 1020 | |
| }, | |
| { | |
| "epoch": 0.103, | |
| "grad_norm": 0.6151288747787476, | |
| "learning_rate": 6.970714806775237e-05, | |
| "loss": 1.8651, | |
| "step": 1030 | |
| }, | |
| { | |
| "epoch": 0.104, | |
| "grad_norm": 0.45094242691993713, | |
| "learning_rate": 6.937088393567812e-05, | |
| "loss": 1.8962, | |
| "step": 1040 | |
| }, | |
| { | |
| "epoch": 0.105, | |
| "grad_norm": 0.5133697390556335, | |
| "learning_rate": 6.903943968922749e-05, | |
| "loss": 1.8309, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.105, | |
| "eval_loss": 1.8607650995254517, | |
| "eval_runtime": 4.6375, | |
| "eval_samples_per_second": 431.269, | |
| "eval_steps_per_second": 26.954, | |
| "step": 1050 | |
| }, | |
| { | |
| "epoch": 0.106, | |
| "grad_norm": 0.502295196056366, | |
| "learning_rate": 6.871270127375409e-05, | |
| "loss": 1.8592, | |
| "step": 1060 | |
| }, | |
| { | |
| "epoch": 0.107, | |
| "grad_norm": 0.46987488865852356, | |
| "learning_rate": 6.839055837767724e-05, | |
| "loss": 1.7896, | |
| "step": 1070 | |
| }, | |
| { | |
| "epoch": 0.108, | |
| "grad_norm": 0.5071763396263123, | |
| "learning_rate": 6.807290427601058e-05, | |
| "loss": 1.7998, | |
| "step": 1080 | |
| }, | |
| { | |
| "epoch": 0.109, | |
| "grad_norm": 0.43212854862213135, | |
| "learning_rate": 6.775963568181182e-05, | |
| "loss": 1.8668, | |
| "step": 1090 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "grad_norm": 0.548732340335846, | |
| "learning_rate": 6.74506526050837e-05, | |
| "loss": 1.8052, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.11, | |
| "eval_loss": 1.84357750415802, | |
| "eval_runtime": 4.6161, | |
| "eval_samples_per_second": 433.267, | |
| "eval_steps_per_second": 27.079, | |
| "step": 1100 | |
| }, | |
| { | |
| "epoch": 0.111, | |
| "grad_norm": 0.4961816966533661, | |
| "learning_rate": 6.714585821868878e-05, | |
| "loss": 1.7435, | |
| "step": 1110 | |
| }, | |
| { | |
| "epoch": 0.112, | |
| "grad_norm": 0.48313409090042114, | |
| "learning_rate": 6.68451587308695e-05, | |
| "loss": 1.8251, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 0.113, | |
| "grad_norm": 0.49864089488983154, | |
| "learning_rate": 6.654846326399234e-05, | |
| "loss": 1.828, | |
| "step": 1130 | |
| }, | |
| { | |
| "epoch": 0.114, | |
| "grad_norm": 0.4988366961479187, | |
| "learning_rate": 6.625568373916034e-05, | |
| "loss": 1.8395, | |
| "step": 1140 | |
| }, | |
| { | |
| "epoch": 0.115, | |
| "grad_norm": 0.46008503437042236, | |
| "learning_rate": 6.596673476636102e-05, | |
| "loss": 1.8266, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.115, | |
| "eval_loss": 1.8144086599349976, | |
| "eval_runtime": 4.6156, | |
| "eval_samples_per_second": 433.312, | |
| "eval_steps_per_second": 27.082, | |
| "step": 1150 | |
| }, | |
| { | |
| "epoch": 0.116, | |
| "grad_norm": 0.5102156400680542, | |
| "learning_rate": 6.568153353983866e-05, | |
| "loss": 1.878, | |
| "step": 1160 | |
| }, | |
| { | |
| "epoch": 0.117, | |
| "grad_norm": 0.4590121805667877, | |
| "learning_rate": 6.53999997384e-05, | |
| "loss": 1.8488, | |
| "step": 1170 | |
| }, | |
| { | |
| "epoch": 0.118, | |
| "grad_norm": 0.4306875467300415, | |
| "learning_rate": 6.512205543038029e-05, | |
| "loss": 1.7797, | |
| "step": 1180 | |
| }, | |
| { | |
| "epoch": 0.119, | |
| "grad_norm": 0.5295641422271729, | |
| "learning_rate": 6.48476249830151e-05, | |
| "loss": 1.7646, | |
| "step": 1190 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "grad_norm": 0.4984031319618225, | |
| "learning_rate": 6.457663497597783e-05, | |
| "loss": 1.8432, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.12, | |
| "eval_loss": 1.806372880935669, | |
| "eval_runtime": 4.603, | |
| "eval_samples_per_second": 434.496, | |
| "eval_steps_per_second": 27.156, | |
| "step": 1200 | |
| }, | |
| { | |
| "epoch": 0.121, | |
| "grad_norm": 0.42529740929603577, | |
| "learning_rate": 6.430901411885911e-05, | |
| "loss": 1.7983, | |
| "step": 1210 | |
| }, | |
| { | |
| "epoch": 0.122, | |
| "grad_norm": 0.44186535477638245, | |
| "learning_rate": 6.40446931723768e-05, | |
| "loss": 1.7505, | |
| "step": 1220 | |
| }, | |
| { | |
| "epoch": 0.123, | |
| "grad_norm": 0.5641934871673584, | |
| "learning_rate": 6.378360487311965e-05, | |
| "loss": 1.7354, | |
| "step": 1230 | |
| }, | |
| { | |
| "epoch": 0.124, | |
| "grad_norm": 0.49620485305786133, | |
| "learning_rate": 6.352568386163805e-05, | |
| "loss": 1.8056, | |
| "step": 1240 | |
| }, | |
| { | |
| "epoch": 0.125, | |
| "grad_norm": 0.43371838331222534, | |
| "learning_rate": 6.327086661370808e-05, | |
| "loss": 1.7984, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.125, | |
| "eval_loss": 1.7932993173599243, | |
| "eval_runtime": 4.6132, | |
| "eval_samples_per_second": 433.54, | |
| "eval_steps_per_second": 27.096, | |
| "step": 1250 | |
| }, | |
| { | |
| "epoch": 0.126, | |
| "grad_norm": 0.5027583837509155, | |
| "learning_rate": 6.301909137460409e-05, | |
| "loss": 1.688, | |
| "step": 1260 | |
| }, | |
| { | |
| "epoch": 0.127, | |
| "grad_norm": 0.5743605494499207, | |
| "learning_rate": 6.277029809622579e-05, | |
| "loss": 1.795, | |
| "step": 1270 | |
| }, | |
| { | |
| "epoch": 0.128, | |
| "grad_norm": 0.4767725169658661, | |
| "learning_rate": 6.252442837693433e-05, | |
| "loss": 1.7317, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 0.129, | |
| "grad_norm": 0.45642659068107605, | |
| "learning_rate": 6.22814254039606e-05, | |
| "loss": 1.7105, | |
| "step": 1290 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "grad_norm": 0.5867980122566223, | |
| "learning_rate": 6.204123389825647e-05, | |
| "loss": 1.8047, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.13, | |
| "eval_loss": 1.7906692028045654, | |
| "eval_runtime": 6.1816, | |
| "eval_samples_per_second": 323.54, | |
| "eval_steps_per_second": 20.221, | |
| "step": 1300 | |
| }, | |
| { | |
| "epoch": 0.131, | |
| "grad_norm": 0.4501126706600189, | |
| "learning_rate": 6.180380006166808e-05, | |
| "loss": 1.8556, | |
| "step": 1310 | |
| }, | |
| { | |
| "epoch": 0.132, | |
| "grad_norm": 0.49103260040283203, | |
| "learning_rate": 6.156907152631576e-05, | |
| "loss": 1.8429, | |
| "step": 1320 | |
| }, | |
| { | |
| "epoch": 0.133, | |
| "grad_norm": 0.4130401313304901, | |
| "learning_rate": 6.133699730607301e-05, | |
| "loss": 1.7908, | |
| "step": 1330 | |
| }, | |
| { | |
| "epoch": 0.134, | |
| "grad_norm": 0.5524632334709167, | |
| "learning_rate": 6.110752775004223e-05, | |
| "loss": 1.7754, | |
| "step": 1340 | |
| }, | |
| { | |
| "epoch": 0.135, | |
| "grad_norm": 0.4153362214565277, | |
| "learning_rate": 6.088061449793082e-05, | |
| "loss": 1.7705, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.135, | |
| "eval_loss": 1.778767466545105, | |
| "eval_runtime": 4.6221, | |
| "eval_samples_per_second": 432.704, | |
| "eval_steps_per_second": 27.044, | |
| "step": 1350 | |
| }, | |
| { | |
| "epoch": 0.136, | |
| "grad_norm": 0.4490492641925812, | |
| "learning_rate": 6.065621043723658e-05, | |
| "loss": 1.7255, | |
| "step": 1360 | |
| }, | |
| { | |
| "epoch": 0.137, | |
| "grad_norm": 0.453517884016037, | |
| "learning_rate": 6.043426966215649e-05, | |
| "loss": 1.7775, | |
| "step": 1370 | |
| }, | |
| { | |
| "epoch": 0.138, | |
| "grad_norm": 0.48418131470680237, | |
| "learning_rate": 6.021474743413714e-05, | |
| "loss": 1.8089, | |
| "step": 1380 | |
| }, | |
| { | |
| "epoch": 0.139, | |
| "grad_norm": 0.44093987345695496, | |
| "learning_rate": 5.999760014399041e-05, | |
| "loss": 1.7662, | |
| "step": 1390 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "grad_norm": 0.39292681217193604, | |
| "learning_rate": 5.978278527550084e-05, | |
| "loss": 1.7712, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.14, | |
| "eval_loss": 1.7568821907043457, | |
| "eval_runtime": 4.613, | |
| "eval_samples_per_second": 433.554, | |
| "eval_steps_per_second": 27.097, | |
| "step": 1400 | |
| }, | |
| { | |
| "epoch": 0.141, | |
| "grad_norm": 0.4755737781524658, | |
| "learning_rate": 5.957026137045648e-05, | |
| "loss": 1.761, | |
| "step": 1410 | |
| }, | |
| { | |
| "epoch": 0.142, | |
| "grad_norm": 0.4488621652126312, | |
| "learning_rate": 5.935998799503725e-05, | |
| "loss": 1.7299, | |
| "step": 1420 | |
| }, | |
| { | |
| "epoch": 0.143, | |
| "grad_norm": 0.44802042841911316, | |
| "learning_rate": 5.91519257074994e-05, | |
| "loss": 1.7244, | |
| "step": 1430 | |
| }, | |
| { | |
| "epoch": 0.144, | |
| "grad_norm": 0.481559157371521, | |
| "learning_rate": 5.8946036027097295e-05, | |
| "loss": 1.7574, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 0.145, | |
| "grad_norm": 0.39685600996017456, | |
| "learning_rate": 5.8742281404186785e-05, | |
| "loss": 1.7684, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.145, | |
| "eval_loss": 1.7553154230117798, | |
| "eval_runtime": 4.6273, | |
| "eval_samples_per_second": 432.213, | |
| "eval_steps_per_second": 27.013, | |
| "step": 1450 | |
| }, | |
| { | |
| "epoch": 0.146, | |
| "grad_norm": 0.5256659984588623, | |
| "learning_rate": 5.8540625191457576e-05, | |
| "loss": 1.7294, | |
| "step": 1460 | |
| }, | |
| { | |
| "epoch": 0.147, | |
| "grad_norm": 0.4324532449245453, | |
| "learning_rate": 5.834103161624459e-05, | |
| "loss": 1.678, | |
| "step": 1470 | |
| }, | |
| { | |
| "epoch": 0.148, | |
| "grad_norm": 0.4632064402103424, | |
| "learning_rate": 5.8143465753870694e-05, | |
| "loss": 1.7198, | |
| "step": 1480 | |
| }, | |
| { | |
| "epoch": 0.149, | |
| "grad_norm": 0.4310835301876068, | |
| "learning_rate": 5.7947893501975715e-05, | |
| "loss": 1.7345, | |
| "step": 1490 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "grad_norm": 0.39495712518692017, | |
| "learning_rate": 5.7754281555789e-05, | |
| "loss": 1.7207, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.15, | |
| "eval_loss": 1.7468315362930298, | |
| "eval_runtime": 4.6195, | |
| "eval_samples_per_second": 432.948, | |
| "eval_steps_per_second": 27.059, | |
| "step": 1500 | |
| }, | |
| { | |
| "epoch": 0.151, | |
| "grad_norm": 0.47850221395492554, | |
| "learning_rate": 5.756259738430475e-05, | |
| "loss": 1.6935, | |
| "step": 1510 | |
| }, | |
| { | |
| "epoch": 0.152, | |
| "grad_norm": 0.5015422105789185, | |
| "learning_rate": 5.7372809207321355e-05, | |
| "loss": 1.73, | |
| "step": 1520 | |
| }, | |
| { | |
| "epoch": 0.153, | |
| "grad_norm": 0.5264328718185425, | |
| "learning_rate": 5.71848859733081e-05, | |
| "loss": 1.7577, | |
| "step": 1530 | |
| }, | |
| { | |
| "epoch": 0.154, | |
| "grad_norm": 0.4855109751224518, | |
| "learning_rate": 5.699879733806412e-05, | |
| "loss": 1.7081, | |
| "step": 1540 | |
| }, | |
| { | |
| "epoch": 0.155, | |
| "grad_norm": 0.4310589134693146, | |
| "learning_rate": 5.681451364413635e-05, | |
| "loss": 1.6542, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.155, | |
| "eval_loss": 1.7462912797927856, | |
| "eval_runtime": 5.0409, | |
| "eval_samples_per_second": 396.753, | |
| "eval_steps_per_second": 24.797, | |
| "step": 1550 | |
| }, | |
| { | |
| "epoch": 0.156, | |
| "grad_norm": 0.4904212951660156, | |
| "learning_rate": 5.663200590096471e-05, | |
| "loss": 1.7219, | |
| "step": 1560 | |
| }, | |
| { | |
| "epoch": 0.157, | |
| "grad_norm": 0.46510931849479675, | |
| "learning_rate": 5.645124576572452e-05, | |
| "loss": 1.7661, | |
| "step": 1570 | |
| }, | |
| { | |
| "epoch": 0.158, | |
| "grad_norm": 0.48309725522994995, | |
| "learning_rate": 5.627220552483715e-05, | |
| "loss": 1.7406, | |
| "step": 1580 | |
| }, | |
| { | |
| "epoch": 0.159, | |
| "grad_norm": 0.4559168815612793, | |
| "learning_rate": 5.609485807612173e-05, | |
| "loss": 1.7269, | |
| "step": 1590 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "grad_norm": 0.3912065625190735, | |
| "learning_rate": 5.591917691156175e-05, | |
| "loss": 1.6987, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.16, | |
| "eval_loss": 1.7371795177459717, | |
| "eval_runtime": 4.6084, | |
| "eval_samples_per_second": 433.989, | |
| "eval_steps_per_second": 27.124, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 0.161, | |
| "grad_norm": 0.4040866792201996, | |
| "learning_rate": 5.5745136100661674e-05, | |
| "loss": 1.7261, | |
| "step": 1610 | |
| }, | |
| { | |
| "epoch": 0.162, | |
| "grad_norm": 0.3926049470901489, | |
| "learning_rate": 5.557271027436971e-05, | |
| "loss": 1.7449, | |
| "step": 1620 | |
| }, | |
| { | |
| "epoch": 0.163, | |
| "grad_norm": 0.453355997800827, | |
| "learning_rate": 5.540187460954447e-05, | |
| "loss": 1.7382, | |
| "step": 1630 | |
| }, | |
| { | |
| "epoch": 0.164, | |
| "grad_norm": 0.41926702857017517, | |
| "learning_rate": 5.523260481394348e-05, | |
| "loss": 1.7363, | |
| "step": 1640 | |
| }, | |
| { | |
| "epoch": 0.165, | |
| "grad_norm": 0.506650447845459, | |
| "learning_rate": 5.506487711171322e-05, | |
| "loss": 1.7409, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.165, | |
| "eval_loss": 1.729161024093628, | |
| "eval_runtime": 4.9479, | |
| "eval_samples_per_second": 404.214, | |
| "eval_steps_per_second": 25.263, | |
| "step": 1650 | |
| }, | |
| { | |
| "epoch": 0.166, | |
| "grad_norm": 0.45161449909210205, | |
| "learning_rate": 5.489866822936095e-05, | |
| "loss": 1.7375, | |
| "step": 1660 | |
| }, | |
| { | |
| "epoch": 0.167, | |
| "grad_norm": 0.42479032278060913, | |
| "learning_rate": 5.4733955382189484e-05, | |
| "loss": 1.6951, | |
| "step": 1670 | |
| }, | |
| { | |
| "epoch": 0.168, | |
| "grad_norm": 0.40763387084007263, | |
| "learning_rate": 5.457071626117703e-05, | |
| "loss": 1.6684, | |
| "step": 1680 | |
| }, | |
| { | |
| "epoch": 0.169, | |
| "grad_norm": 0.45459604263305664, | |
| "learning_rate": 5.440892902028488e-05, | |
| "loss": 1.6781, | |
| "step": 1690 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "grad_norm": 0.44360917806625366, | |
| "learning_rate": 5.424857226417659e-05, | |
| "loss": 1.7428, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.17, | |
| "eval_loss": 1.723348617553711, | |
| "eval_runtime": 6.9404, | |
| "eval_samples_per_second": 288.166, | |
| "eval_steps_per_second": 18.01, | |
| "step": 1700 | |
| }, | |
| { | |
| "epoch": 0.171, | |
| "grad_norm": 0.38958659768104553, | |
| "learning_rate": 5.408962503633292e-05, | |
| "loss": 1.7107, | |
| "step": 1710 | |
| }, | |
| { | |
| "epoch": 0.172, | |
| "grad_norm": 0.5742718577384949, | |
| "learning_rate": 5.39320668075478e-05, | |
| "loss": 1.7336, | |
| "step": 1720 | |
| }, | |
| { | |
| "epoch": 0.173, | |
| "grad_norm": 0.4499475359916687, | |
| "learning_rate": 5.3775877464790436e-05, | |
| "loss": 1.771, | |
| "step": 1730 | |
| }, | |
| { | |
| "epoch": 0.174, | |
| "grad_norm": 0.4566386640071869, | |
| "learning_rate": 5.362103730042052e-05, | |
| "loss": 1.7279, | |
| "step": 1740 | |
| }, | |
| { | |
| "epoch": 0.175, | |
| "grad_norm": 0.3985291123390198, | |
| "learning_rate": 5.346752700174288e-05, | |
| "loss": 1.6383, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.175, | |
| "eval_loss": 1.7142506837844849, | |
| "eval_runtime": 4.639, | |
| "eval_samples_per_second": 431.132, | |
| "eval_steps_per_second": 26.946, | |
| "step": 1750 | |
| }, | |
| { | |
| "epoch": 0.176, | |
| "grad_norm": 0.48121654987335205, | |
| "learning_rate": 5.331532764088928e-05, | |
| "loss": 1.7232, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 0.177, | |
| "grad_norm": 0.44289764761924744, | |
| "learning_rate": 5.316442066501519e-05, | |
| "loss": 1.7705, | |
| "step": 1770 | |
| }, | |
| { | |
| "epoch": 0.178, | |
| "grad_norm": 0.39872604608535767, | |
| "learning_rate": 5.30147878868001e-05, | |
| "loss": 1.7136, | |
| "step": 1780 | |
| }, | |
| { | |
| "epoch": 0.179, | |
| "grad_norm": 0.428268700838089, | |
| "learning_rate": 5.2866411475240354e-05, | |
| "loss": 1.658, | |
| "step": 1790 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "grad_norm": 0.5536085963249207, | |
| "learning_rate": 5.2719273946723746e-05, | |
| "loss": 1.6961, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.18, | |
| "eval_loss": 1.719687581062317, | |
| "eval_runtime": 4.6069, | |
| "eval_samples_per_second": 434.128, | |
| "eval_steps_per_second": 27.133, | |
| "step": 1800 | |
| }, | |
| { | |
| "epoch": 0.181, | |
| "grad_norm": 0.4429076611995697, | |
| "learning_rate": 5.257335815637598e-05, | |
| "loss": 1.7568, | |
| "step": 1810 | |
| }, | |
| { | |
| "epoch": 0.182, | |
| "grad_norm": 0.4638884961605072, | |
| "learning_rate": 5.242864728966902e-05, | |
| "loss": 1.7089, | |
| "step": 1820 | |
| }, | |
| { | |
| "epoch": 0.183, | |
| "grad_norm": 0.48520177602767944, | |
| "learning_rate": 5.2285124854282266e-05, | |
| "loss": 1.6566, | |
| "step": 1830 | |
| }, | |
| { | |
| "epoch": 0.184, | |
| "grad_norm": 0.4005000591278076, | |
| "learning_rate": 5.2142774672207326e-05, | |
| "loss": 1.6203, | |
| "step": 1840 | |
| }, | |
| { | |
| "epoch": 0.185, | |
| "grad_norm": 0.3970281183719635, | |
| "learning_rate": 5.200158087208814e-05, | |
| "loss": 1.6356, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.185, | |
| "eval_loss": 1.704606056213379, | |
| "eval_runtime": 4.6253, | |
| "eval_samples_per_second": 432.402, | |
| "eval_steps_per_second": 27.025, | |
| "step": 1850 | |
| }, | |
| { | |
| "epoch": 0.186, | |
| "grad_norm": 0.4269203245639801, | |
| "learning_rate": 5.186152788178785e-05, | |
| "loss": 1.6598, | |
| "step": 1860 | |
| }, | |
| { | |
| "epoch": 0.187, | |
| "grad_norm": 0.3638952672481537, | |
| "learning_rate": 5.172260042117486e-05, | |
| "loss": 1.6282, | |
| "step": 1870 | |
| }, | |
| { | |
| "epoch": 0.188, | |
| "grad_norm": 0.38531213998794556, | |
| "learning_rate": 5.1584783495120195e-05, | |
| "loss": 1.6547, | |
| "step": 1880 | |
| }, | |
| { | |
| "epoch": 0.189, | |
| "grad_norm": 0.4625402092933655, | |
| "learning_rate": 5.1448062386699125e-05, | |
| "loss": 1.6738, | |
| "step": 1890 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "grad_norm": 0.4621464014053345, | |
| "learning_rate": 5.1312422650589934e-05, | |
| "loss": 1.6875, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.19, | |
| "eval_loss": 1.698980689048767, | |
| "eval_runtime": 4.6172, | |
| "eval_samples_per_second": 433.159, | |
| "eval_steps_per_second": 27.072, | |
| "step": 1900 | |
| }, | |
| { | |
| "epoch": 0.191, | |
| "grad_norm": 0.44431203603744507, | |
| "learning_rate": 5.117785010666307e-05, | |
| "loss": 1.6827, | |
| "step": 1910 | |
| }, | |
| { | |
| "epoch": 0.192, | |
| "grad_norm": 0.4711743891239166, | |
| "learning_rate": 5.104433083375434e-05, | |
| "loss": 1.7471, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 0.193, | |
| "grad_norm": 0.42140287160873413, | |
| "learning_rate": 5.091185116361582e-05, | |
| "loss": 1.724, | |
| "step": 1930 | |
| }, | |
| { | |
| "epoch": 0.194, | |
| "grad_norm": 0.4742933213710785, | |
| "learning_rate": 5.0780397675038636e-05, | |
| "loss": 1.7009, | |
| "step": 1940 | |
| }, | |
| { | |
| "epoch": 0.195, | |
| "grad_norm": 0.42618677020072937, | |
| "learning_rate": 5.0649957188141786e-05, | |
| "loss": 1.6447, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.195, | |
| "eval_loss": 1.696645975112915, | |
| "eval_runtime": 4.6168, | |
| "eval_samples_per_second": 433.201, | |
| "eval_steps_per_second": 27.075, | |
| "step": 1950 | |
| }, | |
| { | |
| "epoch": 0.196, | |
| "grad_norm": 0.4423786997795105, | |
| "learning_rate": 5.052051675882142e-05, | |
| "loss": 1.6871, | |
| "step": 1960 | |
| }, | |
| { | |
| "epoch": 0.197, | |
| "grad_norm": 0.45436352491378784, | |
| "learning_rate": 5.0392063673355584e-05, | |
| "loss": 1.6682, | |
| "step": 1970 | |
| }, | |
| { | |
| "epoch": 0.198, | |
| "grad_norm": 0.43426281213760376, | |
| "learning_rate": 5.026458544315881e-05, | |
| "loss": 1.6709, | |
| "step": 1980 | |
| }, | |
| { | |
| "epoch": 0.199, | |
| "grad_norm": 0.4324936270713806, | |
| "learning_rate": 5.013806979968224e-05, | |
| "loss": 1.6365, | |
| "step": 1990 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "grad_norm": 0.4487122893333435, | |
| "learning_rate": 5.0012504689453974e-05, | |
| "loss": 1.6461, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.2, | |
| "eval_loss": 1.6922798156738281, | |
| "eval_runtime": 4.5917, | |
| "eval_samples_per_second": 435.564, | |
| "eval_steps_per_second": 27.223, | |
| "step": 2000 | |
| }, | |
| { | |
| "epoch": 0.201, | |
| "grad_norm": 0.42944806814193726, | |
| "learning_rate": 4.988787826925559e-05, | |
| "loss": 1.6658, | |
| "step": 2010 | |
| }, | |
| { | |
| "epoch": 0.202, | |
| "grad_norm": 0.5247456431388855, | |
| "learning_rate": 4.9764178901430006e-05, | |
| "loss": 1.6671, | |
| "step": 2020 | |
| }, | |
| { | |
| "epoch": 0.203, | |
| "grad_norm": 0.4485164284706116, | |
| "learning_rate": 4.9641395149316935e-05, | |
| "loss": 1.6854, | |
| "step": 2030 | |
| }, | |
| { | |
| "epoch": 0.204, | |
| "grad_norm": 0.4638156592845917, | |
| "learning_rate": 4.951951577281135e-05, | |
| "loss": 1.7201, | |
| "step": 2040 | |
| }, | |
| { | |
| "epoch": 0.205, | |
| "grad_norm": 0.42195677757263184, | |
| "learning_rate": 4.939852972404162e-05, | |
| "loss": 1.6711, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.205, | |
| "eval_loss": 1.6950935125350952, | |
| "eval_runtime": 4.6116, | |
| "eval_samples_per_second": 433.685, | |
| "eval_steps_per_second": 27.105, | |
| "step": 2050 | |
| }, | |
| { | |
| "epoch": 0.206, | |
| "grad_norm": 0.3974883556365967, | |
| "learning_rate": 4.927842614316289e-05, | |
| "loss": 1.638, | |
| "step": 2060 | |
| }, | |
| { | |
| "epoch": 0.207, | |
| "grad_norm": 0.4347898066043854, | |
| "learning_rate": 4.9159194354262706e-05, | |
| "loss": 1.6931, | |
| "step": 2070 | |
| }, | |
| { | |
| "epoch": 0.208, | |
| "grad_norm": 0.4808084964752197, | |
| "learning_rate": 4.904082386137498e-05, | |
| "loss": 1.725, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 0.209, | |
| "grad_norm": 0.48091718554496765, | |
| "learning_rate": 4.892330434459896e-05, | |
| "loss": 1.7049, | |
| "step": 2090 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "grad_norm": 0.4456811249256134, | |
| "learning_rate": 4.880662565632016e-05, | |
| "loss": 1.6487, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.21, | |
| "eval_loss": 1.688322901725769, | |
| "eval_runtime": 4.5926, | |
| "eval_samples_per_second": 435.484, | |
| "eval_steps_per_second": 27.218, | |
| "step": 2100 | |
| }, | |
| { | |
| "epoch": 0.211, | |
| "grad_norm": 0.43593478202819824, | |
| "learning_rate": 4.869077781752983e-05, | |
| "loss": 1.6993, | |
| "step": 2110 | |
| }, | |
| { | |
| "epoch": 0.212, | |
| "grad_norm": 0.4544734060764313, | |
| "learning_rate": 4.857575101424013e-05, | |
| "loss": 1.6894, | |
| "step": 2120 | |
| }, | |
| { | |
| "epoch": 0.213, | |
| "grad_norm": 0.42114976048469543, | |
| "learning_rate": 4.846153559399206e-05, | |
| "loss": 1.6745, | |
| "step": 2130 | |
| }, | |
| { | |
| "epoch": 0.214, | |
| "grad_norm": 0.4465537369251251, | |
| "learning_rate": 4.834812206245318e-05, | |
| "loss": 1.6264, | |
| "step": 2140 | |
| }, | |
| { | |
| "epoch": 0.215, | |
| "grad_norm": 0.4641367793083191, | |
| "learning_rate": 4.8235501080102624e-05, | |
| "loss": 1.7031, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.215, | |
| "eval_loss": 1.6830954551696777, | |
| "eval_runtime": 4.625, | |
| "eval_samples_per_second": 432.43, | |
| "eval_steps_per_second": 27.027, | |
| "step": 2150 | |
| }, | |
| { | |
| "epoch": 0.216, | |
| "grad_norm": 0.4330129027366638, | |
| "learning_rate": 4.812366345900056e-05, | |
| "loss": 1.6766, | |
| "step": 2160 | |
| }, | |
| { | |
| "epoch": 0.217, | |
| "grad_norm": 0.4278458058834076, | |
| "learning_rate": 4.801260015963979e-05, | |
| "loss": 1.6989, | |
| "step": 2170 | |
| }, | |
| { | |
| "epoch": 0.218, | |
| "grad_norm": 0.4446634352207184, | |
| "learning_rate": 4.790230228787671e-05, | |
| "loss": 1.6944, | |
| "step": 2180 | |
| }, | |
| { | |
| "epoch": 0.219, | |
| "grad_norm": 0.4040505588054657, | |
| "learning_rate": 4.779276109193975e-05, | |
| "loss": 1.6661, | |
| "step": 2190 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "grad_norm": 0.45129111409187317, | |
| "learning_rate": 4.7683967959512366e-05, | |
| "loss": 1.6758, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.22, | |
| "eval_loss": 1.6761962175369263, | |
| "eval_runtime": 4.6057, | |
| "eval_samples_per_second": 434.246, | |
| "eval_steps_per_second": 27.14, | |
| "step": 2200 | |
| }, | |
| { | |
| "epoch": 0.221, | |
| "grad_norm": 0.4348917007446289, | |
| "learning_rate": 4.757591441488914e-05, | |
| "loss": 1.6307, | |
| "step": 2210 | |
| }, | |
| { | |
| "epoch": 0.222, | |
| "grad_norm": 0.4746496081352234, | |
| "learning_rate": 4.746859211620213e-05, | |
| "loss": 1.6654, | |
| "step": 2220 | |
| }, | |
| { | |
| "epoch": 0.223, | |
| "grad_norm": 0.5951523184776306, | |
| "learning_rate": 4.7361992852715864e-05, | |
| "loss": 1.6478, | |
| "step": 2230 | |
| }, | |
| { | |
| "epoch": 0.224, | |
| "grad_norm": 0.4033101797103882, | |
| "learning_rate": 4.7256108542188905e-05, | |
| "loss": 1.6441, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 0.225, | |
| "grad_norm": 0.45017239451408386, | |
| "learning_rate": 4.715093122829988e-05, | |
| "loss": 1.6143, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.225, | |
| "eval_loss": 1.6731691360473633, | |
| "eval_runtime": 4.6295, | |
| "eval_samples_per_second": 432.014, | |
| "eval_steps_per_second": 27.001, | |
| "step": 2250 | |
| }, | |
| { | |
| "epoch": 0.226, | |
| "grad_norm": 0.4218330383300781, | |
| "learning_rate": 4.704645307813639e-05, | |
| "loss": 1.6687, | |
| "step": 2260 | |
| }, | |
| { | |
| "epoch": 0.227, | |
| "grad_norm": 0.41807785630226135, | |
| "learning_rate": 4.6942666379744746e-05, | |
| "loss": 1.6473, | |
| "step": 2270 | |
| }, | |
| { | |
| "epoch": 0.228, | |
| "grad_norm": 0.44153642654418945, | |
| "learning_rate": 4.683956353973895e-05, | |
| "loss": 1.634, | |
| "step": 2280 | |
| }, | |
| { | |
| "epoch": 0.229, | |
| "grad_norm": 0.4353524446487427, | |
| "learning_rate": 4.6737137080967207e-05, | |
| "loss": 1.6452, | |
| "step": 2290 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "grad_norm": 0.47118180990219116, | |
| "learning_rate": 4.663537964023428e-05, | |
| "loss": 1.657, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.23, | |
| "eval_loss": 1.6700432300567627, | |
| "eval_runtime": 4.626, | |
| "eval_samples_per_second": 432.343, | |
| "eval_steps_per_second": 27.021, | |
| "step": 2300 | |
| }, | |
| { | |
| "epoch": 0.231, | |
| "grad_norm": 0.41519850492477417, | |
| "learning_rate": 4.653428396607817e-05, | |
| "loss": 1.6589, | |
| "step": 2310 | |
| }, | |
| { | |
| "epoch": 0.232, | |
| "grad_norm": 0.5066779255867004, | |
| "learning_rate": 4.643384291659964e-05, | |
| "loss": 1.6592, | |
| "step": 2320 | |
| }, | |
| { | |
| "epoch": 0.233, | |
| "grad_norm": 0.4837518632411957, | |
| "learning_rate": 4.6334049457342925e-05, | |
| "loss": 1.6744, | |
| "step": 2330 | |
| }, | |
| { | |
| "epoch": 0.234, | |
| "grad_norm": 0.43611717224121094, | |
| "learning_rate": 4.623489665922651e-05, | |
| "loss": 1.6966, | |
| "step": 2340 | |
| }, | |
| { | |
| "epoch": 0.235, | |
| "grad_norm": 0.4694697856903076, | |
| "learning_rate": 4.613637769652221e-05, | |
| "loss": 1.6805, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.235, | |
| "eval_loss": 1.678394079208374, | |
| "eval_runtime": 4.6355, | |
| "eval_samples_per_second": 431.454, | |
| "eval_steps_per_second": 26.966, | |
| "step": 2350 | |
| }, | |
| { | |
| "epoch": 0.236, | |
| "grad_norm": 0.38386818766593933, | |
| "learning_rate": 4.603848584488156e-05, | |
| "loss": 1.666, | |
| "step": 2360 | |
| }, | |
| { | |
| "epoch": 0.237, | |
| "grad_norm": 0.4271336495876312, | |
| "learning_rate": 4.594121447940805e-05, | |
| "loss": 1.6365, | |
| "step": 2370 | |
| }, | |
| { | |
| "epoch": 0.238, | |
| "grad_norm": 0.37166059017181396, | |
| "learning_rate": 4.58445570727739e-05, | |
| "loss": 1.6179, | |
| "step": 2380 | |
| }, | |
| { | |
| "epoch": 0.239, | |
| "grad_norm": 0.40135860443115234, | |
| "learning_rate": 4.574850719338033e-05, | |
| "loss": 1.6805, | |
| "step": 2390 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "grad_norm": 0.4425022602081299, | |
| "learning_rate": 4.5653058503559986e-05, | |
| "loss": 1.6637, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.24, | |
| "eval_loss": 1.6743378639221191, | |
| "eval_runtime": 4.6179, | |
| "eval_samples_per_second": 433.097, | |
| "eval_steps_per_second": 27.069, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 0.241, | |
| "grad_norm": 0.4733668863773346, | |
| "learning_rate": 4.555820475782052e-05, | |
| "loss": 1.6514, | |
| "step": 2410 | |
| }, | |
| { | |
| "epoch": 0.242, | |
| "grad_norm": 0.4362589716911316, | |
| "learning_rate": 4.5463939801128054e-05, | |
| "loss": 1.6338, | |
| "step": 2420 | |
| }, | |
| { | |
| "epoch": 0.243, | |
| "grad_norm": 0.3984014391899109, | |
| "learning_rate": 4.537025756722961e-05, | |
| "loss": 1.5926, | |
| "step": 2430 | |
| }, | |
| { | |
| "epoch": 0.244, | |
| "grad_norm": 0.38249659538269043, | |
| "learning_rate": 4.527715207701344e-05, | |
| "loss": 1.652, | |
| "step": 2440 | |
| }, | |
| { | |
| "epoch": 0.245, | |
| "grad_norm": 0.3849375247955322, | |
| "learning_rate": 4.51846174369061e-05, | |
| "loss": 1.6031, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.245, | |
| "eval_loss": 1.6682366132736206, | |
| "eval_runtime": 4.6479, | |
| "eval_samples_per_second": 430.304, | |
| "eval_steps_per_second": 26.894, | |
| "step": 2450 | |
| }, | |
| { | |
| "epoch": 0.246, | |
| "grad_norm": 0.4070509672164917, | |
| "learning_rate": 4.5092647837305585e-05, | |
| "loss": 1.6779, | |
| "step": 2460 | |
| }, | |
| { | |
| "epoch": 0.247, | |
| "grad_norm": 0.3720800578594208, | |
| "learning_rate": 4.500123755104922e-05, | |
| "loss": 1.6363, | |
| "step": 2470 | |
| }, | |
| { | |
| "epoch": 0.248, | |
| "grad_norm": 0.4045543074607849, | |
| "learning_rate": 4.4910380931915706e-05, | |
| "loss": 1.5965, | |
| "step": 2480 | |
| }, | |
| { | |
| "epoch": 0.249, | |
| "grad_norm": 0.49813225865364075, | |
| "learning_rate": 4.4820072413160295e-05, | |
| "loss": 1.6665, | |
| "step": 2490 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "grad_norm": 0.425383985042572, | |
| "learning_rate": 4.4730306506082105e-05, | |
| "loss": 1.6569, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.25, | |
| "eval_loss": 1.6610209941864014, | |
| "eval_runtime": 4.6252, | |
| "eval_samples_per_second": 432.417, | |
| "eval_steps_per_second": 27.026, | |
| "step": 2500 | |
| }, | |
| { | |
| "epoch": 0.251, | |
| "grad_norm": 0.4529320001602173, | |
| "learning_rate": 4.4641077798623075e-05, | |
| "loss": 1.676, | |
| "step": 2510 | |
| }, | |
| { | |
| "epoch": 0.252, | |
| "grad_norm": 0.44008973240852356, | |
| "learning_rate": 4.455238095399737e-05, | |
| "loss": 1.6743, | |
| "step": 2520 | |
| }, | |
| { | |
| "epoch": 0.253, | |
| "grad_norm": 0.45680171251296997, | |
| "learning_rate": 4.446421070935069e-05, | |
| "loss": 1.6867, | |
| "step": 2530 | |
| }, | |
| { | |
| "epoch": 0.254, | |
| "grad_norm": 0.42524072527885437, | |
| "learning_rate": 4.437656187444866e-05, | |
| "loss": 1.6318, | |
| "step": 2540 | |
| }, | |
| { | |
| "epoch": 0.255, | |
| "grad_norm": 0.48738694190979004, | |
| "learning_rate": 4.428942933039357e-05, | |
| "loss": 1.6515, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.255, | |
| "eval_loss": 1.6626614332199097, | |
| "eval_runtime": 4.6447, | |
| "eval_samples_per_second": 430.596, | |
| "eval_steps_per_second": 26.912, | |
| "step": 2550 | |
| }, | |
| { | |
| "epoch": 0.256, | |
| "grad_norm": 0.4267066419124603, | |
| "learning_rate": 4.420280802836862e-05, | |
| "loss": 1.6502, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 0.257, | |
| "grad_norm": 0.37836384773254395, | |
| "learning_rate": 4.4116692988409206e-05, | |
| "loss": 1.5564, | |
| "step": 2570 | |
| }, | |
| { | |
| "epoch": 0.258, | |
| "grad_norm": 0.384313702583313, | |
| "learning_rate": 4.4031079298200316e-05, | |
| "loss": 1.6081, | |
| "step": 2580 | |
| }, | |
| { | |
| "epoch": 0.259, | |
| "grad_norm": 0.42795801162719727, | |
| "learning_rate": 4.394596211189963e-05, | |
| "loss": 1.6417, | |
| "step": 2590 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "grad_norm": 0.51594477891922, | |
| "learning_rate": 4.3861336648985394e-05, | |
| "loss": 1.6494, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.26, | |
| "eval_loss": 1.6588302850723267, | |
| "eval_runtime": 4.6233, | |
| "eval_samples_per_second": 432.588, | |
| "eval_steps_per_second": 27.037, | |
| "step": 2600 | |
| }, | |
| { | |
| "epoch": 0.261, | |
| "grad_norm": 0.42120248079299927, | |
| "learning_rate": 4.3777198193128756e-05, | |
| "loss": 1.6644, | |
| "step": 2610 | |
| }, | |
| { | |
| "epoch": 0.262, | |
| "grad_norm": 0.4563160240650177, | |
| "learning_rate": 4.369354209108969e-05, | |
| "loss": 1.6758, | |
| "step": 2620 | |
| }, | |
| { | |
| "epoch": 0.263, | |
| "grad_norm": 0.488189160823822, | |
| "learning_rate": 4.3610363751636137e-05, | |
| "loss": 1.6199, | |
| "step": 2630 | |
| }, | |
| { | |
| "epoch": 0.264, | |
| "grad_norm": 0.44422969222068787, | |
| "learning_rate": 4.352765864448559e-05, | |
| "loss": 1.6438, | |
| "step": 2640 | |
| }, | |
| { | |
| "epoch": 0.265, | |
| "grad_norm": 0.47733762860298157, | |
| "learning_rate": 4.344542229926874e-05, | |
| "loss": 1.581, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.265, | |
| "eval_loss": 1.6608359813690186, | |
| "eval_runtime": 4.628, | |
| "eval_samples_per_second": 432.154, | |
| "eval_steps_per_second": 27.01, | |
| "step": 2650 | |
| }, | |
| { | |
| "epoch": 0.266, | |
| "grad_norm": 0.41654443740844727, | |
| "learning_rate": 4.336365030451462e-05, | |
| "loss": 1.6328, | |
| "step": 2660 | |
| }, | |
| { | |
| "epoch": 0.267, | |
| "grad_norm": 0.3549344539642334, | |
| "learning_rate": 4.3282338306656564e-05, | |
| "loss": 1.5871, | |
| "step": 2670 | |
| }, | |
| { | |
| "epoch": 0.268, | |
| "grad_norm": 0.36303457617759705, | |
| "learning_rate": 4.32014820090587e-05, | |
| "loss": 1.6032, | |
| "step": 2680 | |
| }, | |
| { | |
| "epoch": 0.269, | |
| "grad_norm": 0.415372759103775, | |
| "learning_rate": 4.312107717106231e-05, | |
| "loss": 1.6015, | |
| "step": 2690 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "grad_norm": 0.3588610291481018, | |
| "learning_rate": 4.30411196070517e-05, | |
| "loss": 1.615, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.27, | |
| "eval_loss": 1.6555571556091309, | |
| "eval_runtime": 4.6266, | |
| "eval_samples_per_second": 432.287, | |
| "eval_steps_per_second": 27.018, | |
| "step": 2700 | |
| }, | |
| { | |
| "epoch": 0.271, | |
| "grad_norm": 0.47867655754089355, | |
| "learning_rate": 4.296160518553892e-05, | |
| "loss": 1.6321, | |
| "step": 2710 | |
| }, | |
| { | |
| "epoch": 0.272, | |
| "grad_norm": 0.4255579710006714, | |
| "learning_rate": 4.288252982826728e-05, | |
| "loss": 1.5484, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 0.273, | |
| "grad_norm": 0.41540971398353577, | |
| "learning_rate": 4.2803889509332595e-05, | |
| "loss": 1.6357, | |
| "step": 2730 | |
| }, | |
| { | |
| "epoch": 0.274, | |
| "grad_norm": 0.41569575667381287, | |
| "learning_rate": 4.2725680254322424e-05, | |
| "loss": 1.6814, | |
| "step": 2740 | |
| }, | |
| { | |
| "epoch": 0.275, | |
| "grad_norm": 0.40866607427597046, | |
| "learning_rate": 4.2647898139472306e-05, | |
| "loss": 1.621, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.275, | |
| "eval_loss": 1.6504429578781128, | |
| "eval_runtime": 4.639, | |
| "eval_samples_per_second": 431.128, | |
| "eval_steps_per_second": 26.945, | |
| "step": 2750 | |
| }, | |
| { | |
| "epoch": 0.276, | |
| "grad_norm": 0.3872694969177246, | |
| "learning_rate": 4.2570539290839e-05, | |
| "loss": 1.6261, | |
| "step": 2760 | |
| }, | |
| { | |
| "epoch": 0.277, | |
| "grad_norm": 0.40379494428634644, | |
| "learning_rate": 4.2493599883490024e-05, | |
| "loss": 1.6705, | |
| "step": 2770 | |
| }, | |
| { | |
| "epoch": 0.278, | |
| "grad_norm": 0.35201510787010193, | |
| "learning_rate": 4.241707614070937e-05, | |
| "loss": 1.648, | |
| "step": 2780 | |
| }, | |
| { | |
| "epoch": 0.279, | |
| "grad_norm": 0.41881263256073, | |
| "learning_rate": 4.234096433321879e-05, | |
| "loss": 1.5774, | |
| "step": 2790 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "grad_norm": 0.4631783366203308, | |
| "learning_rate": 4.226526077841448e-05, | |
| "loss": 1.6091, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.28, | |
| "eval_loss": 1.654015064239502, | |
| "eval_runtime": 4.615, | |
| "eval_samples_per_second": 433.373, | |
| "eval_steps_per_second": 27.086, | |
| "step": 2800 | |
| }, | |
| { | |
| "epoch": 0.281, | |
| "grad_norm": 0.4018985629081726, | |
| "learning_rate": 4.218996183961868e-05, | |
| "loss": 1.6026, | |
| "step": 2810 | |
| }, | |
| { | |
| "epoch": 0.282, | |
| "grad_norm": 0.36099082231521606, | |
| "learning_rate": 4.2115063925345885e-05, | |
| "loss": 1.645, | |
| "step": 2820 | |
| }, | |
| { | |
| "epoch": 0.283, | |
| "grad_norm": 0.4079434275627136, | |
| "learning_rate": 4.204056348858339e-05, | |
| "loss": 1.6135, | |
| "step": 2830 | |
| }, | |
| { | |
| "epoch": 0.284, | |
| "grad_norm": 0.4359963536262512, | |
| "learning_rate": 4.196645702608569e-05, | |
| "loss": 1.6012, | |
| "step": 2840 | |
| }, | |
| { | |
| "epoch": 0.285, | |
| "grad_norm": 0.459576815366745, | |
| "learning_rate": 4.18927410776826e-05, | |
| "loss": 1.6173, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.285, | |
| "eval_loss": 1.6534111499786377, | |
| "eval_runtime": 4.6296, | |
| "eval_samples_per_second": 432.004, | |
| "eval_steps_per_second": 27.0, | |
| "step": 2850 | |
| }, | |
| { | |
| "epoch": 0.286, | |
| "grad_norm": 0.43706053495407104, | |
| "learning_rate": 4.181941222560067e-05, | |
| "loss": 1.6155, | |
| "step": 2860 | |
| }, | |
| { | |
| "epoch": 0.287, | |
| "grad_norm": 0.3740486204624176, | |
| "learning_rate": 4.1746467093797576e-05, | |
| "loss": 1.5868, | |
| "step": 2870 | |
| }, | |
| { | |
| "epoch": 0.288, | |
| "grad_norm": 0.4161520302295685, | |
| "learning_rate": 4.167390234730933e-05, | |
| "loss": 1.5994, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 0.289, | |
| "grad_norm": 0.38721993565559387, | |
| "learning_rate": 4.1601714691609825e-05, | |
| "loss": 1.6165, | |
| "step": 2890 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "grad_norm": 0.42869701981544495, | |
| "learning_rate": 4.152990087198261e-05, | |
| "loss": 1.6065, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.29, | |
| "eval_loss": 1.6540226936340332, | |
| "eval_runtime": 4.613, | |
| "eval_samples_per_second": 433.558, | |
| "eval_steps_per_second": 27.097, | |
| "step": 2900 | |
| }, | |
| { | |
| "epoch": 0.291, | |
| "grad_norm": 0.38152533769607544, | |
| "learning_rate": 4.145845767290457e-05, | |
| "loss": 1.6351, | |
| "step": 2910 | |
| }, | |
| { | |
| "epoch": 0.292, | |
| "grad_norm": 0.5534144043922424, | |
| "learning_rate": 4.1387381917441095e-05, | |
| "loss": 1.6169, | |
| "step": 2920 | |
| }, | |
| { | |
| "epoch": 0.293, | |
| "grad_norm": 0.4238661527633667, | |
| "learning_rate": 4.131667046665284e-05, | |
| "loss": 1.6155, | |
| "step": 2930 | |
| }, | |
| { | |
| "epoch": 0.294, | |
| "grad_norm": 0.44289034605026245, | |
| "learning_rate": 4.1246320219013365e-05, | |
| "loss": 1.5717, | |
| "step": 2940 | |
| }, | |
| { | |
| "epoch": 0.295, | |
| "grad_norm": 0.3852718770503998, | |
| "learning_rate": 4.117632810983782e-05, | |
| "loss": 1.6337, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.295, | |
| "eval_loss": 1.6513147354125977, | |
| "eval_runtime": 4.6328, | |
| "eval_samples_per_second": 431.708, | |
| "eval_steps_per_second": 26.982, | |
| "step": 2950 | |
| }, | |
| { | |
| "epoch": 0.296, | |
| "grad_norm": 0.3397306501865387, | |
| "learning_rate": 4.1106691110722134e-05, | |
| "loss": 1.6018, | |
| "step": 2960 | |
| }, | |
| { | |
| "epoch": 0.297, | |
| "grad_norm": 0.35091128945350647, | |
| "learning_rate": 4.10374062289927e-05, | |
| "loss": 1.6031, | |
| "step": 2970 | |
| }, | |
| { | |
| "epoch": 0.298, | |
| "grad_norm": 0.43097996711730957, | |
| "learning_rate": 4.096847050716615e-05, | |
| "loss": 1.5879, | |
| "step": 2980 | |
| }, | |
| { | |
| "epoch": 0.299, | |
| "grad_norm": 0.41129791736602783, | |
| "learning_rate": 4.089988102241916e-05, | |
| "loss": 1.6094, | |
| "step": 2990 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "grad_norm": 0.42959263920783997, | |
| "learning_rate": 4.083163488606789e-05, | |
| "loss": 1.5706, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.3, | |
| "eval_loss": 1.6442506313323975, | |
| "eval_runtime": 4.616, | |
| "eval_samples_per_second": 433.279, | |
| "eval_steps_per_second": 27.08, | |
| "step": 3000 | |
| }, | |
| { | |
| "epoch": 0.301, | |
| "grad_norm": 0.41272443532943726, | |
| "learning_rate": 4.076372924305703e-05, | |
| "loss": 1.5555, | |
| "step": 3010 | |
| }, | |
| { | |
| "epoch": 0.302, | |
| "grad_norm": 0.402739942073822, | |
| "learning_rate": 4.0696161271458113e-05, | |
| "loss": 1.634, | |
| "step": 3020 | |
| }, | |
| { | |
| "epoch": 0.303, | |
| "grad_norm": 0.44482648372650146, | |
| "learning_rate": 4.0628928181976975e-05, | |
| "loss": 1.5485, | |
| "step": 3030 | |
| }, | |
| { | |
| "epoch": 0.304, | |
| "grad_norm": 0.4939306676387787, | |
| "learning_rate": 4.0562027217470115e-05, | |
| "loss": 1.5673, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 0.305, | |
| "grad_norm": 0.5246641635894775, | |
| "learning_rate": 4.0495455652469793e-05, | |
| "loss": 1.6093, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.305, | |
| "eval_loss": 1.6460926532745361, | |
| "eval_runtime": 4.6264, | |
| "eval_samples_per_second": 432.302, | |
| "eval_steps_per_second": 27.019, | |
| "step": 3050 | |
| }, | |
| { | |
| "epoch": 0.306, | |
| "grad_norm": 0.4141904413700104, | |
| "learning_rate": 4.0429210792717695e-05, | |
| "loss": 1.6173, | |
| "step": 3060 | |
| }, | |
| { | |
| "epoch": 0.307, | |
| "grad_norm": 0.3761964738368988, | |
| "learning_rate": 4.03632899747069e-05, | |
| "loss": 1.6303, | |
| "step": 3070 | |
| }, | |
| { | |
| "epoch": 0.308, | |
| "grad_norm": 0.3574119806289673, | |
| "learning_rate": 4.0297690565232124e-05, | |
| "loss": 1.5726, | |
| "step": 3080 | |
| }, | |
| { | |
| "epoch": 0.309, | |
| "grad_norm": 0.4022608995437622, | |
| "learning_rate": 4.0232409960947924e-05, | |
| "loss": 1.6603, | |
| "step": 3090 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "grad_norm": 0.43622887134552, | |
| "learning_rate": 4.0167445587934724e-05, | |
| "loss": 1.6398, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.31, | |
| "eval_loss": 1.6402058601379395, | |
| "eval_runtime": 4.6099, | |
| "eval_samples_per_second": 433.849, | |
| "eval_steps_per_second": 27.116, | |
| "step": 3100 | |
| }, | |
| { | |
| "epoch": 0.311, | |
| "grad_norm": 0.4243031442165375, | |
| "learning_rate": 4.0102794901272596e-05, | |
| "loss": 1.5742, | |
| "step": 3110 | |
| }, | |
| { | |
| "epoch": 0.312, | |
| "grad_norm": 0.4194463789463043, | |
| "learning_rate": 4.00384553846225e-05, | |
| "loss": 1.6243, | |
| "step": 3120 | |
| }, | |
| { | |
| "epoch": 0.313, | |
| "grad_norm": 0.3699721693992615, | |
| "learning_rate": 3.997442454981493e-05, | |
| "loss": 1.6164, | |
| "step": 3130 | |
| }, | |
| { | |
| "epoch": 0.314, | |
| "grad_norm": 0.4038933217525482, | |
| "learning_rate": 3.991069993644577e-05, | |
| "loss": 1.6501, | |
| "step": 3140 | |
| }, | |
| { | |
| "epoch": 0.315, | |
| "grad_norm": 0.44397690892219543, | |
| "learning_rate": 3.984727911147916e-05, | |
| "loss": 1.6439, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.315, | |
| "eval_loss": 1.6468923091888428, | |
| "eval_runtime": 4.6362, | |
| "eval_samples_per_second": 431.391, | |
| "eval_steps_per_second": 26.962, | |
| "step": 3150 | |
| }, | |
| { | |
| "epoch": 0.316, | |
| "grad_norm": 0.4412182569503784, | |
| "learning_rate": 3.978415966885739e-05, | |
| "loss": 1.5617, | |
| "step": 3160 | |
| }, | |
| { | |
| "epoch": 0.317, | |
| "grad_norm": 0.430493026971817, | |
| "learning_rate": 3.972133922911742e-05, | |
| "loss": 1.6315, | |
| "step": 3170 | |
| }, | |
| { | |
| "epoch": 0.318, | |
| "grad_norm": 0.44671764969825745, | |
| "learning_rate": 3.9658815439014244e-05, | |
| "loss": 1.6522, | |
| "step": 3180 | |
| }, | |
| { | |
| "epoch": 0.319, | |
| "grad_norm": 0.39204543828964233, | |
| "learning_rate": 3.959658597115049e-05, | |
| "loss": 1.5415, | |
| "step": 3190 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "grad_norm": 0.38203272223472595, | |
| "learning_rate": 3.953464852361267e-05, | |
| "loss": 1.5773, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.32, | |
| "eval_loss": 1.6403597593307495, | |
| "eval_runtime": 4.6223, | |
| "eval_samples_per_second": 432.689, | |
| "eval_steps_per_second": 27.043, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 0.321, | |
| "grad_norm": 0.46437588334083557, | |
| "learning_rate": 3.9473000819613454e-05, | |
| "loss": 1.6081, | |
| "step": 3210 | |
| }, | |
| { | |
| "epoch": 0.322, | |
| "grad_norm": 0.38332080841064453, | |
| "learning_rate": 3.941164060714018e-05, | |
| "loss": 1.5312, | |
| "step": 3220 | |
| }, | |
| { | |
| "epoch": 0.323, | |
| "grad_norm": 0.41572943329811096, | |
| "learning_rate": 3.9350565658609336e-05, | |
| "loss": 1.5595, | |
| "step": 3230 | |
| }, | |
| { | |
| "epoch": 0.324, | |
| "grad_norm": 0.3750537931919098, | |
| "learning_rate": 3.928977377052684e-05, | |
| "loss": 1.626, | |
| "step": 3240 | |
| }, | |
| { | |
| "epoch": 0.325, | |
| "grad_norm": 0.5018735527992249, | |
| "learning_rate": 3.922926276315421e-05, | |
| "loss": 1.6085, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.325, | |
| "eval_loss": 1.6348525285720825, | |
| "eval_runtime": 4.6293, | |
| "eval_samples_per_second": 432.031, | |
| "eval_steps_per_second": 27.002, | |
| "step": 3250 | |
| }, | |
| { | |
| "epoch": 0.326, | |
| "grad_norm": 0.3808193802833557, | |
| "learning_rate": 3.916903048018023e-05, | |
| "loss": 1.6246, | |
| "step": 3260 | |
| }, | |
| { | |
| "epoch": 0.327, | |
| "grad_norm": 0.34066709876060486, | |
| "learning_rate": 3.910907478839825e-05, | |
| "loss": 1.5548, | |
| "step": 3270 | |
| }, | |
| { | |
| "epoch": 0.328, | |
| "grad_norm": 0.38990190625190735, | |
| "learning_rate": 3.904939357738885e-05, | |
| "loss": 1.5419, | |
| "step": 3280 | |
| }, | |
| { | |
| "epoch": 0.329, | |
| "grad_norm": 0.45460209250450134, | |
| "learning_rate": 3.8989984759207835e-05, | |
| "loss": 1.5897, | |
| "step": 3290 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "grad_norm": 0.3932059407234192, | |
| "learning_rate": 3.893084626807942e-05, | |
| "loss": 1.6439, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.33, | |
| "eval_loss": 1.6305333375930786, | |
| "eval_runtime": 4.6169, | |
| "eval_samples_per_second": 433.188, | |
| "eval_steps_per_second": 27.074, | |
| "step": 3300 | |
| }, | |
| { | |
| "epoch": 0.331, | |
| "grad_norm": 0.3681378662586212, | |
| "learning_rate": 3.887197606009451e-05, | |
| "loss": 1.6234, | |
| "step": 3310 | |
| }, | |
| { | |
| "epoch": 0.332, | |
| "grad_norm": 0.4306110739707947, | |
| "learning_rate": 3.881337211291406e-05, | |
| "loss": 1.5888, | |
| "step": 3320 | |
| }, | |
| { | |
| "epoch": 0.333, | |
| "grad_norm": 0.43308112025260925, | |
| "learning_rate": 3.875503242547721e-05, | |
| "loss": 1.6162, | |
| "step": 3330 | |
| }, | |
| { | |
| "epoch": 0.334, | |
| "grad_norm": 0.3679511547088623, | |
| "learning_rate": 3.869695501771434e-05, | |
| "loss": 1.5796, | |
| "step": 3340 | |
| }, | |
| { | |
| "epoch": 0.335, | |
| "grad_norm": 0.4424503743648529, | |
| "learning_rate": 3.8639137930264805e-05, | |
| "loss": 1.591, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.335, | |
| "eval_loss": 1.6320486068725586, | |
| "eval_runtime": 4.6344, | |
| "eval_samples_per_second": 431.558, | |
| "eval_steps_per_second": 26.972, | |
| "step": 3350 | |
| }, | |
| { | |
| "epoch": 0.336, | |
| "grad_norm": 0.4127017557621002, | |
| "learning_rate": 3.858157922419927e-05, | |
| "loss": 1.6142, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 0.337, | |
| "grad_norm": 0.516035795211792, | |
| "learning_rate": 3.8524276980746626e-05, | |
| "loss": 1.6361, | |
| "step": 3370 | |
| }, | |
| { | |
| "epoch": 0.338, | |
| "grad_norm": 0.35583969950675964, | |
| "learning_rate": 3.8467229301025334e-05, | |
| "loss": 1.6123, | |
| "step": 3380 | |
| }, | |
| { | |
| "epoch": 0.339, | |
| "grad_norm": 0.40070992708206177, | |
| "learning_rate": 3.8410434305779104e-05, | |
| "loss": 1.5668, | |
| "step": 3390 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "grad_norm": 0.4105798900127411, | |
| "learning_rate": 3.835389013511689e-05, | |
| "loss": 1.5962, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.34, | |
| "eval_loss": 1.629593014717102, | |
| "eval_runtime": 4.6284, | |
| "eval_samples_per_second": 432.113, | |
| "eval_steps_per_second": 27.007, | |
| "step": 3400 | |
| }, | |
| { | |
| "epoch": 0.341, | |
| "grad_norm": 0.4205196797847748, | |
| "learning_rate": 3.829759494825711e-05, | |
| "loss": 1.5301, | |
| "step": 3410 | |
| }, | |
| { | |
| "epoch": 0.342, | |
| "grad_norm": 0.39605259895324707, | |
| "learning_rate": 3.82415469232759e-05, | |
| "loss": 1.6035, | |
| "step": 3420 | |
| }, | |
| { | |
| "epoch": 0.343, | |
| "grad_norm": 0.3812793493270874, | |
| "learning_rate": 3.8185744256859485e-05, | |
| "loss": 1.6069, | |
| "step": 3430 | |
| }, | |
| { | |
| "epoch": 0.344, | |
| "grad_norm": 0.3666427433490753, | |
| "learning_rate": 3.813018516406046e-05, | |
| "loss": 1.5852, | |
| "step": 3440 | |
| }, | |
| { | |
| "epoch": 0.345, | |
| "grad_norm": 0.4229329526424408, | |
| "learning_rate": 3.807486787805802e-05, | |
| "loss": 1.5569, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.345, | |
| "eval_loss": 1.6314581632614136, | |
| "eval_runtime": 4.6341, | |
| "eval_samples_per_second": 431.582, | |
| "eval_steps_per_second": 26.974, | |
| "step": 3450 | |
| }, | |
| { | |
| "epoch": 0.346, | |
| "grad_norm": 0.41742926836013794, | |
| "learning_rate": 3.801979064992194e-05, | |
| "loss": 1.5947, | |
| "step": 3460 | |
| }, | |
| { | |
| "epoch": 0.347, | |
| "grad_norm": 0.3738185465335846, | |
| "learning_rate": 3.796495174838033e-05, | |
| "loss": 1.5678, | |
| "step": 3470 | |
| }, | |
| { | |
| "epoch": 0.348, | |
| "grad_norm": 0.3990820348262787, | |
| "learning_rate": 3.7910349459591094e-05, | |
| "loss": 1.5846, | |
| "step": 3480 | |
| }, | |
| { | |
| "epoch": 0.349, | |
| "grad_norm": 0.3688414394855499, | |
| "learning_rate": 3.785598208691693e-05, | |
| "loss": 1.5944, | |
| "step": 3490 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "grad_norm": 0.4192740023136139, | |
| "learning_rate": 3.780184795070394e-05, | |
| "loss": 1.5938, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.35, | |
| "eval_loss": 1.6295636892318726, | |
| "eval_runtime": 4.6346, | |
| "eval_samples_per_second": 431.535, | |
| "eval_steps_per_second": 26.971, | |
| "step": 3500 | |
| }, | |
| { | |
| "epoch": 0.351, | |
| "grad_norm": 0.3593446910381317, | |
| "learning_rate": 3.7747945388063626e-05, | |
| "loss": 1.5717, | |
| "step": 3510 | |
| }, | |
| { | |
| "epoch": 0.352, | |
| "grad_norm": 0.4086902439594269, | |
| "learning_rate": 3.7694272752658386e-05, | |
| "loss": 1.5605, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 0.353, | |
| "grad_norm": 0.3733591139316559, | |
| "learning_rate": 3.764082841449032e-05, | |
| "loss": 1.6089, | |
| "step": 3530 | |
| }, | |
| { | |
| "epoch": 0.354, | |
| "grad_norm": 0.3657650947570801, | |
| "learning_rate": 3.758761075969328e-05, | |
| "loss": 1.5542, | |
| "step": 3540 | |
| }, | |
| { | |
| "epoch": 0.355, | |
| "grad_norm": 0.3745919466018677, | |
| "learning_rate": 3.7534618190328195e-05, | |
| "loss": 1.5384, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.355, | |
| "eval_loss": 1.6273561716079712, | |
| "eval_runtime": 4.6393, | |
| "eval_samples_per_second": 431.096, | |
| "eval_steps_per_second": 26.943, | |
| "step": 3550 | |
| }, | |
| { | |
| "epoch": 0.356, | |
| "grad_norm": 0.36409613490104675, | |
| "learning_rate": 3.748184912418159e-05, | |
| "loss": 1.5235, | |
| "step": 3560 | |
| }, | |
| { | |
| "epoch": 0.357, | |
| "grad_norm": 0.4654470980167389, | |
| "learning_rate": 3.742930199456709e-05, | |
| "loss": 1.5894, | |
| "step": 3570 | |
| }, | |
| { | |
| "epoch": 0.358, | |
| "grad_norm": 0.38284242153167725, | |
| "learning_rate": 3.7376975250130133e-05, | |
| "loss": 1.6203, | |
| "step": 3580 | |
| }, | |
| { | |
| "epoch": 0.359, | |
| "grad_norm": 0.3441823422908783, | |
| "learning_rate": 3.732486735465553e-05, | |
| "loss": 1.5614, | |
| "step": 3590 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "grad_norm": 0.36308762431144714, | |
| "learning_rate": 3.727297678687811e-05, | |
| "loss": 1.5292, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.36, | |
| "eval_loss": 1.6230641603469849, | |
| "eval_runtime": 4.6201, | |
| "eval_samples_per_second": 432.889, | |
| "eval_steps_per_second": 27.056, | |
| "step": 3600 | |
| }, | |
| { | |
| "epoch": 0.361, | |
| "grad_norm": 0.3441646099090576, | |
| "learning_rate": 3.7221302040296116e-05, | |
| "loss": 1.5294, | |
| "step": 3610 | |
| }, | |
| { | |
| "epoch": 0.362, | |
| "grad_norm": 0.3621971905231476, | |
| "learning_rate": 3.7169841622987545e-05, | |
| "loss": 1.5199, | |
| "step": 3620 | |
| }, | |
| { | |
| "epoch": 0.363, | |
| "grad_norm": 0.375041127204895, | |
| "learning_rate": 3.711859405742924e-05, | |
| "loss": 1.5561, | |
| "step": 3630 | |
| }, | |
| { | |
| "epoch": 0.364, | |
| "grad_norm": 0.4329853355884552, | |
| "learning_rate": 3.706755788031864e-05, | |
| "loss": 1.5937, | |
| "step": 3640 | |
| }, | |
| { | |
| "epoch": 0.365, | |
| "grad_norm": 0.394687294960022, | |
| "learning_rate": 3.701673164239839e-05, | |
| "loss": 1.5165, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.365, | |
| "eval_loss": 1.623963475227356, | |
| "eval_runtime": 4.6333, | |
| "eval_samples_per_second": 431.655, | |
| "eval_steps_per_second": 26.978, | |
| "step": 3650 | |
| }, | |
| { | |
| "epoch": 0.366, | |
| "grad_norm": 0.394584059715271, | |
| "learning_rate": 3.696611390828343e-05, | |
| "loss": 1.5693, | |
| "step": 3660 | |
| }, | |
| { | |
| "epoch": 0.367, | |
| "grad_norm": 0.39215362071990967, | |
| "learning_rate": 3.691570325629073e-05, | |
| "loss": 1.5818, | |
| "step": 3670 | |
| }, | |
| { | |
| "epoch": 0.368, | |
| "grad_norm": 0.3797973394393921, | |
| "learning_rate": 3.6865498278271596e-05, | |
| "loss": 1.5617, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 0.369, | |
| "grad_norm": 0.39166536927223206, | |
| "learning_rate": 3.681549757944646e-05, | |
| "loss": 1.6029, | |
| "step": 3690 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "grad_norm": 0.4009022116661072, | |
| "learning_rate": 3.676569977824213e-05, | |
| "loss": 1.584, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.37, | |
| "eval_loss": 1.6209263801574707, | |
| "eval_runtime": 4.6099, | |
| "eval_samples_per_second": 433.849, | |
| "eval_steps_per_second": 27.116, | |
| "step": 3700 | |
| }, | |
| { | |
| "epoch": 0.371, | |
| "grad_norm": 0.36677446961402893, | |
| "learning_rate": 3.6716103506131446e-05, | |
| "loss": 1.5531, | |
| "step": 3710 | |
| }, | |
| { | |
| "epoch": 0.372, | |
| "grad_norm": 0.42492184042930603, | |
| "learning_rate": 3.6666707407475314e-05, | |
| "loss": 1.5946, | |
| "step": 3720 | |
| }, | |
| { | |
| "epoch": 0.373, | |
| "grad_norm": 0.41753894090652466, | |
| "learning_rate": 3.661751013936708e-05, | |
| "loss": 1.5885, | |
| "step": 3730 | |
| }, | |
| { | |
| "epoch": 0.374, | |
| "grad_norm": 0.3767106235027313, | |
| "learning_rate": 3.656851037147919e-05, | |
| "loss": 1.6074, | |
| "step": 3740 | |
| }, | |
| { | |
| "epoch": 0.375, | |
| "grad_norm": 0.46671557426452637, | |
| "learning_rate": 3.65197067859121e-05, | |
| "loss": 1.5997, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.375, | |
| "eval_loss": 1.6205507516860962, | |
| "eval_runtime": 4.6338, | |
| "eval_samples_per_second": 431.614, | |
| "eval_steps_per_second": 26.976, | |
| "step": 3750 | |
| }, | |
| { | |
| "epoch": 0.376, | |
| "grad_norm": 0.40476033091545105, | |
| "learning_rate": 3.6471098077045404e-05, | |
| "loss": 1.5884, | |
| "step": 3760 | |
| }, | |
| { | |
| "epoch": 0.377, | |
| "grad_norm": 0.4570123851299286, | |
| "learning_rate": 3.642268295139107e-05, | |
| "loss": 1.5792, | |
| "step": 3770 | |
| }, | |
| { | |
| "epoch": 0.378, | |
| "grad_norm": 0.4356667101383209, | |
| "learning_rate": 3.6374460127448995e-05, | |
| "loss": 1.5642, | |
| "step": 3780 | |
| }, | |
| { | |
| "epoch": 0.379, | |
| "grad_norm": 0.43170198798179626, | |
| "learning_rate": 3.632642833556441e-05, | |
| "loss": 1.5847, | |
| "step": 3790 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "grad_norm": 0.5624451637268066, | |
| "learning_rate": 3.627858631778756e-05, | |
| "loss": 1.5819, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.38, | |
| "eval_loss": 1.6164690256118774, | |
| "eval_runtime": 4.6091, | |
| "eval_samples_per_second": 433.925, | |
| "eval_steps_per_second": 27.12, | |
| "step": 3800 | |
| }, | |
| { | |
| "epoch": 0.381, | |
| "grad_norm": 0.41750478744506836, | |
| "learning_rate": 3.623093282773527e-05, | |
| "loss": 1.6101, | |
| "step": 3810 | |
| }, | |
| { | |
| "epoch": 0.382, | |
| "grad_norm": 0.41956469416618347, | |
| "learning_rate": 3.6183466630454636e-05, | |
| "loss": 1.5962, | |
| "step": 3820 | |
| }, | |
| { | |
| "epoch": 0.383, | |
| "grad_norm": 0.409078985452652, | |
| "learning_rate": 3.6136186502288535e-05, | |
| "loss": 1.5791, | |
| "step": 3830 | |
| }, | |
| { | |
| "epoch": 0.384, | |
| "grad_norm": 0.3656834661960602, | |
| "learning_rate": 3.608909123074314e-05, | |
| "loss": 1.5996, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 0.385, | |
| "grad_norm": 0.46695512533187866, | |
| "learning_rate": 3.6042179614357375e-05, | |
| "loss": 1.6295, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.385, | |
| "eval_loss": 1.6181100606918335, | |
| "eval_runtime": 4.6344, | |
| "eval_samples_per_second": 431.56, | |
| "eval_steps_per_second": 26.972, | |
| "step": 3850 | |
| }, | |
| { | |
| "epoch": 0.386, | |
| "grad_norm": 0.4372875392436981, | |
| "learning_rate": 3.5995450462574126e-05, | |
| "loss": 1.5875, | |
| "step": 3860 | |
| }, | |
| { | |
| "epoch": 0.387, | |
| "grad_norm": 0.4402835965156555, | |
| "learning_rate": 3.594890259561335e-05, | |
| "loss": 1.6344, | |
| "step": 3870 | |
| }, | |
| { | |
| "epoch": 0.388, | |
| "grad_norm": 0.4220004677772522, | |
| "learning_rate": 3.5902534844346976e-05, | |
| "loss": 1.6147, | |
| "step": 3880 | |
| }, | |
| { | |
| "epoch": 0.389, | |
| "grad_norm": 0.3918090760707855, | |
| "learning_rate": 3.5856346050175565e-05, | |
| "loss": 1.4967, | |
| "step": 3890 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "grad_norm": 0.3966856598854065, | |
| "learning_rate": 3.581033506490671e-05, | |
| "loss": 1.5349, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.39, | |
| "eval_loss": 1.6230711936950684, | |
| "eval_runtime": 4.6195, | |
| "eval_samples_per_second": 432.948, | |
| "eval_steps_per_second": 27.059, | |
| "step": 3900 | |
| }, | |
| { | |
| "epoch": 0.391, | |
| "grad_norm": 0.38468602299690247, | |
| "learning_rate": 3.576450075063519e-05, | |
| "loss": 1.5366, | |
| "step": 3910 | |
| }, | |
| { | |
| "epoch": 0.392, | |
| "grad_norm": 0.39990097284317017, | |
| "learning_rate": 3.571884197962469e-05, | |
| "loss": 1.5028, | |
| "step": 3920 | |
| }, | |
| { | |
| "epoch": 0.393, | |
| "grad_norm": 0.3719644248485565, | |
| "learning_rate": 3.567335763419138e-05, | |
| "loss": 1.5696, | |
| "step": 3930 | |
| }, | |
| { | |
| "epoch": 0.394, | |
| "grad_norm": 0.40582171082496643, | |
| "learning_rate": 3.562804660658888e-05, | |
| "loss": 1.5883, | |
| "step": 3940 | |
| }, | |
| { | |
| "epoch": 0.395, | |
| "grad_norm": 0.387210488319397, | |
| "learning_rate": 3.5582907798895035e-05, | |
| "loss": 1.561, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.395, | |
| "eval_loss": 1.6173763275146484, | |
| "eval_runtime": 4.6286, | |
| "eval_samples_per_second": 432.1, | |
| "eval_steps_per_second": 27.006, | |
| "step": 3950 | |
| }, | |
| { | |
| "epoch": 0.396, | |
| "grad_norm": 0.39436179399490356, | |
| "learning_rate": 3.55379401229001e-05, | |
| "loss": 1.527, | |
| "step": 3960 | |
| }, | |
| { | |
| "epoch": 0.397, | |
| "grad_norm": 0.37865695357322693, | |
| "learning_rate": 3.5493142499996666e-05, | |
| "loss": 1.5324, | |
| "step": 3970 | |
| }, | |
| { | |
| "epoch": 0.398, | |
| "grad_norm": 0.3817419409751892, | |
| "learning_rate": 3.544851386107085e-05, | |
| "loss": 1.5975, | |
| "step": 3980 | |
| }, | |
| { | |
| "epoch": 0.399, | |
| "grad_norm": 0.3955094814300537, | |
| "learning_rate": 3.540405314639526e-05, | |
| "loss": 1.5799, | |
| "step": 3990 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "grad_norm": 0.432600200176239, | |
| "learning_rate": 3.535975930552322e-05, | |
| "loss": 1.5693, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.4, | |
| "eval_loss": 1.6128498315811157, | |
| "eval_runtime": 4.6082, | |
| "eval_samples_per_second": 434.006, | |
| "eval_steps_per_second": 27.125, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 0.401, | |
| "grad_norm": 0.42284053564071655, | |
| "learning_rate": 3.531563129718458e-05, | |
| "loss": 1.5441, | |
| "step": 4010 | |
| }, | |
| { | |
| "epoch": 0.402, | |
| "grad_norm": 0.36951157450675964, | |
| "learning_rate": 3.527166808918287e-05, | |
| "loss": 1.5692, | |
| "step": 4020 | |
| }, | |
| { | |
| "epoch": 0.403, | |
| "grad_norm": 0.4210173785686493, | |
| "learning_rate": 3.522786865829391e-05, | |
| "loss": 1.5176, | |
| "step": 4030 | |
| }, | |
| { | |
| "epoch": 0.404, | |
| "grad_norm": 0.491728276014328, | |
| "learning_rate": 3.51842319901658e-05, | |
| "loss": 1.56, | |
| "step": 4040 | |
| }, | |
| { | |
| "epoch": 0.405, | |
| "grad_norm": 0.4052882492542267, | |
| "learning_rate": 3.5140757079220216e-05, | |
| "loss": 1.56, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.405, | |
| "eval_loss": 1.6166870594024658, | |
| "eval_runtime": 4.6429, | |
| "eval_samples_per_second": 430.77, | |
| "eval_steps_per_second": 26.923, | |
| "step": 4050 | |
| }, | |
| { | |
| "epoch": 0.406, | |
| "grad_norm": 0.3819957375526428, | |
| "learning_rate": 3.509744292855511e-05, | |
| "loss": 1.5628, | |
| "step": 4060 | |
| }, | |
| { | |
| "epoch": 0.407, | |
| "grad_norm": 0.4066322147846222, | |
| "learning_rate": 3.505428854984869e-05, | |
| "loss": 1.6145, | |
| "step": 4070 | |
| }, | |
| { | |
| "epoch": 0.408, | |
| "grad_norm": 0.328678697347641, | |
| "learning_rate": 3.5011292963264705e-05, | |
| "loss": 1.6075, | |
| "step": 4080 | |
| }, | |
| { | |
| "epoch": 0.409, | |
| "grad_norm": 0.3642025291919708, | |
| "learning_rate": 3.496845519735901e-05, | |
| "loss": 1.499, | |
| "step": 4090 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "grad_norm": 0.40798667073249817, | |
| "learning_rate": 3.492577428898734e-05, | |
| "loss": 1.5397, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.41, | |
| "eval_loss": 1.6112936735153198, | |
| "eval_runtime": 4.6258, | |
| "eval_samples_per_second": 432.359, | |
| "eval_steps_per_second": 27.022, | |
| "step": 4100 | |
| }, | |
| { | |
| "epoch": 0.411, | |
| "grad_norm": 0.40259620547294617, | |
| "learning_rate": 3.4883249283214416e-05, | |
| "loss": 1.606, | |
| "step": 4110 | |
| }, | |
| { | |
| "epoch": 0.412, | |
| "grad_norm": 0.41489359736442566, | |
| "learning_rate": 3.4840879233224196e-05, | |
| "loss": 1.5389, | |
| "step": 4120 | |
| }, | |
| { | |
| "epoch": 0.413, | |
| "grad_norm": 0.39256951212882996, | |
| "learning_rate": 3.4798663200231265e-05, | |
| "loss": 1.554, | |
| "step": 4130 | |
| }, | |
| { | |
| "epoch": 0.414, | |
| "grad_norm": 0.3881060779094696, | |
| "learning_rate": 3.475660025339355e-05, | |
| "loss": 1.5654, | |
| "step": 4140 | |
| }, | |
| { | |
| "epoch": 0.415, | |
| "grad_norm": 0.4276053011417389, | |
| "learning_rate": 3.471468946972612e-05, | |
| "loss": 1.5292, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.415, | |
| "eval_loss": 1.6163493394851685, | |
| "eval_runtime": 4.642, | |
| "eval_samples_per_second": 430.846, | |
| "eval_steps_per_second": 26.928, | |
| "step": 4150 | |
| }, | |
| { | |
| "epoch": 0.416, | |
| "grad_norm": 0.38684210181236267, | |
| "learning_rate": 3.467292993401603e-05, | |
| "loss": 1.5643, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 0.417, | |
| "grad_norm": 0.39144203066825867, | |
| "learning_rate": 3.4631320738738494e-05, | |
| "loss": 1.5399, | |
| "step": 4170 | |
| }, | |
| { | |
| "epoch": 0.418, | |
| "grad_norm": 0.39135465025901794, | |
| "learning_rate": 3.458986098397395e-05, | |
| "loss": 1.5646, | |
| "step": 4180 | |
| }, | |
| { | |
| "epoch": 0.419, | |
| "grad_norm": 0.44311603903770447, | |
| "learning_rate": 3.45485497773264e-05, | |
| "loss": 1.5739, | |
| "step": 4190 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "grad_norm": 0.45056262612342834, | |
| "learning_rate": 3.450738623384265e-05, | |
| "loss": 1.6048, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.42, | |
| "eval_loss": 1.6106904745101929, | |
| "eval_runtime": 4.616, | |
| "eval_samples_per_second": 433.279, | |
| "eval_steps_per_second": 27.08, | |
| "step": 4200 | |
| }, | |
| { | |
| "epoch": 0.421, | |
| "grad_norm": 0.37200435996055603, | |
| "learning_rate": 3.4466369475932744e-05, | |
| "loss": 1.5346, | |
| "step": 4210 | |
| }, | |
| { | |
| "epoch": 0.422, | |
| "grad_norm": 0.3999169170856476, | |
| "learning_rate": 3.442549863329138e-05, | |
| "loss": 1.5593, | |
| "step": 4220 | |
| }, | |
| { | |
| "epoch": 0.423, | |
| "grad_norm": 0.38607126474380493, | |
| "learning_rate": 3.438477284282031e-05, | |
| "loss": 1.557, | |
| "step": 4230 | |
| }, | |
| { | |
| "epoch": 0.424, | |
| "grad_norm": 0.4031635820865631, | |
| "learning_rate": 3.4344191248551814e-05, | |
| "loss": 1.5875, | |
| "step": 4240 | |
| }, | |
| { | |
| "epoch": 0.425, | |
| "grad_norm": 0.37006083130836487, | |
| "learning_rate": 3.4303753001573164e-05, | |
| "loss": 1.576, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.425, | |
| "eval_loss": 1.612293004989624, | |
| "eval_runtime": 4.6366, | |
| "eval_samples_per_second": 431.347, | |
| "eval_steps_per_second": 26.959, | |
| "step": 4250 | |
| }, | |
| { | |
| "epoch": 0.426, | |
| "grad_norm": 0.37665873765945435, | |
| "learning_rate": 3.426345725995197e-05, | |
| "loss": 1.5487, | |
| "step": 4260 | |
| }, | |
| { | |
| "epoch": 0.427, | |
| "grad_norm": 0.4067043662071228, | |
| "learning_rate": 3.422330318866262e-05, | |
| "loss": 1.5885, | |
| "step": 4270 | |
| }, | |
| { | |
| "epoch": 0.428, | |
| "grad_norm": 0.37107494473457336, | |
| "learning_rate": 3.4183289959513575e-05, | |
| "loss": 1.5542, | |
| "step": 4280 | |
| }, | |
| { | |
| "epoch": 0.429, | |
| "grad_norm": 0.35898610949516296, | |
| "learning_rate": 3.414341675107563e-05, | |
| "loss": 1.5508, | |
| "step": 4290 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "grad_norm": 0.3901163339614868, | |
| "learning_rate": 3.4103682748611054e-05, | |
| "loss": 1.6198, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.43, | |
| "eval_loss": 1.6109461784362793, | |
| "eval_runtime": 4.6346, | |
| "eval_samples_per_second": 431.536, | |
| "eval_steps_per_second": 26.971, | |
| "step": 4300 | |
| }, | |
| { | |
| "epoch": 0.431, | |
| "grad_norm": 0.41391491889953613, | |
| "learning_rate": 3.4064087144003755e-05, | |
| "loss": 1.5761, | |
| "step": 4310 | |
| }, | |
| { | |
| "epoch": 0.432, | |
| "grad_norm": 0.4450955390930176, | |
| "learning_rate": 3.402462913569015e-05, | |
| "loss": 1.5615, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 0.433, | |
| "grad_norm": 0.44352859258651733, | |
| "learning_rate": 3.3985307928591074e-05, | |
| "loss": 1.5776, | |
| "step": 4330 | |
| }, | |
| { | |
| "epoch": 0.434, | |
| "grad_norm": 0.3886057436466217, | |
| "learning_rate": 3.3946122734044455e-05, | |
| "loss": 1.5866, | |
| "step": 4340 | |
| }, | |
| { | |
| "epoch": 0.435, | |
| "grad_norm": 0.36949875950813293, | |
| "learning_rate": 3.390707276973892e-05, | |
| "loss": 1.5043, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.435, | |
| "eval_loss": 1.6040953397750854, | |
| "eval_runtime": 4.651, | |
| "eval_samples_per_second": 430.016, | |
| "eval_steps_per_second": 26.876, | |
| "step": 4350 | |
| }, | |
| { | |
| "epoch": 0.436, | |
| "grad_norm": 0.37118640542030334, | |
| "learning_rate": 3.3868157259648156e-05, | |
| "loss": 1.55, | |
| "step": 4360 | |
| }, | |
| { | |
| "epoch": 0.437, | |
| "grad_norm": 0.3654156029224396, | |
| "learning_rate": 3.382937543396614e-05, | |
| "loss": 1.5805, | |
| "step": 4370 | |
| }, | |
| { | |
| "epoch": 0.438, | |
| "grad_norm": 0.36356887221336365, | |
| "learning_rate": 3.379072652904321e-05, | |
| "loss": 1.5733, | |
| "step": 4380 | |
| }, | |
| { | |
| "epoch": 0.439, | |
| "grad_norm": 0.3553885519504547, | |
| "learning_rate": 3.375220978732289e-05, | |
| "loss": 1.5236, | |
| "step": 4390 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "grad_norm": 0.33554360270500183, | |
| "learning_rate": 3.371382445727951e-05, | |
| "loss": 1.4966, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.44, | |
| "eval_loss": 1.6084920167922974, | |
| "eval_runtime": 4.6233, | |
| "eval_samples_per_second": 432.591, | |
| "eval_steps_per_second": 27.037, | |
| "step": 4400 | |
| }, | |
| { | |
| "epoch": 0.441, | |
| "grad_norm": 0.5069860816001892, | |
| "learning_rate": 3.3675569793356654e-05, | |
| "loss": 1.5297, | |
| "step": 4410 | |
| }, | |
| { | |
| "epoch": 0.442, | |
| "grad_norm": 0.4153578579425812, | |
| "learning_rate": 3.36374450559063e-05, | |
| "loss": 1.6225, | |
| "step": 4420 | |
| }, | |
| { | |
| "epoch": 0.443, | |
| "grad_norm": 0.4991408884525299, | |
| "learning_rate": 3.359944951112878e-05, | |
| "loss": 1.5769, | |
| "step": 4430 | |
| }, | |
| { | |
| "epoch": 0.444, | |
| "grad_norm": 0.38914990425109863, | |
| "learning_rate": 3.356158243101345e-05, | |
| "loss": 1.5169, | |
| "step": 4440 | |
| }, | |
| { | |
| "epoch": 0.445, | |
| "grad_norm": 0.38475480675697327, | |
| "learning_rate": 3.3523843093280096e-05, | |
| "loss": 1.532, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.445, | |
| "eval_loss": 1.6051650047302246, | |
| "eval_runtime": 4.6208, | |
| "eval_samples_per_second": 432.824, | |
| "eval_steps_per_second": 27.051, | |
| "step": 4450 | |
| }, | |
| { | |
| "epoch": 0.446, | |
| "grad_norm": 0.3508313000202179, | |
| "learning_rate": 3.348623078132111e-05, | |
| "loss": 1.5924, | |
| "step": 4460 | |
| }, | |
| { | |
| "epoch": 0.447, | |
| "grad_norm": 0.3878994286060333, | |
| "learning_rate": 3.344874478414426e-05, | |
| "loss": 1.5773, | |
| "step": 4470 | |
| }, | |
| { | |
| "epoch": 0.448, | |
| "grad_norm": 0.39611005783081055, | |
| "learning_rate": 3.341138439631633e-05, | |
| "loss": 1.5787, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 0.449, | |
| "grad_norm": 0.37631756067276, | |
| "learning_rate": 3.337414891790731e-05, | |
| "loss": 1.4992, | |
| "step": 4490 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "grad_norm": 0.35049423575401306, | |
| "learning_rate": 3.3337037654435325e-05, | |
| "loss": 1.5761, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.45, | |
| "eval_loss": 1.6108379364013672, | |
| "eval_runtime": 4.6124, | |
| "eval_samples_per_second": 433.615, | |
| "eval_steps_per_second": 27.101, | |
| "step": 4500 | |
| }, | |
| { | |
| "epoch": 0.451, | |
| "grad_norm": 0.41320350766181946, | |
| "learning_rate": 3.330004991681224e-05, | |
| "loss": 1.5535, | |
| "step": 4510 | |
| }, | |
| { | |
| "epoch": 0.452, | |
| "grad_norm": 0.40881311893463135, | |
| "learning_rate": 3.3263185021289925e-05, | |
| "loss": 1.5806, | |
| "step": 4520 | |
| }, | |
| { | |
| "epoch": 0.453, | |
| "grad_norm": 0.3850957453250885, | |
| "learning_rate": 3.322644228940717e-05, | |
| "loss": 1.585, | |
| "step": 4530 | |
| }, | |
| { | |
| "epoch": 0.454, | |
| "grad_norm": 0.36344847083091736, | |
| "learning_rate": 3.318982104793721e-05, | |
| "loss": 1.5393, | |
| "step": 4540 | |
| }, | |
| { | |
| "epoch": 0.455, | |
| "grad_norm": 0.41180986166000366, | |
| "learning_rate": 3.3153320628835953e-05, | |
| "loss": 1.5423, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.455, | |
| "eval_loss": 1.6033803224563599, | |
| "eval_runtime": 4.6294, | |
| "eval_samples_per_second": 432.018, | |
| "eval_steps_per_second": 27.001, | |
| "step": 4550 | |
| }, | |
| { | |
| "epoch": 0.456, | |
| "grad_norm": 0.3614816963672638, | |
| "learning_rate": 3.311694036919079e-05, | |
| "loss": 1.4995, | |
| "step": 4560 | |
| }, | |
| { | |
| "epoch": 0.457, | |
| "grad_norm": 0.416110098361969, | |
| "learning_rate": 3.308067961117001e-05, | |
| "loss": 1.4939, | |
| "step": 4570 | |
| }, | |
| { | |
| "epoch": 0.458, | |
| "grad_norm": 0.3330053389072418, | |
| "learning_rate": 3.3044537701972836e-05, | |
| "loss": 1.5118, | |
| "step": 4580 | |
| }, | |
| { | |
| "epoch": 0.459, | |
| "grad_norm": 0.35887932777404785, | |
| "learning_rate": 3.30085139937801e-05, | |
| "loss": 1.5084, | |
| "step": 4590 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "grad_norm": 0.33625465631484985, | |
| "learning_rate": 3.2972607843705445e-05, | |
| "loss": 1.5614, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.46, | |
| "eval_loss": 1.6055898666381836, | |
| "eval_runtime": 4.6102, | |
| "eval_samples_per_second": 433.823, | |
| "eval_steps_per_second": 27.114, | |
| "step": 4600 | |
| }, | |
| { | |
| "epoch": 0.461, | |
| "grad_norm": 0.4302857518196106, | |
| "learning_rate": 3.293681861374713e-05, | |
| "loss": 1.459, | |
| "step": 4610 | |
| }, | |
| { | |
| "epoch": 0.462, | |
| "grad_norm": 0.37634381651878357, | |
| "learning_rate": 3.290114567074045e-05, | |
| "loss": 1.513, | |
| "step": 4620 | |
| }, | |
| { | |
| "epoch": 0.463, | |
| "grad_norm": 0.3611469268798828, | |
| "learning_rate": 3.2865588386310646e-05, | |
| "loss": 1.5474, | |
| "step": 4630 | |
| }, | |
| { | |
| "epoch": 0.464, | |
| "grad_norm": 0.34918755292892456, | |
| "learning_rate": 3.2830146136826445e-05, | |
| "loss": 1.5675, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 0.465, | |
| "grad_norm": 0.3509509563446045, | |
| "learning_rate": 3.2794818303354126e-05, | |
| "loss": 1.5691, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.465, | |
| "eval_loss": 1.602272391319275, | |
| "eval_runtime": 4.6268, | |
| "eval_samples_per_second": 432.262, | |
| "eval_steps_per_second": 27.016, | |
| "step": 4650 | |
| }, | |
| { | |
| "epoch": 0.466, | |
| "grad_norm": 0.3535073399543762, | |
| "learning_rate": 3.2759604271612094e-05, | |
| "loss": 1.5149, | |
| "step": 4660 | |
| }, | |
| { | |
| "epoch": 0.467, | |
| "grad_norm": 0.36410200595855713, | |
| "learning_rate": 3.272450343192603e-05, | |
| "loss": 1.5849, | |
| "step": 4670 | |
| }, | |
| { | |
| "epoch": 0.468, | |
| "grad_norm": 0.34112101793289185, | |
| "learning_rate": 3.2689515179184576e-05, | |
| "loss": 1.5669, | |
| "step": 4680 | |
| }, | |
| { | |
| "epoch": 0.469, | |
| "grad_norm": 0.352558970451355, | |
| "learning_rate": 3.265463891279551e-05, | |
| "loss": 1.4972, | |
| "step": 4690 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "grad_norm": 0.4231419563293457, | |
| "learning_rate": 3.2619874036642426e-05, | |
| "loss": 1.5375, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.47, | |
| "eval_loss": 1.6014271974563599, | |
| "eval_runtime": 4.6266, | |
| "eval_samples_per_second": 432.286, | |
| "eval_steps_per_second": 27.018, | |
| "step": 4700 | |
| }, | |
| { | |
| "epoch": 0.471, | |
| "grad_norm": 0.4738163352012634, | |
| "learning_rate": 3.258521995904196e-05, | |
| "loss": 1.5724, | |
| "step": 4710 | |
| }, | |
| { | |
| "epoch": 0.472, | |
| "grad_norm": 0.489130437374115, | |
| "learning_rate": 3.2550676092701496e-05, | |
| "loss": 1.5234, | |
| "step": 4720 | |
| }, | |
| { | |
| "epoch": 0.473, | |
| "grad_norm": 0.40578359365463257, | |
| "learning_rate": 3.2516241854677356e-05, | |
| "loss": 1.4739, | |
| "step": 4730 | |
| }, | |
| { | |
| "epoch": 0.474, | |
| "grad_norm": 0.3754052221775055, | |
| "learning_rate": 3.248191666633348e-05, | |
| "loss": 1.5517, | |
| "step": 4740 | |
| }, | |
| { | |
| "epoch": 0.475, | |
| "grad_norm": 0.4701234996318817, | |
| "learning_rate": 3.244769995330059e-05, | |
| "loss": 1.557, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.475, | |
| "eval_loss": 1.6014772653579712, | |
| "eval_runtime": 4.6244, | |
| "eval_samples_per_second": 432.488, | |
| "eval_steps_per_second": 27.03, | |
| "step": 4750 | |
| }, | |
| { | |
| "epoch": 0.476, | |
| "grad_norm": 0.37672650814056396, | |
| "learning_rate": 3.241359114543582e-05, | |
| "loss": 1.5531, | |
| "step": 4760 | |
| }, | |
| { | |
| "epoch": 0.477, | |
| "grad_norm": 0.40195539593696594, | |
| "learning_rate": 3.237958967678283e-05, | |
| "loss": 1.5759, | |
| "step": 4770 | |
| }, | |
| { | |
| "epoch": 0.478, | |
| "grad_norm": 0.3967335522174835, | |
| "learning_rate": 3.234569498553233e-05, | |
| "loss": 1.5871, | |
| "step": 4780 | |
| }, | |
| { | |
| "epoch": 0.479, | |
| "grad_norm": 0.3859959840774536, | |
| "learning_rate": 3.231190651398314e-05, | |
| "loss": 1.5716, | |
| "step": 4790 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "grad_norm": 0.42818543314933777, | |
| "learning_rate": 3.227822370850359e-05, | |
| "loss": 1.5402, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.48, | |
| "eval_loss": 1.5974245071411133, | |
| "eval_runtime": 4.6278, | |
| "eval_samples_per_second": 432.172, | |
| "eval_steps_per_second": 27.011, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 0.481, | |
| "grad_norm": 0.4882192015647888, | |
| "learning_rate": 3.224464601949349e-05, | |
| "loss": 1.599, | |
| "step": 4810 | |
| }, | |
| { | |
| "epoch": 0.482, | |
| "grad_norm": 0.4261423647403717, | |
| "learning_rate": 3.2211172901346385e-05, | |
| "loss": 1.5992, | |
| "step": 4820 | |
| }, | |
| { | |
| "epoch": 0.483, | |
| "grad_norm": 0.3666815459728241, | |
| "learning_rate": 3.21778038124124e-05, | |
| "loss": 1.595, | |
| "step": 4830 | |
| }, | |
| { | |
| "epoch": 0.484, | |
| "grad_norm": 0.35577964782714844, | |
| "learning_rate": 3.2144538214961344e-05, | |
| "loss": 1.5837, | |
| "step": 4840 | |
| }, | |
| { | |
| "epoch": 0.485, | |
| "grad_norm": 0.3705880045890808, | |
| "learning_rate": 3.21113755751464e-05, | |
| "loss": 1.5842, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.485, | |
| "eval_loss": 1.6001406908035278, | |
| "eval_runtime": 4.6403, | |
| "eval_samples_per_second": 431.008, | |
| "eval_steps_per_second": 26.938, | |
| "step": 4850 | |
| }, | |
| { | |
| "epoch": 0.486, | |
| "grad_norm": 0.4082805812358856, | |
| "learning_rate": 3.207831536296808e-05, | |
| "loss": 1.5471, | |
| "step": 4860 | |
| }, | |
| { | |
| "epoch": 0.487, | |
| "grad_norm": 0.41893184185028076, | |
| "learning_rate": 3.2045357052238676e-05, | |
| "loss": 1.5404, | |
| "step": 4870 | |
| }, | |
| { | |
| "epoch": 0.488, | |
| "grad_norm": 0.4093189537525177, | |
| "learning_rate": 3.201250012054707e-05, | |
| "loss": 1.5889, | |
| "step": 4880 | |
| }, | |
| { | |
| "epoch": 0.489, | |
| "grad_norm": 0.4183659553527832, | |
| "learning_rate": 3.197974404922397e-05, | |
| "loss": 1.5439, | |
| "step": 4890 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "grad_norm": 0.4108491837978363, | |
| "learning_rate": 3.194708832330752e-05, | |
| "loss": 1.5454, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.49, | |
| "eval_loss": 1.596949577331543, | |
| "eval_runtime": 4.612, | |
| "eval_samples_per_second": 433.65, | |
| "eval_steps_per_second": 27.103, | |
| "step": 4900 | |
| }, | |
| { | |
| "epoch": 0.491, | |
| "grad_norm": 0.44454118609428406, | |
| "learning_rate": 3.191453243150929e-05, | |
| "loss": 1.5632, | |
| "step": 4910 | |
| }, | |
| { | |
| "epoch": 0.492, | |
| "grad_norm": 0.3570930063724518, | |
| "learning_rate": 3.188207586618064e-05, | |
| "loss": 1.4952, | |
| "step": 4920 | |
| }, | |
| { | |
| "epoch": 0.493, | |
| "grad_norm": 0.377950519323349, | |
| "learning_rate": 3.1849718123279517e-05, | |
| "loss": 1.5151, | |
| "step": 4930 | |
| }, | |
| { | |
| "epoch": 0.494, | |
| "grad_norm": 0.38532108068466187, | |
| "learning_rate": 3.181745870233753e-05, | |
| "loss": 1.5811, | |
| "step": 4940 | |
| }, | |
| { | |
| "epoch": 0.495, | |
| "grad_norm": 0.3370131850242615, | |
| "learning_rate": 3.178529710642749e-05, | |
| "loss": 1.5415, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.495, | |
| "eval_loss": 1.5937820672988892, | |
| "eval_runtime": 4.6236, | |
| "eval_samples_per_second": 432.56, | |
| "eval_steps_per_second": 27.035, | |
| "step": 4950 | |
| }, | |
| { | |
| "epoch": 0.496, | |
| "grad_norm": 0.43014732003211975, | |
| "learning_rate": 3.175323284213118e-05, | |
| "loss": 1.5406, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 0.497, | |
| "grad_norm": 0.4400146007537842, | |
| "learning_rate": 3.172126541950766e-05, | |
| "loss": 1.5578, | |
| "step": 4970 | |
| }, | |
| { | |
| "epoch": 0.498, | |
| "grad_norm": 0.41548582911491394, | |
| "learning_rate": 3.1689394352061735e-05, | |
| "loss": 1.5549, | |
| "step": 4980 | |
| }, | |
| { | |
| "epoch": 0.499, | |
| "grad_norm": 0.37343522906303406, | |
| "learning_rate": 3.165761915671293e-05, | |
| "loss": 1.5127, | |
| "step": 4990 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "grad_norm": 0.3881683051586151, | |
| "learning_rate": 3.162593935376469e-05, | |
| "loss": 1.5553, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.5, | |
| "eval_loss": 1.6019805669784546, | |
| "eval_runtime": 4.6105, | |
| "eval_samples_per_second": 433.793, | |
| "eval_steps_per_second": 27.112, | |
| "step": 5000 | |
| }, | |
| { | |
| "epoch": 0.501, | |
| "grad_norm": 0.35221096873283386, | |
| "learning_rate": 3.159435446687396e-05, | |
| "loss": 1.5544, | |
| "step": 5010 | |
| }, | |
| { | |
| "epoch": 0.502, | |
| "grad_norm": 0.33756914734840393, | |
| "learning_rate": 3.1562864023021174e-05, | |
| "loss": 1.5108, | |
| "step": 5020 | |
| }, | |
| { | |
| "epoch": 0.503, | |
| "grad_norm": 0.340436726808548, | |
| "learning_rate": 3.1531467552480395e-05, | |
| "loss": 1.4998, | |
| "step": 5030 | |
| }, | |
| { | |
| "epoch": 0.504, | |
| "grad_norm": 0.3652532696723938, | |
| "learning_rate": 3.1500164588789964e-05, | |
| "loss": 1.5853, | |
| "step": 5040 | |
| }, | |
| { | |
| "epoch": 0.505, | |
| "grad_norm": 0.37459462881088257, | |
| "learning_rate": 3.146895466872338e-05, | |
| "loss": 1.5595, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.505, | |
| "eval_loss": 1.5959675312042236, | |
| "eval_runtime": 4.6328, | |
| "eval_samples_per_second": 431.705, | |
| "eval_steps_per_second": 26.982, | |
| "step": 5050 | |
| }, | |
| { | |
| "epoch": 0.506, | |
| "grad_norm": 0.4339495599269867, | |
| "learning_rate": 3.1437837332260465e-05, | |
| "loss": 1.5392, | |
| "step": 5060 | |
| }, | |
| { | |
| "epoch": 0.507, | |
| "grad_norm": 0.4025578796863556, | |
| "learning_rate": 3.140681212255896e-05, | |
| "loss": 1.499, | |
| "step": 5070 | |
| }, | |
| { | |
| "epoch": 0.508, | |
| "grad_norm": 0.38840150833129883, | |
| "learning_rate": 3.137587858592628e-05, | |
| "loss": 1.5285, | |
| "step": 5080 | |
| }, | |
| { | |
| "epoch": 0.509, | |
| "grad_norm": 0.34794872999191284, | |
| "learning_rate": 3.13450362717917e-05, | |
| "loss": 1.5384, | |
| "step": 5090 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "grad_norm": 0.3731895089149475, | |
| "learning_rate": 3.131428473267876e-05, | |
| "loss": 1.5532, | |
| "step": 5100 | |
| }, | |
| { | |
| "epoch": 0.51, | |
| "eval_loss": 1.591291904449463, | |
| "eval_runtime": 4.625, | |
| "eval_samples_per_second": 432.432, | |
| "eval_steps_per_second": 27.027, | |
| "step": 5100 | |
| } | |
| ], | |
| "logging_steps": 10, | |
| "max_steps": 10000, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 9223372036854775807, | |
| "save_steps": 50, | |
| "stateful_callbacks": { | |
| "EarlyStoppingCallback": { | |
| "args": { | |
| "early_stopping_patience": 5, | |
| "early_stopping_threshold": 0.0 | |
| }, | |
| "attributes": { | |
| "early_stopping_patience_counter": 0 | |
| } | |
| }, | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.3645715079168e+18, | |
| "train_batch_size": 32, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |