| { |
| "best_global_step": null, |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 1.0, |
| "eval_steps": 500, |
| "global_step": 363, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.002758620689655172, |
| "grad_norm": 2.2012167823201163, |
| "learning_rate": 0.0, |
| "loss": 0.0852, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.005517241379310344, |
| "grad_norm": 2.684327612345962, |
| "learning_rate": 5.405405405405406e-07, |
| "loss": 0.1171, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.008275862068965517, |
| "grad_norm": 2.922848494791373, |
| "learning_rate": 1.0810810810810812e-06, |
| "loss": 0.1201, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.011034482758620689, |
| "grad_norm": 2.923985908786216, |
| "learning_rate": 1.6216216216216219e-06, |
| "loss": 0.0976, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.013793103448275862, |
| "grad_norm": 1.2724240051553533, |
| "learning_rate": 2.1621621621621623e-06, |
| "loss": 0.1046, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.016551724137931035, |
| "grad_norm": 1.673019205535178, |
| "learning_rate": 2.702702702702703e-06, |
| "loss": 0.0609, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.019310344827586208, |
| "grad_norm": 2.504855019369174, |
| "learning_rate": 3.2432432432432437e-06, |
| "loss": 0.081, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.022068965517241378, |
| "grad_norm": 1.878885411504625, |
| "learning_rate": 3.7837837837837844e-06, |
| "loss": 0.1121, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.02482758620689655, |
| "grad_norm": 2.653287813889321, |
| "learning_rate": 4.324324324324325e-06, |
| "loss": 0.1125, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.027586206896551724, |
| "grad_norm": 1.7453919996030804, |
| "learning_rate": 4.864864864864866e-06, |
| "loss": 0.0724, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.030344827586206897, |
| "grad_norm": 3.333638766021421, |
| "learning_rate": 5.405405405405406e-06, |
| "loss": 0.1475, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.03310344827586207, |
| "grad_norm": 2.292734722563975, |
| "learning_rate": 5.945945945945947e-06, |
| "loss": 0.0974, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.03586206896551724, |
| "grad_norm": 1.099049620799708, |
| "learning_rate": 6.486486486486487e-06, |
| "loss": 0.0597, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.038620689655172416, |
| "grad_norm": 1.727408246250002, |
| "learning_rate": 7.027027027027028e-06, |
| "loss": 0.0778, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.041379310344827586, |
| "grad_norm": 1.761342077131181, |
| "learning_rate": 7.567567567567569e-06, |
| "loss": 0.0955, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.044137931034482755, |
| "grad_norm": 1.9277700023635902, |
| "learning_rate": 8.108108108108109e-06, |
| "loss": 0.0869, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.04689655172413793, |
| "grad_norm": 1.568808525211501, |
| "learning_rate": 8.64864864864865e-06, |
| "loss": 0.0807, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.0496551724137931, |
| "grad_norm": 1.3523306531543753, |
| "learning_rate": 9.189189189189191e-06, |
| "loss": 0.0665, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.05241379310344828, |
| "grad_norm": 1.7180837689557336, |
| "learning_rate": 9.729729729729732e-06, |
| "loss": 0.0868, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.05517241379310345, |
| "grad_norm": 1.5131863973051378, |
| "learning_rate": 1.027027027027027e-05, |
| "loss": 0.0736, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.057931034482758624, |
| "grad_norm": 1.6815268723910783, |
| "learning_rate": 1.0810810810810812e-05, |
| "loss": 0.079, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.060689655172413794, |
| "grad_norm": 1.4834028913829107, |
| "learning_rate": 1.1351351351351352e-05, |
| "loss": 0.0437, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.06344827586206897, |
| "grad_norm": 2.1417510445007792, |
| "learning_rate": 1.1891891891891894e-05, |
| "loss": 0.0905, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.06620689655172414, |
| "grad_norm": 1.7871991104284883, |
| "learning_rate": 1.2432432432432433e-05, |
| "loss": 0.1024, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.06896551724137931, |
| "grad_norm": 2.0156063129423485, |
| "learning_rate": 1.2972972972972975e-05, |
| "loss": 0.0701, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.07172413793103448, |
| "grad_norm": 1.9160557707810737, |
| "learning_rate": 1.3513513513513515e-05, |
| "loss": 0.0748, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.07448275862068965, |
| "grad_norm": 2.000911025201981, |
| "learning_rate": 1.4054054054054055e-05, |
| "loss": 0.101, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.07724137931034483, |
| "grad_norm": 2.115047269982351, |
| "learning_rate": 1.4594594594594596e-05, |
| "loss": 0.0918, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.08, |
| "grad_norm": 1.8098326272573262, |
| "learning_rate": 1.5135135135135138e-05, |
| "loss": 0.0915, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.08275862068965517, |
| "grad_norm": 2.1207994498886764, |
| "learning_rate": 1.5675675675675676e-05, |
| "loss": 0.1025, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.08551724137931034, |
| "grad_norm": 2.171868532095742, |
| "learning_rate": 1.6216216216216218e-05, |
| "loss": 0.1059, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.08827586206896551, |
| "grad_norm": 1.6943229112776936, |
| "learning_rate": 1.6756756756756757e-05, |
| "loss": 0.0677, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.0910344827586207, |
| "grad_norm": 1.994899794317826, |
| "learning_rate": 1.72972972972973e-05, |
| "loss": 0.0894, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.09379310344827586, |
| "grad_norm": 2.3711691862715907, |
| "learning_rate": 1.783783783783784e-05, |
| "loss": 0.1158, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.09655172413793103, |
| "grad_norm": 1.9877327316155244, |
| "learning_rate": 1.8378378378378383e-05, |
| "loss": 0.0924, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.0993103448275862, |
| "grad_norm": 2.0641113315593898, |
| "learning_rate": 1.891891891891892e-05, |
| "loss": 0.0978, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.10206896551724139, |
| "grad_norm": 2.5636633454788593, |
| "learning_rate": 1.9459459459459463e-05, |
| "loss": 0.1327, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.10482758620689656, |
| "grad_norm": 2.3262721461116396, |
| "learning_rate": 2e-05, |
| "loss": 0.095, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.10758620689655173, |
| "grad_norm": 2.0685116596847832, |
| "learning_rate": 1.9999535665248e-05, |
| "loss": 0.0706, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.1103448275862069, |
| "grad_norm": 2.5954824427417007, |
| "learning_rate": 1.999814270411335e-05, |
| "loss": 0.1205, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.11310344827586206, |
| "grad_norm": 2.144122161715886, |
| "learning_rate": 1.99958212459561e-05, |
| "loss": 0.0896, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.11586206896551725, |
| "grad_norm": 1.9322201299634434, |
| "learning_rate": 1.9992571506362997e-05, |
| "loss": 0.1012, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.11862068965517242, |
| "grad_norm": 2.075048476162322, |
| "learning_rate": 1.9988393787127444e-05, |
| "loss": 0.1087, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.12137931034482759, |
| "grad_norm": 2.259726318820954, |
| "learning_rate": 1.9983288476221482e-05, |
| "loss": 0.1382, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.12413793103448276, |
| "grad_norm": 1.360664005910794, |
| "learning_rate": 1.9977256047759765e-05, |
| "loss": 0.0538, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.12689655172413794, |
| "grad_norm": 1.6435792074598727, |
| "learning_rate": 1.9970297061955533e-05, |
| "loss": 0.0855, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.1296551724137931, |
| "grad_norm": 2.3601345815092483, |
| "learning_rate": 1.9962412165068575e-05, |
| "loss": 0.1386, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.13241379310344828, |
| "grad_norm": 1.7105484837560168, |
| "learning_rate": 1.9953602089345215e-05, |
| "loss": 0.0961, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.13517241379310344, |
| "grad_norm": 1.6496292161975452, |
| "learning_rate": 1.9943867652950323e-05, |
| "loss": 0.0857, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.13793103448275862, |
| "grad_norm": 1.2468755808496865, |
| "learning_rate": 1.9933209759891318e-05, |
| "loss": 0.0563, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.1406896551724138, |
| "grad_norm": 2.0793413384987183, |
| "learning_rate": 1.9921629399934224e-05, |
| "loss": 0.108, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.14344827586206896, |
| "grad_norm": 2.0093677538398524, |
| "learning_rate": 1.9909127648511758e-05, |
| "loss": 0.1289, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.14620689655172414, |
| "grad_norm": 1.2736371995174656, |
| "learning_rate": 1.989570566662345e-05, |
| "loss": 0.0615, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.1489655172413793, |
| "grad_norm": 1.5368399292599437, |
| "learning_rate": 1.9881364700727827e-05, |
| "loss": 0.0559, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.15172413793103448, |
| "grad_norm": 1.6143124587473046, |
| "learning_rate": 1.986610608262665e-05, |
| "loss": 0.0512, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.15448275862068966, |
| "grad_norm": 2.0640396297545323, |
| "learning_rate": 1.9849931229341258e-05, |
| "loss": 0.1492, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.15724137931034482, |
| "grad_norm": 1.9321341515451917, |
| "learning_rate": 1.9832841642980948e-05, |
| "loss": 0.1277, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.16, |
| "grad_norm": 3.5565020190489753, |
| "learning_rate": 1.981483891060348e-05, |
| "loss": 0.1266, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.16275862068965516, |
| "grad_norm": 1.6731150384360727, |
| "learning_rate": 1.979592470406772e-05, |
| "loss": 0.095, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.16551724137931034, |
| "grad_norm": 1.3348447458614452, |
| "learning_rate": 1.9776100779878344e-05, |
| "loss": 0.0602, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.16827586206896553, |
| "grad_norm": 1.211440894700164, |
| "learning_rate": 1.9755368979022734e-05, |
| "loss": 0.0597, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.17103448275862068, |
| "grad_norm": 2.045395762464389, |
| "learning_rate": 1.9733731226800016e-05, |
| "loss": 0.0902, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.17379310344827587, |
| "grad_norm": 1.681889406229311, |
| "learning_rate": 1.9711189532642244e-05, |
| "loss": 0.0851, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.17655172413793102, |
| "grad_norm": 2.156485405646681, |
| "learning_rate": 1.9687745989927823e-05, |
| "loss": 0.1022, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.1793103448275862, |
| "grad_norm": 2.2694346385487765, |
| "learning_rate": 1.9663402775787066e-05, |
| "loss": 0.1402, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.1820689655172414, |
| "grad_norm": 2.2102729735649502, |
| "learning_rate": 1.9638162150900028e-05, |
| "loss": 0.0595, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.18482758620689654, |
| "grad_norm": 2.4808367225986023, |
| "learning_rate": 1.961202645928658e-05, |
| "loss": 0.1526, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.18758620689655173, |
| "grad_norm": 2.125991623159161, |
| "learning_rate": 1.9584998128088686e-05, |
| "loss": 0.1149, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.19034482758620688, |
| "grad_norm": 3.5959738930248553, |
| "learning_rate": 1.955707966734505e-05, |
| "loss": 0.1724, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.19310344827586207, |
| "grad_norm": 2.217349082976429, |
| "learning_rate": 1.9528273669757974e-05, |
| "loss": 0.1357, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.19586206896551725, |
| "grad_norm": 2.9019793725624288, |
| "learning_rate": 1.9498582810452607e-05, |
| "loss": 0.1289, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.1986206896551724, |
| "grad_norm": 2.677822777383495, |
| "learning_rate": 1.9468009846728515e-05, |
| "loss": 0.1201, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.2013793103448276, |
| "grad_norm": 2.1904068713477693, |
| "learning_rate": 1.9436557617803594e-05, |
| "loss": 0.1343, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.20413793103448277, |
| "grad_norm": 1.9333752870288448, |
| "learning_rate": 1.9404229044550432e-05, |
| "loss": 0.1014, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.20689655172413793, |
| "grad_norm": 1.2594973301567354, |
| "learning_rate": 1.9371027129225042e-05, |
| "loss": 0.08, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.2096551724137931, |
| "grad_norm": 1.1588871349219974, |
| "learning_rate": 1.9336954955188042e-05, |
| "loss": 0.0664, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.21241379310344827, |
| "grad_norm": 2.424390700582275, |
| "learning_rate": 1.9302015686618328e-05, |
| "loss": 0.1228, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.21517241379310345, |
| "grad_norm": 1.839851381331024, |
| "learning_rate": 1.9266212568219223e-05, |
| "loss": 0.1181, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.21793103448275863, |
| "grad_norm": 2.9112052255264627, |
| "learning_rate": 1.9229548924917146e-05, |
| "loss": 0.1593, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.2206896551724138, |
| "grad_norm": 2.6573066473445124, |
| "learning_rate": 1.9192028161552848e-05, |
| "loss": 0.1068, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.22344827586206897, |
| "grad_norm": 2.228711734634549, |
| "learning_rate": 1.915365376256519e-05, |
| "loss": 0.1395, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.22620689655172413, |
| "grad_norm": 2.1487793287842463, |
| "learning_rate": 1.9114429291667583e-05, |
| "loss": 0.1322, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.2289655172413793, |
| "grad_norm": 1.418589529134045, |
| "learning_rate": 1.9074358391517026e-05, |
| "loss": 0.1298, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.2317241379310345, |
| "grad_norm": 1.6184969490675134, |
| "learning_rate": 1.9033444783375806e-05, |
| "loss": 0.1375, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.23448275862068965, |
| "grad_norm": 2.636414425888468, |
| "learning_rate": 1.8991692266765947e-05, |
| "loss": 0.1454, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.23724137931034484, |
| "grad_norm": 1.5999297559125214, |
| "learning_rate": 1.8949104719116334e-05, |
| "loss": 0.1004, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.24, |
| "grad_norm": 1.7047937938863038, |
| "learning_rate": 1.8905686095402648e-05, |
| "loss": 0.0729, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.24275862068965517, |
| "grad_norm": 2.213860086801788, |
| "learning_rate": 1.886144042778006e-05, |
| "loss": 0.1219, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.24551724137931036, |
| "grad_norm": 1.9830213856101813, |
| "learning_rate": 1.881637182520879e-05, |
| "loss": 0.1277, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.2482758620689655, |
| "grad_norm": 2.1526995896994543, |
| "learning_rate": 1.8770484473072518e-05, |
| "loss": 0.133, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.25103448275862067, |
| "grad_norm": 5.615536299423343, |
| "learning_rate": 1.87237826327897e-05, |
| "loss": 0.1271, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.2537931034482759, |
| "grad_norm": 2.9892624763616213, |
| "learning_rate": 1.8676270641417824e-05, |
| "loss": 0.1337, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.25655172413793104, |
| "grad_norm": 1.7670570813048203, |
| "learning_rate": 1.8627952911250632e-05, |
| "loss": 0.1188, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.2593103448275862, |
| "grad_norm": 1.8832817735619782, |
| "learning_rate": 1.857883392940837e-05, |
| "loss": 0.09, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.2620689655172414, |
| "grad_norm": 2.2995757188920964, |
| "learning_rate": 1.85289182574211e-05, |
| "loss": 0.1084, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.26482758620689656, |
| "grad_norm": 2.268436925129853, |
| "learning_rate": 1.847821053080505e-05, |
| "loss": 0.1003, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.2675862068965517, |
| "grad_norm": 1.8417005302085403, |
| "learning_rate": 1.8426715458632154e-05, |
| "loss": 0.0747, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.27034482758620687, |
| "grad_norm": 2.497572736496612, |
| "learning_rate": 1.8374437823092726e-05, |
| "loss": 0.1513, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.2731034482758621, |
| "grad_norm": 2.0212629509138766, |
| "learning_rate": 1.832138247905135e-05, |
| "loss": 0.088, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.27586206896551724, |
| "grad_norm": 1.806243660764203, |
| "learning_rate": 1.8267554353596027e-05, |
| "loss": 0.0617, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.2786206896551724, |
| "grad_norm": 1.43840041921586, |
| "learning_rate": 1.8212958445580623e-05, |
| "loss": 0.0809, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.2813793103448276, |
| "grad_norm": 1.5624416473825227, |
| "learning_rate": 1.815759982516061e-05, |
| "loss": 0.0987, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.28413793103448276, |
| "grad_norm": 1.3722413173440258, |
| "learning_rate": 1.8101483633322255e-05, |
| "loss": 0.1055, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.2868965517241379, |
| "grad_norm": 1.5055322976041905, |
| "learning_rate": 1.8044615081405153e-05, |
| "loss": 0.12, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.2896551724137931, |
| "grad_norm": 1.9238100509571232, |
| "learning_rate": 1.7986999450618295e-05, |
| "loss": 0.1293, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.2924137931034483, |
| "grad_norm": 1.4169309751550185, |
| "learning_rate": 1.7928642091549616e-05, |
| "loss": 0.0945, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.29517241379310344, |
| "grad_norm": 1.7127369268478376, |
| "learning_rate": 1.7869548423669075e-05, |
| "loss": 0.1124, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.2979310344827586, |
| "grad_norm": 1.859903898855105, |
| "learning_rate": 1.7809723934825405e-05, |
| "loss": 0.1182, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.3006896551724138, |
| "grad_norm": 1.4775939025780847, |
| "learning_rate": 1.7749174180736443e-05, |
| "loss": 0.0857, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.30344827586206896, |
| "grad_norm": 1.4354799536836982, |
| "learning_rate": 1.768790478447319e-05, |
| "loss": 0.0781, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.3062068965517241, |
| "grad_norm": 1.0273549786544405, |
| "learning_rate": 1.762592143593764e-05, |
| "loss": 0.0713, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.30896551724137933, |
| "grad_norm": 1.9640486386494376, |
| "learning_rate": 1.756322989133434e-05, |
| "loss": 0.1301, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.3117241379310345, |
| "grad_norm": 1.7089277562540073, |
| "learning_rate": 1.749983597263586e-05, |
| "loss": 0.0681, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.31448275862068964, |
| "grad_norm": 1.9429869321530793, |
| "learning_rate": 1.7435745567042096e-05, |
| "loss": 0.1331, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.31724137931034485, |
| "grad_norm": 1.26536650090277, |
| "learning_rate": 1.737096462643357e-05, |
| "loss": 0.0604, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.32, |
| "grad_norm": 2.087216494160791, |
| "learning_rate": 1.730549916681868e-05, |
| "loss": 0.1494, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.32275862068965516, |
| "grad_norm": 1.9975034441736954, |
| "learning_rate": 1.723935526777502e-05, |
| "loss": 0.1092, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.3255172413793103, |
| "grad_norm": 2.038235893575754, |
| "learning_rate": 1.717253907188477e-05, |
| "loss": 0.1082, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.32827586206896553, |
| "grad_norm": 3.3184113391653396, |
| "learning_rate": 1.7105056784164295e-05, |
| "loss": 0.1703, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.3310344827586207, |
| "grad_norm": 2.0270058507038966, |
| "learning_rate": 1.7036914671487854e-05, |
| "loss": 0.1303, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.33379310344827584, |
| "grad_norm": 1.8154593381383104, |
| "learning_rate": 1.6968119062005644e-05, |
| "loss": 0.1077, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.33655172413793105, |
| "grad_norm": 1.2350092975243878, |
| "learning_rate": 1.689867634455612e-05, |
| "loss": 0.0924, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.3393103448275862, |
| "grad_norm": 1.792900391384705, |
| "learning_rate": 1.682859296807268e-05, |
| "loss": 0.1221, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.34206896551724136, |
| "grad_norm": 1.9404357508554042, |
| "learning_rate": 1.675787544098477e-05, |
| "loss": 0.1106, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.3448275862068966, |
| "grad_norm": 1.7814508590384097, |
| "learning_rate": 1.6686530330613472e-05, |
| "loss": 0.1122, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.34758620689655173, |
| "grad_norm": 1.53815702992395, |
| "learning_rate": 1.661456426256161e-05, |
| "loss": 0.087, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.3503448275862069, |
| "grad_norm": 1.9077349161199095, |
| "learning_rate": 1.6541983920098462e-05, |
| "loss": 0.087, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.35310344827586204, |
| "grad_norm": 2.2996250559408704, |
| "learning_rate": 1.6468796043539082e-05, |
| "loss": 0.1254, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.35586206896551725, |
| "grad_norm": 1.2031295382564096, |
| "learning_rate": 1.639500742961838e-05, |
| "loss": 0.0498, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.3586206896551724, |
| "grad_norm": 1.3436281870681093, |
| "learning_rate": 1.6320624930859905e-05, |
| "loss": 0.077, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.36137931034482756, |
| "grad_norm": 0.6314939772663396, |
| "learning_rate": 1.6245655454939474e-05, |
| "loss": 0.0352, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.3641379310344828, |
| "grad_norm": 2.420910473317855, |
| "learning_rate": 1.6170105964043698e-05, |
| "loss": 0.1228, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.36689655172413793, |
| "grad_norm": 2.80050435155021, |
| "learning_rate": 1.6093983474223392e-05, |
| "loss": 0.1799, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.3696551724137931, |
| "grad_norm": 1.9733179937634817, |
| "learning_rate": 1.6017295054742045e-05, |
| "loss": 0.1088, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.3724137931034483, |
| "grad_norm": 0.8866446082715613, |
| "learning_rate": 1.5940047827419305e-05, |
| "loss": 0.0722, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.37517241379310345, |
| "grad_norm": 1.4747083716633578, |
| "learning_rate": 1.5862248965969604e-05, |
| "loss": 0.0902, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.3779310344827586, |
| "grad_norm": 1.203872871772734, |
| "learning_rate": 1.5783905695335947e-05, |
| "loss": 0.0763, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.38068965517241377, |
| "grad_norm": 1.295759494638433, |
| "learning_rate": 1.570502529101896e-05, |
| "loss": 0.0582, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.383448275862069, |
| "grad_norm": 2.082719193654199, |
| "learning_rate": 1.5625615078401244e-05, |
| "loss": 0.1131, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.38620689655172413, |
| "grad_norm": 2.4303248947365046, |
| "learning_rate": 1.5545682432067068e-05, |
| "loss": 0.1319, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.3889655172413793, |
| "grad_norm": 2.308746027421277, |
| "learning_rate": 1.5465234775117538e-05, |
| "loss": 0.1232, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.3917241379310345, |
| "grad_norm": 2.9364629760972907, |
| "learning_rate": 1.5384279578481223e-05, |
| "loss": 0.1537, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.39448275862068966, |
| "grad_norm": 1.8350345374035169, |
| "learning_rate": 1.5302824360220352e-05, |
| "loss": 0.1161, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.3972413793103448, |
| "grad_norm": 1.8562459888809821, |
| "learning_rate": 1.522087668483264e-05, |
| "loss": 0.1054, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.4, |
| "grad_norm": 2.2050577332052317, |
| "learning_rate": 1.5138444162548791e-05, |
| "loss": 0.072, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.4027586206896552, |
| "grad_norm": 2.576862971551694, |
| "learning_rate": 1.5055534448625766e-05, |
| "loss": 0.1956, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.40551724137931033, |
| "grad_norm": 2.526412596716041, |
| "learning_rate": 1.4972155242635853e-05, |
| "loss": 0.1163, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.40827586206896554, |
| "grad_norm": 2.4122326865926325, |
| "learning_rate": 1.488831428775164e-05, |
| "loss": 0.1588, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.4110344827586207, |
| "grad_norm": 1.5900334279122388, |
| "learning_rate": 1.4804019370026927e-05, |
| "loss": 0.0874, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.41379310344827586, |
| "grad_norm": 2.2583849148863284, |
| "learning_rate": 1.4719278317673655e-05, |
| "loss": 0.1225, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.416551724137931, |
| "grad_norm": 1.5043786808378643, |
| "learning_rate": 1.4634099000334932e-05, |
| "loss": 0.0628, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.4193103448275862, |
| "grad_norm": 2.6867593671378707, |
| "learning_rate": 1.4548489328354197e-05, |
| "loss": 0.1119, |
| "step": 152 |
| }, |
| { |
| "epoch": 0.4220689655172414, |
| "grad_norm": 1.892183158003583, |
| "learning_rate": 1.4462457252040606e-05, |
| "loss": 0.1371, |
| "step": 153 |
| }, |
| { |
| "epoch": 0.42482758620689653, |
| "grad_norm": 0.7376897959603039, |
| "learning_rate": 1.437601076093073e-05, |
| "loss": 0.0715, |
| "step": 154 |
| }, |
| { |
| "epoch": 0.42758620689655175, |
| "grad_norm": 1.2313754414547655, |
| "learning_rate": 1.4289157883046567e-05, |
| "loss": 0.0671, |
| "step": 155 |
| }, |
| { |
| "epoch": 0.4303448275862069, |
| "grad_norm": 2.443959323743539, |
| "learning_rate": 1.420190668415002e-05, |
| "loss": 0.0937, |
| "step": 156 |
| }, |
| { |
| "epoch": 0.43310344827586206, |
| "grad_norm": 1.8279711072151712, |
| "learning_rate": 1.4114265266993847e-05, |
| "loss": 0.1114, |
| "step": 157 |
| }, |
| { |
| "epoch": 0.43586206896551727, |
| "grad_norm": 2.722814716999782, |
| "learning_rate": 1.4026241770569198e-05, |
| "loss": 0.1388, |
| "step": 158 |
| }, |
| { |
| "epoch": 0.4386206896551724, |
| "grad_norm": 2.522323402694141, |
| "learning_rate": 1.3937844369349736e-05, |
| "loss": 0.0862, |
| "step": 159 |
| }, |
| { |
| "epoch": 0.4413793103448276, |
| "grad_norm": 1.0652613518383625, |
| "learning_rate": 1.3849081272532545e-05, |
| "loss": 0.0752, |
| "step": 160 |
| }, |
| { |
| "epoch": 0.44413793103448274, |
| "grad_norm": 2.6542750874407335, |
| "learning_rate": 1.375996072327573e-05, |
| "loss": 0.1456, |
| "step": 161 |
| }, |
| { |
| "epoch": 0.44689655172413795, |
| "grad_norm": 1.8544625800539327, |
| "learning_rate": 1.3670490997932922e-05, |
| "loss": 0.109, |
| "step": 162 |
| }, |
| { |
| "epoch": 0.4496551724137931, |
| "grad_norm": 1.9735546120483556, |
| "learning_rate": 1.3580680405284666e-05, |
| "loss": 0.1144, |
| "step": 163 |
| }, |
| { |
| "epoch": 0.45241379310344826, |
| "grad_norm": 1.2177687301571833, |
| "learning_rate": 1.3490537285766809e-05, |
| "loss": 0.0838, |
| "step": 164 |
| }, |
| { |
| "epoch": 0.45517241379310347, |
| "grad_norm": 2.866049013585789, |
| "learning_rate": 1.3400070010695966e-05, |
| "loss": 0.1309, |
| "step": 165 |
| }, |
| { |
| "epoch": 0.4579310344827586, |
| "grad_norm": 1.7714947880406966, |
| "learning_rate": 1.3309286981492084e-05, |
| "loss": 0.1336, |
| "step": 166 |
| }, |
| { |
| "epoch": 0.4606896551724138, |
| "grad_norm": 2.3528937198669966, |
| "learning_rate": 1.3218196628898232e-05, |
| "loss": 0.0878, |
| "step": 167 |
| }, |
| { |
| "epoch": 0.463448275862069, |
| "grad_norm": 1.32763797639713, |
| "learning_rate": 1.3126807412197666e-05, |
| "loss": 0.0866, |
| "step": 168 |
| }, |
| { |
| "epoch": 0.46620689655172415, |
| "grad_norm": 1.0489060017302254, |
| "learning_rate": 1.3035127818428239e-05, |
| "loss": 0.0863, |
| "step": 169 |
| }, |
| { |
| "epoch": 0.4689655172413793, |
| "grad_norm": 1.4769228384530726, |
| "learning_rate": 1.2943166361594242e-05, |
| "loss": 0.1393, |
| "step": 170 |
| }, |
| { |
| "epoch": 0.47172413793103446, |
| "grad_norm": 1.9367121483246261, |
| "learning_rate": 1.2850931581875723e-05, |
| "loss": 0.1762, |
| "step": 171 |
| }, |
| { |
| "epoch": 0.47448275862068967, |
| "grad_norm": 1.5440366202640854, |
| "learning_rate": 1.275843204483539e-05, |
| "loss": 0.0599, |
| "step": 172 |
| }, |
| { |
| "epoch": 0.4772413793103448, |
| "grad_norm": 1.691410284752133, |
| "learning_rate": 1.2665676340623172e-05, |
| "loss": 0.0938, |
| "step": 173 |
| }, |
| { |
| "epoch": 0.48, |
| "grad_norm": 1.2010801724859, |
| "learning_rate": 1.2572673083178448e-05, |
| "loss": 0.098, |
| "step": 174 |
| }, |
| { |
| "epoch": 0.4827586206896552, |
| "grad_norm": 2.223070626201405, |
| "learning_rate": 1.2479430909430109e-05, |
| "loss": 0.0936, |
| "step": 175 |
| }, |
| { |
| "epoch": 0.48551724137931035, |
| "grad_norm": 2.1176352077764107, |
| "learning_rate": 1.2385958478494487e-05, |
| "loss": 0.1279, |
| "step": 176 |
| }, |
| { |
| "epoch": 0.4882758620689655, |
| "grad_norm": 1.3628256498935367, |
| "learning_rate": 1.2292264470871183e-05, |
| "loss": 0.0783, |
| "step": 177 |
| }, |
| { |
| "epoch": 0.4910344827586207, |
| "grad_norm": 1.064266823164612, |
| "learning_rate": 1.2198357587636958e-05, |
| "loss": 0.0525, |
| "step": 178 |
| }, |
| { |
| "epoch": 0.49379310344827587, |
| "grad_norm": 2.00945904980985, |
| "learning_rate": 1.2104246549637683e-05, |
| "loss": 0.0978, |
| "step": 179 |
| }, |
| { |
| "epoch": 0.496551724137931, |
| "grad_norm": 1.3141391526330848, |
| "learning_rate": 1.2009940096678451e-05, |
| "loss": 0.0963, |
| "step": 180 |
| }, |
| { |
| "epoch": 0.4993103448275862, |
| "grad_norm": 1.6531005740449383, |
| "learning_rate": 1.1915446986711953e-05, |
| "loss": 0.1092, |
| "step": 181 |
| }, |
| { |
| "epoch": 0.5020689655172413, |
| "grad_norm": 1.4945974838001637, |
| "learning_rate": 1.1820775995025147e-05, |
| "loss": 0.0667, |
| "step": 182 |
| }, |
| { |
| "epoch": 0.5048275862068966, |
| "grad_norm": 1.3762608443334923, |
| "learning_rate": 1.172593591342432e-05, |
| "loss": 0.0882, |
| "step": 183 |
| }, |
| { |
| "epoch": 0.5075862068965518, |
| "grad_norm": 1.5415877422060749, |
| "learning_rate": 1.1630935549418627e-05, |
| "loss": 0.0865, |
| "step": 184 |
| }, |
| { |
| "epoch": 0.5103448275862069, |
| "grad_norm": 2.0042018778537622, |
| "learning_rate": 1.1535783725402163e-05, |
| "loss": 0.1353, |
| "step": 185 |
| }, |
| { |
| "epoch": 0.5131034482758621, |
| "grad_norm": 2.6818407651740266, |
| "learning_rate": 1.1440489277834645e-05, |
| "loss": 0.1393, |
| "step": 186 |
| }, |
| { |
| "epoch": 0.5158620689655172, |
| "grad_norm": 2.3807831512162374, |
| "learning_rate": 1.134506105642081e-05, |
| "loss": 0.1169, |
| "step": 187 |
| }, |
| { |
| "epoch": 0.5186206896551724, |
| "grad_norm": 2.202610904326787, |
| "learning_rate": 1.1249507923288563e-05, |
| "loss": 0.1241, |
| "step": 188 |
| }, |
| { |
| "epoch": 0.5213793103448275, |
| "grad_norm": 1.8110167173727552, |
| "learning_rate": 1.115383875216598e-05, |
| "loss": 0.0945, |
| "step": 189 |
| }, |
| { |
| "epoch": 0.5241379310344828, |
| "grad_norm": 1.2395129394737805, |
| "learning_rate": 1.105806242755723e-05, |
| "loss": 0.0791, |
| "step": 190 |
| }, |
| { |
| "epoch": 0.526896551724138, |
| "grad_norm": 2.269164321603063, |
| "learning_rate": 1.0962187843917498e-05, |
| "loss": 0.0995, |
| "step": 191 |
| }, |
| { |
| "epoch": 0.5296551724137931, |
| "grad_norm": 2.1091615955126217, |
| "learning_rate": 1.0866223904826992e-05, |
| "loss": 0.0978, |
| "step": 192 |
| }, |
| { |
| "epoch": 0.5324137931034483, |
| "grad_norm": 1.1661181979597077, |
| "learning_rate": 1.0770179522164079e-05, |
| "loss": 0.0654, |
| "step": 193 |
| }, |
| { |
| "epoch": 0.5351724137931034, |
| "grad_norm": 1.529220319087916, |
| "learning_rate": 1.0674063615277681e-05, |
| "loss": 0.0795, |
| "step": 194 |
| }, |
| { |
| "epoch": 0.5379310344827586, |
| "grad_norm": 1.61462843303553, |
| "learning_rate": 1.0577885110158959e-05, |
| "loss": 0.0722, |
| "step": 195 |
| }, |
| { |
| "epoch": 0.5406896551724137, |
| "grad_norm": 1.0229094996985268, |
| "learning_rate": 1.0481652938612374e-05, |
| "loss": 0.1059, |
| "step": 196 |
| }, |
| { |
| "epoch": 0.543448275862069, |
| "grad_norm": 1.8905308810375994, |
| "learning_rate": 1.0385376037426227e-05, |
| "loss": 0.1007, |
| "step": 197 |
| }, |
| { |
| "epoch": 0.5462068965517242, |
| "grad_norm": 1.4659160591839386, |
| "learning_rate": 1.0289063347542727e-05, |
| "loss": 0.0997, |
| "step": 198 |
| }, |
| { |
| "epoch": 0.5489655172413793, |
| "grad_norm": 3.3227958982974637, |
| "learning_rate": 1.0192723813227672e-05, |
| "loss": 0.1803, |
| "step": 199 |
| }, |
| { |
| "epoch": 0.5517241379310345, |
| "grad_norm": 1.963767479554977, |
| "learning_rate": 1.0096366381239808e-05, |
| "loss": 0.1275, |
| "step": 200 |
| }, |
| { |
| "epoch": 0.5544827586206896, |
| "grad_norm": 1.6295769776645304, |
| "learning_rate": 1e-05, |
| "loss": 0.0784, |
| "step": 201 |
| }, |
| { |
| "epoch": 0.5572413793103448, |
| "grad_norm": 2.440021613726331, |
| "learning_rate": 9.903633618760195e-06, |
| "loss": 0.0884, |
| "step": 202 |
| }, |
| { |
| "epoch": 0.56, |
| "grad_norm": 1.4242019780978539, |
| "learning_rate": 9.807276186772335e-06, |
| "loss": 0.0852, |
| "step": 203 |
| }, |
| { |
| "epoch": 0.5627586206896552, |
| "grad_norm": 0.9963914089293635, |
| "learning_rate": 9.710936652457276e-06, |
| "loss": 0.0528, |
| "step": 204 |
| }, |
| { |
| "epoch": 0.5655172413793104, |
| "grad_norm": 2.0171601668683863, |
| "learning_rate": 9.614623962573776e-06, |
| "loss": 0.1213, |
| "step": 205 |
| }, |
| { |
| "epoch": 0.5682758620689655, |
| "grad_norm": 1.8302600165169711, |
| "learning_rate": 9.518347061387629e-06, |
| "loss": 0.0813, |
| "step": 206 |
| }, |
| { |
| "epoch": 0.5710344827586207, |
| "grad_norm": 2.135574797850369, |
| "learning_rate": 9.422114889841045e-06, |
| "loss": 0.1114, |
| "step": 207 |
| }, |
| { |
| "epoch": 0.5737931034482758, |
| "grad_norm": 1.856147430619363, |
| "learning_rate": 9.325936384722322e-06, |
| "loss": 0.1072, |
| "step": 208 |
| }, |
| { |
| "epoch": 0.576551724137931, |
| "grad_norm": 1.3734175489711284, |
| "learning_rate": 9.229820477835926e-06, |
| "loss": 0.0629, |
| "step": 209 |
| }, |
| { |
| "epoch": 0.5793103448275863, |
| "grad_norm": 1.4955582653472665, |
| "learning_rate": 9.133776095173015e-06, |
| "loss": 0.0798, |
| "step": 210 |
| }, |
| { |
| "epoch": 0.5820689655172414, |
| "grad_norm": 2.2098641671152075, |
| "learning_rate": 9.037812156082503e-06, |
| "loss": 0.0832, |
| "step": 211 |
| }, |
| { |
| "epoch": 0.5848275862068966, |
| "grad_norm": 2.1234204791453446, |
| "learning_rate": 8.941937572442773e-06, |
| "loss": 0.0912, |
| "step": 212 |
| }, |
| { |
| "epoch": 0.5875862068965517, |
| "grad_norm": 1.7048969824409477, |
| "learning_rate": 8.846161247834024e-06, |
| "loss": 0.0757, |
| "step": 213 |
| }, |
| { |
| "epoch": 0.5903448275862069, |
| "grad_norm": 1.2386625705533991, |
| "learning_rate": 8.750492076711439e-06, |
| "loss": 0.0607, |
| "step": 214 |
| }, |
| { |
| "epoch": 0.593103448275862, |
| "grad_norm": 2.4577783979073518, |
| "learning_rate": 8.654938943579194e-06, |
| "loss": 0.1315, |
| "step": 215 |
| }, |
| { |
| "epoch": 0.5958620689655172, |
| "grad_norm": 1.7822551401830198, |
| "learning_rate": 8.55951072216536e-06, |
| "loss": 0.1015, |
| "step": 216 |
| }, |
| { |
| "epoch": 0.5986206896551725, |
| "grad_norm": 1.627034418887004, |
| "learning_rate": 8.464216274597839e-06, |
| "loss": 0.083, |
| "step": 217 |
| }, |
| { |
| "epoch": 0.6013793103448276, |
| "grad_norm": 2.4236805021780907, |
| "learning_rate": 8.369064450581374e-06, |
| "loss": 0.144, |
| "step": 218 |
| }, |
| { |
| "epoch": 0.6041379310344828, |
| "grad_norm": 2.0852606023844547, |
| "learning_rate": 8.274064086575682e-06, |
| "loss": 0.1187, |
| "step": 219 |
| }, |
| { |
| "epoch": 0.6068965517241379, |
| "grad_norm": 1.554124619394613, |
| "learning_rate": 8.179224004974857e-06, |
| "loss": 0.101, |
| "step": 220 |
| }, |
| { |
| "epoch": 0.6096551724137931, |
| "grad_norm": 1.0086858963161975, |
| "learning_rate": 8.084553013288048e-06, |
| "loss": 0.0562, |
| "step": 221 |
| }, |
| { |
| "epoch": 0.6124137931034482, |
| "grad_norm": 1.744957411692858, |
| "learning_rate": 7.990059903321554e-06, |
| "loss": 0.1144, |
| "step": 222 |
| }, |
| { |
| "epoch": 0.6151724137931035, |
| "grad_norm": 1.115741854823727, |
| "learning_rate": 7.89575345036232e-06, |
| "loss": 0.0431, |
| "step": 223 |
| }, |
| { |
| "epoch": 0.6179310344827587, |
| "grad_norm": 2.4048467960954523, |
| "learning_rate": 7.801642412363042e-06, |
| "loss": 0.1239, |
| "step": 224 |
| }, |
| { |
| "epoch": 0.6206896551724138, |
| "grad_norm": 1.7045994682708523, |
| "learning_rate": 7.707735529128819e-06, |
| "loss": 0.0861, |
| "step": 225 |
| }, |
| { |
| "epoch": 0.623448275862069, |
| "grad_norm": 1.5844433019812807, |
| "learning_rate": 7.614041521505517e-06, |
| "loss": 0.0939, |
| "step": 226 |
| }, |
| { |
| "epoch": 0.6262068965517241, |
| "grad_norm": 1.817274462365455, |
| "learning_rate": 7.520569090569894e-06, |
| "loss": 0.0961, |
| "step": 227 |
| }, |
| { |
| "epoch": 0.6289655172413793, |
| "grad_norm": 2.169489761841419, |
| "learning_rate": 7.427326916821557e-06, |
| "loss": 0.1141, |
| "step": 228 |
| }, |
| { |
| "epoch": 0.6317241379310344, |
| "grad_norm": 1.2611061121212517, |
| "learning_rate": 7.3343236593768295e-06, |
| "loss": 0.0631, |
| "step": 229 |
| }, |
| { |
| "epoch": 0.6344827586206897, |
| "grad_norm": 1.4981316942094398, |
| "learning_rate": 7.24156795516461e-06, |
| "loss": 0.0607, |
| "step": 230 |
| }, |
| { |
| "epoch": 0.6372413793103449, |
| "grad_norm": 1.889688470210782, |
| "learning_rate": 7.149068418124281e-06, |
| "loss": 0.1243, |
| "step": 231 |
| }, |
| { |
| "epoch": 0.64, |
| "grad_norm": 1.6091621479461797, |
| "learning_rate": 7.056833638405762e-06, |
| "loss": 0.1154, |
| "step": 232 |
| }, |
| { |
| "epoch": 0.6427586206896552, |
| "grad_norm": 1.7584988460897566, |
| "learning_rate": 6.964872181571765e-06, |
| "loss": 0.0782, |
| "step": 233 |
| }, |
| { |
| "epoch": 0.6455172413793103, |
| "grad_norm": 1.9608264553670172, |
| "learning_rate": 6.87319258780234e-06, |
| "loss": 0.1088, |
| "step": 234 |
| }, |
| { |
| "epoch": 0.6482758620689655, |
| "grad_norm": 1.1583598512074385, |
| "learning_rate": 6.781803371101774e-06, |
| "loss": 0.0635, |
| "step": 235 |
| }, |
| { |
| "epoch": 0.6510344827586206, |
| "grad_norm": 1.5744270190667782, |
| "learning_rate": 6.690713018507917e-06, |
| "loss": 0.065, |
| "step": 236 |
| }, |
| { |
| "epoch": 0.6537931034482759, |
| "grad_norm": 2.0489496115781147, |
| "learning_rate": 6.599929989304034e-06, |
| "loss": 0.0941, |
| "step": 237 |
| }, |
| { |
| "epoch": 0.6565517241379311, |
| "grad_norm": 1.3833240601648478, |
| "learning_rate": 6.509462714233194e-06, |
| "loss": 0.1049, |
| "step": 238 |
| }, |
| { |
| "epoch": 0.6593103448275862, |
| "grad_norm": 1.105761243006651, |
| "learning_rate": 6.419319594715338e-06, |
| "loss": 0.0795, |
| "step": 239 |
| }, |
| { |
| "epoch": 0.6620689655172414, |
| "grad_norm": 2.2243011538195323, |
| "learning_rate": 6.32950900206708e-06, |
| "loss": 0.1389, |
| "step": 240 |
| }, |
| { |
| "epoch": 0.6648275862068965, |
| "grad_norm": 1.6204675243320001, |
| "learning_rate": 6.240039276724273e-06, |
| "loss": 0.0904, |
| "step": 241 |
| }, |
| { |
| "epoch": 0.6675862068965517, |
| "grad_norm": 1.7416445698806022, |
| "learning_rate": 6.150918727467455e-06, |
| "loss": 0.0988, |
| "step": 242 |
| }, |
| { |
| "epoch": 0.670344827586207, |
| "grad_norm": 1.5865476897173794, |
| "learning_rate": 6.062155630650265e-06, |
| "loss": 0.0887, |
| "step": 243 |
| }, |
| { |
| "epoch": 0.6731034482758621, |
| "grad_norm": 1.0908181878235528, |
| "learning_rate": 5.973758229430806e-06, |
| "loss": 0.0406, |
| "step": 244 |
| }, |
| { |
| "epoch": 0.6758620689655173, |
| "grad_norm": 1.2409267373149415, |
| "learning_rate": 5.8857347330061545e-06, |
| "loss": 0.0613, |
| "step": 245 |
| }, |
| { |
| "epoch": 0.6786206896551724, |
| "grad_norm": 1.5366160191792417, |
| "learning_rate": 5.798093315849984e-06, |
| "loss": 0.0621, |
| "step": 246 |
| }, |
| { |
| "epoch": 0.6813793103448276, |
| "grad_norm": 2.401971583677582, |
| "learning_rate": 5.7108421169534376e-06, |
| "loss": 0.135, |
| "step": 247 |
| }, |
| { |
| "epoch": 0.6841379310344827, |
| "grad_norm": 1.7382959732076737, |
| "learning_rate": 5.623989239069275e-06, |
| "loss": 0.0912, |
| "step": 248 |
| }, |
| { |
| "epoch": 0.6868965517241379, |
| "grad_norm": 1.1164138593496515, |
| "learning_rate": 5.5375427479593945e-06, |
| "loss": 0.0481, |
| "step": 249 |
| }, |
| { |
| "epoch": 0.6896551724137931, |
| "grad_norm": 1.9207044725285578, |
| "learning_rate": 5.451510671645806e-06, |
| "loss": 0.0915, |
| "step": 250 |
| }, |
| { |
| "epoch": 0.6924137931034483, |
| "grad_norm": 1.4554330468757273, |
| "learning_rate": 5.3659009996650704e-06, |
| "loss": 0.0932, |
| "step": 251 |
| }, |
| { |
| "epoch": 0.6951724137931035, |
| "grad_norm": 1.5637751353358034, |
| "learning_rate": 5.280721682326349e-06, |
| "loss": 0.1258, |
| "step": 252 |
| }, |
| { |
| "epoch": 0.6979310344827586, |
| "grad_norm": 1.6864022667103549, |
| "learning_rate": 5.195980629973077e-06, |
| "loss": 0.136, |
| "step": 253 |
| }, |
| { |
| "epoch": 0.7006896551724138, |
| "grad_norm": 1.5424417657042435, |
| "learning_rate": 5.111685712248364e-06, |
| "loss": 0.0744, |
| "step": 254 |
| }, |
| { |
| "epoch": 0.7034482758620689, |
| "grad_norm": 2.4152541945255965, |
| "learning_rate": 5.02784475736415e-06, |
| "loss": 0.1436, |
| "step": 255 |
| }, |
| { |
| "epoch": 0.7062068965517241, |
| "grad_norm": 1.8800510082119968, |
| "learning_rate": 4.944465551374238e-06, |
| "loss": 0.0864, |
| "step": 256 |
| }, |
| { |
| "epoch": 0.7089655172413794, |
| "grad_norm": 1.6391965299287088, |
| "learning_rate": 4.861555837451213e-06, |
| "loss": 0.0931, |
| "step": 257 |
| }, |
| { |
| "epoch": 0.7117241379310345, |
| "grad_norm": 2.1164230178199688, |
| "learning_rate": 4.779123315167362e-06, |
| "loss": 0.1589, |
| "step": 258 |
| }, |
| { |
| "epoch": 0.7144827586206897, |
| "grad_norm": 2.1736287568680615, |
| "learning_rate": 4.6971756397796506e-06, |
| "loss": 0.075, |
| "step": 259 |
| }, |
| { |
| "epoch": 0.7172413793103448, |
| "grad_norm": 0.981201558647046, |
| "learning_rate": 4.61572042151878e-06, |
| "loss": 0.0423, |
| "step": 260 |
| }, |
| { |
| "epoch": 0.72, |
| "grad_norm": 1.2068505093305162, |
| "learning_rate": 4.534765224882463e-06, |
| "loss": 0.0538, |
| "step": 261 |
| }, |
| { |
| "epoch": 0.7227586206896551, |
| "grad_norm": 0.9541230040360105, |
| "learning_rate": 4.4543175679329345e-06, |
| "loss": 0.0591, |
| "step": 262 |
| }, |
| { |
| "epoch": 0.7255172413793104, |
| "grad_norm": 1.7162320823485342, |
| "learning_rate": 4.37438492159876e-06, |
| "loss": 0.0965, |
| "step": 263 |
| }, |
| { |
| "epoch": 0.7282758620689656, |
| "grad_norm": 1.184768855122373, |
| "learning_rate": 4.294974708981041e-06, |
| "loss": 0.0383, |
| "step": 264 |
| }, |
| { |
| "epoch": 0.7310344827586207, |
| "grad_norm": 1.5098290827125584, |
| "learning_rate": 4.216094304664056e-06, |
| "loss": 0.0952, |
| "step": 265 |
| }, |
| { |
| "epoch": 0.7337931034482759, |
| "grad_norm": 1.3641276614652302, |
| "learning_rate": 4.1377510340304e-06, |
| "loss": 0.0687, |
| "step": 266 |
| }, |
| { |
| "epoch": 0.736551724137931, |
| "grad_norm": 1.9886328182742048, |
| "learning_rate": 4.059952172580694e-06, |
| "loss": 0.1397, |
| "step": 267 |
| }, |
| { |
| "epoch": 0.7393103448275862, |
| "grad_norm": 2.0623334755221494, |
| "learning_rate": 3.982704945257957e-06, |
| "loss": 0.1271, |
| "step": 268 |
| }, |
| { |
| "epoch": 0.7420689655172413, |
| "grad_norm": 2.027142606530138, |
| "learning_rate": 3.9060165257766116e-06, |
| "loss": 0.0736, |
| "step": 269 |
| }, |
| { |
| "epoch": 0.7448275862068966, |
| "grad_norm": 1.8884432574197467, |
| "learning_rate": 3.829894035956306e-06, |
| "loss": 0.1243, |
| "step": 270 |
| }, |
| { |
| "epoch": 0.7475862068965518, |
| "grad_norm": 1.5259208901463874, |
| "learning_rate": 3.754344545060529e-06, |
| "loss": 0.0971, |
| "step": 271 |
| }, |
| { |
| "epoch": 0.7503448275862069, |
| "grad_norm": 1.454648528529707, |
| "learning_rate": 3.6793750691400996e-06, |
| "loss": 0.1226, |
| "step": 272 |
| }, |
| { |
| "epoch": 0.7531034482758621, |
| "grad_norm": 1.1852288533881499, |
| "learning_rate": 3.604992570381621e-06, |
| "loss": 0.0573, |
| "step": 273 |
| }, |
| { |
| "epoch": 0.7558620689655172, |
| "grad_norm": 0.9888790128432269, |
| "learning_rate": 3.5312039564609203e-06, |
| "loss": 0.0357, |
| "step": 274 |
| }, |
| { |
| "epoch": 0.7586206896551724, |
| "grad_norm": 1.9227683916123242, |
| "learning_rate": 3.458016079901544e-06, |
| "loss": 0.0841, |
| "step": 275 |
| }, |
| { |
| "epoch": 0.7613793103448275, |
| "grad_norm": 1.6925865088399465, |
| "learning_rate": 3.3854357374383905e-06, |
| "loss": 0.0749, |
| "step": 276 |
| }, |
| { |
| "epoch": 0.7641379310344828, |
| "grad_norm": 2.482457504595889, |
| "learning_rate": 3.313469669386532e-06, |
| "loss": 0.1234, |
| "step": 277 |
| }, |
| { |
| "epoch": 0.766896551724138, |
| "grad_norm": 1.2830007481095762, |
| "learning_rate": 3.242124559015234e-06, |
| "loss": 0.0752, |
| "step": 278 |
| }, |
| { |
| "epoch": 0.7696551724137931, |
| "grad_norm": 2.0768576257082114, |
| "learning_rate": 3.171407031927325e-06, |
| "loss": 0.0991, |
| "step": 279 |
| }, |
| { |
| "epoch": 0.7724137931034483, |
| "grad_norm": 1.8987466668093202, |
| "learning_rate": 3.101323655443882e-06, |
| "loss": 0.109, |
| "step": 280 |
| }, |
| { |
| "epoch": 0.7751724137931034, |
| "grad_norm": 1.6001709673447158, |
| "learning_rate": 3.0318809379943594e-06, |
| "loss": 0.0599, |
| "step": 281 |
| }, |
| { |
| "epoch": 0.7779310344827586, |
| "grad_norm": 1.309542383665605, |
| "learning_rate": 2.9630853285121506e-06, |
| "loss": 0.0641, |
| "step": 282 |
| }, |
| { |
| "epoch": 0.7806896551724138, |
| "grad_norm": 1.5565630583605414, |
| "learning_rate": 2.8949432158357083e-06, |
| "loss": 0.0909, |
| "step": 283 |
| }, |
| { |
| "epoch": 0.783448275862069, |
| "grad_norm": 1.2776706094735155, |
| "learning_rate": 2.8274609281152322e-06, |
| "loss": 0.1223, |
| "step": 284 |
| }, |
| { |
| "epoch": 0.7862068965517242, |
| "grad_norm": 1.6300368148462467, |
| "learning_rate": 2.7606447322249876e-06, |
| "loss": 0.087, |
| "step": 285 |
| }, |
| { |
| "epoch": 0.7889655172413793, |
| "grad_norm": 1.5096693670735182, |
| "learning_rate": 2.694500833181323e-06, |
| "loss": 0.1002, |
| "step": 286 |
| }, |
| { |
| "epoch": 0.7917241379310345, |
| "grad_norm": 1.719399247233759, |
| "learning_rate": 2.629035373566433e-06, |
| "loss": 0.0915, |
| "step": 287 |
| }, |
| { |
| "epoch": 0.7944827586206896, |
| "grad_norm": 1.1425139227274304, |
| "learning_rate": 2.5642544329579088e-06, |
| "loss": 0.0555, |
| "step": 288 |
| }, |
| { |
| "epoch": 0.7972413793103448, |
| "grad_norm": 1.2082856945935607, |
| "learning_rate": 2.500164027364147e-06, |
| "loss": 0.0555, |
| "step": 289 |
| }, |
| { |
| "epoch": 0.8, |
| "grad_norm": 2.028172728126609, |
| "learning_rate": 2.4367701086656625e-06, |
| "loss": 0.104, |
| "step": 290 |
| }, |
| { |
| "epoch": 0.8027586206896552, |
| "grad_norm": 1.498164820422529, |
| "learning_rate": 2.374078564062364e-06, |
| "loss": 0.0936, |
| "step": 291 |
| }, |
| { |
| "epoch": 0.8055172413793104, |
| "grad_norm": 1.3254283577777912, |
| "learning_rate": 2.312095215526814e-06, |
| "loss": 0.0898, |
| "step": 292 |
| }, |
| { |
| "epoch": 0.8082758620689655, |
| "grad_norm": 1.3758428200048072, |
| "learning_rate": 2.2508258192635614e-06, |
| "loss": 0.073, |
| "step": 293 |
| }, |
| { |
| "epoch": 0.8110344827586207, |
| "grad_norm": 1.2825112587820704, |
| "learning_rate": 2.190276065174596e-06, |
| "loss": 0.0675, |
| "step": 294 |
| }, |
| { |
| "epoch": 0.8137931034482758, |
| "grad_norm": 1.704844614821693, |
| "learning_rate": 2.130451576330925e-06, |
| "loss": 0.0871, |
| "step": 295 |
| }, |
| { |
| "epoch": 0.8165517241379311, |
| "grad_norm": 1.8236093781558738, |
| "learning_rate": 2.0713579084503877e-06, |
| "loss": 0.075, |
| "step": 296 |
| }, |
| { |
| "epoch": 0.8193103448275862, |
| "grad_norm": 1.7159210186184939, |
| "learning_rate": 2.0130005493817063e-06, |
| "loss": 0.0726, |
| "step": 297 |
| }, |
| { |
| "epoch": 0.8220689655172414, |
| "grad_norm": 1.5402879375212146, |
| "learning_rate": 1.9553849185948514e-06, |
| "loss": 0.0585, |
| "step": 298 |
| }, |
| { |
| "epoch": 0.8248275862068966, |
| "grad_norm": 2.088587364963122, |
| "learning_rate": 1.8985163666777473e-06, |
| "loss": 0.1192, |
| "step": 299 |
| }, |
| { |
| "epoch": 0.8275862068965517, |
| "grad_norm": 1.385102288521542, |
| "learning_rate": 1.8424001748393905e-06, |
| "loss": 0.0639, |
| "step": 300 |
| }, |
| { |
| "epoch": 0.8303448275862069, |
| "grad_norm": 1.199103692035072, |
| "learning_rate": 1.7870415544193808e-06, |
| "loss": 0.0644, |
| "step": 301 |
| }, |
| { |
| "epoch": 0.833103448275862, |
| "grad_norm": 1.576550399440244, |
| "learning_rate": 1.7324456464039751e-06, |
| "loss": 0.0806, |
| "step": 302 |
| }, |
| { |
| "epoch": 0.8358620689655173, |
| "grad_norm": 1.9033469567963237, |
| "learning_rate": 1.6786175209486565e-06, |
| "loss": 0.1166, |
| "step": 303 |
| }, |
| { |
| "epoch": 0.8386206896551724, |
| "grad_norm": 1.5481679752062283, |
| "learning_rate": 1.6255621769072805e-06, |
| "loss": 0.0883, |
| "step": 304 |
| }, |
| { |
| "epoch": 0.8413793103448276, |
| "grad_norm": 1.7374538978001977, |
| "learning_rate": 1.5732845413678477e-06, |
| "loss": 0.105, |
| "step": 305 |
| }, |
| { |
| "epoch": 0.8441379310344828, |
| "grad_norm": 1.3465892642582866, |
| "learning_rate": 1.521789469194952e-06, |
| "loss": 0.0618, |
| "step": 306 |
| }, |
| { |
| "epoch": 0.8468965517241379, |
| "grad_norm": 2.312489005340463, |
| "learning_rate": 1.4710817425789015e-06, |
| "loss": 0.0992, |
| "step": 307 |
| }, |
| { |
| "epoch": 0.8496551724137931, |
| "grad_norm": 1.1318530850342379, |
| "learning_rate": 1.4211660705916286e-06, |
| "loss": 0.0458, |
| "step": 308 |
| }, |
| { |
| "epoch": 0.8524137931034482, |
| "grad_norm": 1.5063935617388766, |
| "learning_rate": 1.372047088749372e-06, |
| "loss": 0.0516, |
| "step": 309 |
| }, |
| { |
| "epoch": 0.8551724137931035, |
| "grad_norm": 1.4001415457936668, |
| "learning_rate": 1.3237293585821786e-06, |
| "loss": 0.1156, |
| "step": 310 |
| }, |
| { |
| "epoch": 0.8579310344827586, |
| "grad_norm": 2.487040806365276, |
| "learning_rate": 1.2762173672102996e-06, |
| "loss": 0.1332, |
| "step": 311 |
| }, |
| { |
| "epoch": 0.8606896551724138, |
| "grad_norm": 1.514447174356807, |
| "learning_rate": 1.2295155269274827e-06, |
| "loss": 0.0735, |
| "step": 312 |
| }, |
| { |
| "epoch": 0.863448275862069, |
| "grad_norm": 1.9664878278885487, |
| "learning_rate": 1.1836281747912125e-06, |
| "loss": 0.0825, |
| "step": 313 |
| }, |
| { |
| "epoch": 0.8662068965517241, |
| "grad_norm": 1.7248118984472842, |
| "learning_rate": 1.1385595722199438e-06, |
| "loss": 0.0945, |
| "step": 314 |
| }, |
| { |
| "epoch": 0.8689655172413793, |
| "grad_norm": 1.232176840002336, |
| "learning_rate": 1.094313904597355e-06, |
| "loss": 0.0761, |
| "step": 315 |
| }, |
| { |
| "epoch": 0.8717241379310345, |
| "grad_norm": 2.3846939660082636, |
| "learning_rate": 1.0508952808836682e-06, |
| "loss": 0.1104, |
| "step": 316 |
| }, |
| { |
| "epoch": 0.8744827586206897, |
| "grad_norm": 1.5107849502345858, |
| "learning_rate": 1.0083077332340563e-06, |
| "loss": 0.137, |
| "step": 317 |
| }, |
| { |
| "epoch": 0.8772413793103448, |
| "grad_norm": 2.5218758909483077, |
| "learning_rate": 9.665552166241965e-07, |
| "loss": 0.1395, |
| "step": 318 |
| }, |
| { |
| "epoch": 0.88, |
| "grad_norm": 1.913603235284768, |
| "learning_rate": 9.256416084829778e-07, |
| "loss": 0.1331, |
| "step": 319 |
| }, |
| { |
| "epoch": 0.8827586206896552, |
| "grad_norm": 2.4049905043471806, |
| "learning_rate": 8.855707083324183e-07, |
| "loss": 0.1077, |
| "step": 320 |
| }, |
| { |
| "epoch": 0.8855172413793103, |
| "grad_norm": 1.108589495662786, |
| "learning_rate": 8.46346237434813e-07, |
| "loss": 0.0638, |
| "step": 321 |
| }, |
| { |
| "epoch": 0.8882758620689655, |
| "grad_norm": 1.222025656507595, |
| "learning_rate": 8.079718384471557e-07, |
| "loss": 0.0515, |
| "step": 322 |
| }, |
| { |
| "epoch": 0.8910344827586207, |
| "grad_norm": 1.6319377780473996, |
| "learning_rate": 7.704510750828542e-07, |
| "loss": 0.0623, |
| "step": 323 |
| }, |
| { |
| "epoch": 0.8937931034482759, |
| "grad_norm": 1.402983153361783, |
| "learning_rate": 7.337874317807803e-07, |
| "loss": 0.0516, |
| "step": 324 |
| }, |
| { |
| "epoch": 0.896551724137931, |
| "grad_norm": 1.4500270750731776, |
| "learning_rate": 6.979843133816744e-07, |
| "loss": 0.0612, |
| "step": 325 |
| }, |
| { |
| "epoch": 0.8993103448275862, |
| "grad_norm": 1.095211249638046, |
| "learning_rate": 6.630450448119618e-07, |
| "loss": 0.0333, |
| "step": 326 |
| }, |
| { |
| "epoch": 0.9020689655172414, |
| "grad_norm": 1.8925065532997027, |
| "learning_rate": 6.289728707749609e-07, |
| "loss": 0.0953, |
| "step": 327 |
| }, |
| { |
| "epoch": 0.9048275862068965, |
| "grad_norm": 1.747624078137272, |
| "learning_rate": 5.957709554495683e-07, |
| "loss": 0.0893, |
| "step": 328 |
| }, |
| { |
| "epoch": 0.9075862068965517, |
| "grad_norm": 0.9113610050884592, |
| "learning_rate": 5.634423821964074e-07, |
| "loss": 0.0423, |
| "step": 329 |
| }, |
| { |
| "epoch": 0.9103448275862069, |
| "grad_norm": 1.6463257868897792, |
| "learning_rate": 5.319901532714877e-07, |
| "loss": 0.1055, |
| "step": 330 |
| }, |
| { |
| "epoch": 0.9131034482758621, |
| "grad_norm": 1.84950657136513, |
| "learning_rate": 5.014171895473929e-07, |
| "loss": 0.1232, |
| "step": 331 |
| }, |
| { |
| "epoch": 0.9158620689655173, |
| "grad_norm": 1.5561236996647523, |
| "learning_rate": 4.717263302420283e-07, |
| "loss": 0.076, |
| "step": 332 |
| }, |
| { |
| "epoch": 0.9186206896551724, |
| "grad_norm": 1.2285714349711996, |
| "learning_rate": 4.429203326549525e-07, |
| "loss": 0.0835, |
| "step": 333 |
| }, |
| { |
| "epoch": 0.9213793103448276, |
| "grad_norm": 2.8092718918728288, |
| "learning_rate": 4.150018719113147e-07, |
| "loss": 0.0734, |
| "step": 334 |
| }, |
| { |
| "epoch": 0.9241379310344827, |
| "grad_norm": 1.766273016183224, |
| "learning_rate": 3.8797354071342443e-07, |
| "loss": 0.0768, |
| "step": 335 |
| }, |
| { |
| "epoch": 0.926896551724138, |
| "grad_norm": 2.1729531887535285, |
| "learning_rate": 3.618378490999719e-07, |
| "loss": 0.1027, |
| "step": 336 |
| }, |
| { |
| "epoch": 0.9296551724137931, |
| "grad_norm": 1.7482937409483954, |
| "learning_rate": 3.365972242129378e-07, |
| "loss": 0.1176, |
| "step": 337 |
| }, |
| { |
| "epoch": 0.9324137931034483, |
| "grad_norm": 1.2860014321281648, |
| "learning_rate": 3.122540100721794e-07, |
| "loss": 0.068, |
| "step": 338 |
| }, |
| { |
| "epoch": 0.9351724137931035, |
| "grad_norm": 1.8682982161376196, |
| "learning_rate": 2.888104673577574e-07, |
| "loss": 0.0826, |
| "step": 339 |
| }, |
| { |
| "epoch": 0.9379310344827586, |
| "grad_norm": 2.1192579972688548, |
| "learning_rate": 2.66268773199988e-07, |
| "loss": 0.1094, |
| "step": 340 |
| }, |
| { |
| "epoch": 0.9406896551724138, |
| "grad_norm": 1.812620322889224, |
| "learning_rate": 2.4463102097726843e-07, |
| "loss": 0.1397, |
| "step": 341 |
| }, |
| { |
| "epoch": 0.9434482758620689, |
| "grad_norm": 1.55323368762906, |
| "learning_rate": 2.2389922012165944e-07, |
| "loss": 0.0883, |
| "step": 342 |
| }, |
| { |
| "epoch": 0.9462068965517242, |
| "grad_norm": 1.3537994599803411, |
| "learning_rate": 2.0407529593228114e-07, |
| "loss": 0.0537, |
| "step": 343 |
| }, |
| { |
| "epoch": 0.9489655172413793, |
| "grad_norm": 1.9963802514280435, |
| "learning_rate": 1.8516108939651945e-07, |
| "loss": 0.0841, |
| "step": 344 |
| }, |
| { |
| "epoch": 0.9517241379310345, |
| "grad_norm": 1.9489250790151131, |
| "learning_rate": 1.6715835701905604e-07, |
| "loss": 0.1278, |
| "step": 345 |
| }, |
| { |
| "epoch": 0.9544827586206897, |
| "grad_norm": 1.208403698921063, |
| "learning_rate": 1.5006877065874338e-07, |
| "loss": 0.062, |
| "step": 346 |
| }, |
| { |
| "epoch": 0.9572413793103448, |
| "grad_norm": 1.9902454269295837, |
| "learning_rate": 1.3389391737335112e-07, |
| "loss": 0.1123, |
| "step": 347 |
| }, |
| { |
| "epoch": 0.96, |
| "grad_norm": 1.5549649154456184, |
| "learning_rate": 1.1863529927217731e-07, |
| "loss": 0.0859, |
| "step": 348 |
| }, |
| { |
| "epoch": 0.9627586206896551, |
| "grad_norm": 1.5672800792400794, |
| "learning_rate": 1.0429433337655115e-07, |
| "loss": 0.0664, |
| "step": 349 |
| }, |
| { |
| "epoch": 0.9655172413793104, |
| "grad_norm": 1.980266389039632, |
| "learning_rate": 9.08723514882437e-08, |
| "loss": 0.0978, |
| "step": 350 |
| }, |
| { |
| "epoch": 0.9682758620689655, |
| "grad_norm": 1.924180909114094, |
| "learning_rate": 7.837060006577801e-08, |
| "loss": 0.082, |
| "step": 351 |
| }, |
| { |
| "epoch": 0.9710344827586207, |
| "grad_norm": 1.360244814652006, |
| "learning_rate": 6.679024010868617e-08, |
| "loss": 0.0937, |
| "step": 352 |
| }, |
| { |
| "epoch": 0.9737931034482759, |
| "grad_norm": 1.3638474049460683, |
| "learning_rate": 5.6132347049679955e-08, |
| "loss": 0.0944, |
| "step": 353 |
| }, |
| { |
| "epoch": 0.976551724137931, |
| "grad_norm": 1.0043906816455541, |
| "learning_rate": 4.639791065478738e-08, |
| "loss": 0.0373, |
| "step": 354 |
| }, |
| { |
| "epoch": 0.9793103448275862, |
| "grad_norm": 1.3136551040176567, |
| "learning_rate": 3.758783493142737e-08, |
| "loss": 0.0938, |
| "step": 355 |
| }, |
| { |
| "epoch": 0.9820689655172414, |
| "grad_norm": 1.2330770172633143, |
| "learning_rate": 2.9702938044468e-08, |
| "loss": 0.1083, |
| "step": 356 |
| }, |
| { |
| "epoch": 0.9848275862068966, |
| "grad_norm": 1.4318266541987408, |
| "learning_rate": 2.274395224023618e-08, |
| "loss": 0.0667, |
| "step": 357 |
| }, |
| { |
| "epoch": 0.9875862068965517, |
| "grad_norm": 1.4889881534099365, |
| "learning_rate": 1.671152377852092e-08, |
| "loss": 0.0907, |
| "step": 358 |
| }, |
| { |
| "epoch": 0.9903448275862069, |
| "grad_norm": 1.9950572073445185, |
| "learning_rate": 1.1606212872559142e-08, |
| "loss": 0.0937, |
| "step": 359 |
| }, |
| { |
| "epoch": 0.993103448275862, |
| "grad_norm": 1.2536334792292063, |
| "learning_rate": 7.42849363700282e-09, |
| "loss": 0.0863, |
| "step": 360 |
| }, |
| { |
| "epoch": 0.9958620689655172, |
| "grad_norm": 1.2816200143172458, |
| "learning_rate": 4.178754043898669e-09, |
| "loss": 0.0593, |
| "step": 361 |
| }, |
| { |
| "epoch": 0.9986206896551724, |
| "grad_norm": 0.7637431538197891, |
| "learning_rate": 1.8572958866514e-09, |
| "loss": 0.0859, |
| "step": 362 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 2.3907025925168472, |
| "learning_rate": 4.643347520005836e-10, |
| "loss": 0.0658, |
| "step": 363 |
| }, |
| { |
| "epoch": 1.0, |
| "step": 363, |
| "total_flos": 50754799140864.0, |
| "train_loss": 0.09636665488652289, |
| "train_runtime": 845.3624, |
| "train_samples_per_second": 6.855, |
| "train_steps_per_second": 0.429 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 363, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 1, |
| "save_steps": 100, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 50754799140864.0, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|