| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 4.996517107150174, |
| "eval_steps": 500, |
| "global_step": 760, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.006556033599672198, |
| "grad_norm": 6.308972265166019, |
| "learning_rate": 1.0526315789473685e-06, |
| "loss": 1.085, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.013112067199344397, |
| "grad_norm": 6.3669176951390085, |
| "learning_rate": 2.105263157894737e-06, |
| "loss": 1.0943, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.019668100799016593, |
| "grad_norm": 6.248172461223065, |
| "learning_rate": 3.157894736842105e-06, |
| "loss": 1.0945, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.026224134398688793, |
| "grad_norm": 5.720783796325484, |
| "learning_rate": 4.210526315789474e-06, |
| "loss": 1.0822, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.03278016799836099, |
| "grad_norm": 4.166580116763421, |
| "learning_rate": 5.263157894736842e-06, |
| "loss": 1.0376, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.03933620159803319, |
| "grad_norm": 2.5258613072829443, |
| "learning_rate": 6.31578947368421e-06, |
| "loss": 1.0088, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.04589223519770539, |
| "grad_norm": 3.241153982591279, |
| "learning_rate": 7.368421052631579e-06, |
| "loss": 0.9787, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.05244826879737759, |
| "grad_norm": 4.215997338140802, |
| "learning_rate": 8.421052631578948e-06, |
| "loss": 1.0021, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.05900430239704978, |
| "grad_norm": 3.7917272082574036, |
| "learning_rate": 9.473684210526315e-06, |
| "loss": 0.9883, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.06556033599672198, |
| "grad_norm": 3.1091596358257414, |
| "learning_rate": 1.0526315789473684e-05, |
| "loss": 0.9323, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.07211636959639418, |
| "grad_norm": 2.7854345292697746, |
| "learning_rate": 1.1578947368421053e-05, |
| "loss": 0.9234, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.07867240319606637, |
| "grad_norm": 1.8650709570608603, |
| "learning_rate": 1.263157894736842e-05, |
| "loss": 0.9023, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.08522843679573858, |
| "grad_norm": 1.574226758545731, |
| "learning_rate": 1.3684210526315791e-05, |
| "loss": 0.8693, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.09178447039541078, |
| "grad_norm": 1.4177768637857613, |
| "learning_rate": 1.4736842105263159e-05, |
| "loss": 0.8558, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.09834050399508297, |
| "grad_norm": 1.2125668795375901, |
| "learning_rate": 1.578947368421053e-05, |
| "loss": 0.8394, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.10489653759475517, |
| "grad_norm": 1.1398741474728522, |
| "learning_rate": 1.6842105263157896e-05, |
| "loss": 0.8373, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.11145257119442738, |
| "grad_norm": 1.0380532605315191, |
| "learning_rate": 1.7894736842105264e-05, |
| "loss": 0.8138, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.11800860479409957, |
| "grad_norm": 1.1944105144048436, |
| "learning_rate": 1.894736842105263e-05, |
| "loss": 0.8178, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.12456463839377177, |
| "grad_norm": 0.9889022037403388, |
| "learning_rate": 2e-05, |
| "loss": 0.803, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.13112067199344396, |
| "grad_norm": 0.7531044929728199, |
| "learning_rate": 2.105263157894737e-05, |
| "loss": 0.806, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.13767670559311618, |
| "grad_norm": 0.9451120842150798, |
| "learning_rate": 2.210526315789474e-05, |
| "loss": 0.7904, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.14423273919278837, |
| "grad_norm": 0.8762905130696029, |
| "learning_rate": 2.3157894736842107e-05, |
| "loss": 0.8001, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.15078877279246056, |
| "grad_norm": 0.6699562168478991, |
| "learning_rate": 2.4210526315789474e-05, |
| "loss": 0.7822, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.15734480639213275, |
| "grad_norm": 0.6276115227175743, |
| "learning_rate": 2.526315789473684e-05, |
| "loss": 0.7773, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.16390083999180496, |
| "grad_norm": 0.6290807862288751, |
| "learning_rate": 2.6315789473684215e-05, |
| "loss": 0.7783, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.17045687359147715, |
| "grad_norm": 0.5742564332284298, |
| "learning_rate": 2.7368421052631583e-05, |
| "loss": 0.7674, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.17701290719114934, |
| "grad_norm": 0.7412922803237626, |
| "learning_rate": 2.842105263157895e-05, |
| "loss": 0.7698, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.18356894079082156, |
| "grad_norm": 1.0547770655616728, |
| "learning_rate": 2.9473684210526317e-05, |
| "loss": 0.7655, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.19012497439049375, |
| "grad_norm": 1.5527104367024405, |
| "learning_rate": 3.052631578947369e-05, |
| "loss": 0.7707, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.19668100799016594, |
| "grad_norm": 0.7294937893627019, |
| "learning_rate": 3.157894736842106e-05, |
| "loss": 0.7578, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.20323704158983816, |
| "grad_norm": 0.9536771340964642, |
| "learning_rate": 3.2631578947368426e-05, |
| "loss": 0.748, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.20979307518951035, |
| "grad_norm": 1.6534518524100028, |
| "learning_rate": 3.368421052631579e-05, |
| "loss": 0.7588, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.21634910878918254, |
| "grad_norm": 0.6082911982013689, |
| "learning_rate": 3.473684210526316e-05, |
| "loss": 0.7424, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.22290514238885475, |
| "grad_norm": 1.4728558033024115, |
| "learning_rate": 3.578947368421053e-05, |
| "loss": 0.7423, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.22946117598852694, |
| "grad_norm": 0.8950252912362415, |
| "learning_rate": 3.6842105263157895e-05, |
| "loss": 0.7396, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.23601720958819913, |
| "grad_norm": 1.2202323688159324, |
| "learning_rate": 3.789473684210526e-05, |
| "loss": 0.7394, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.24257324318787135, |
| "grad_norm": 1.0375163505188818, |
| "learning_rate": 3.8947368421052636e-05, |
| "loss": 0.735, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.24912927678754354, |
| "grad_norm": 1.1817870027992188, |
| "learning_rate": 4e-05, |
| "loss": 0.7316, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.25568531038721576, |
| "grad_norm": 1.7472819001700346, |
| "learning_rate": 4.105263157894738e-05, |
| "loss": 0.7371, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.2622413439868879, |
| "grad_norm": 0.6997273968133122, |
| "learning_rate": 4.210526315789474e-05, |
| "loss": 0.7303, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.26879737758656014, |
| "grad_norm": 2.321753393397282, |
| "learning_rate": 4.315789473684211e-05, |
| "loss": 0.7458, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.27535341118623236, |
| "grad_norm": 1.366635958874213, |
| "learning_rate": 4.421052631578948e-05, |
| "loss": 0.7176, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.2819094447859045, |
| "grad_norm": 2.6212092693642552, |
| "learning_rate": 4.5263157894736846e-05, |
| "loss": 0.7315, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.28846547838557673, |
| "grad_norm": 2.5130277449596496, |
| "learning_rate": 4.6315789473684214e-05, |
| "loss": 0.753, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.29502151198524895, |
| "grad_norm": 1.6285728910727462, |
| "learning_rate": 4.736842105263158e-05, |
| "loss": 0.7297, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.3015775455849211, |
| "grad_norm": 2.0964313396830456, |
| "learning_rate": 4.842105263157895e-05, |
| "loss": 0.7265, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.30813357918459333, |
| "grad_norm": 2.0887773848236697, |
| "learning_rate": 4.947368421052632e-05, |
| "loss": 0.7239, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.3146896127842655, |
| "grad_norm": 1.1072655635857458, |
| "learning_rate": 5.052631578947368e-05, |
| "loss": 0.7193, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.3212456463839377, |
| "grad_norm": 1.4488623000707561, |
| "learning_rate": 5.157894736842106e-05, |
| "loss": 0.7203, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.32780167998360993, |
| "grad_norm": 1.2322683716693836, |
| "learning_rate": 5.263157894736843e-05, |
| "loss": 0.7238, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.3343577135832821, |
| "grad_norm": 1.38345500389746, |
| "learning_rate": 5.368421052631579e-05, |
| "loss": 0.7142, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.3409137471829543, |
| "grad_norm": 0.9094126438399451, |
| "learning_rate": 5.4736842105263165e-05, |
| "loss": 0.72, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.3474697807826265, |
| "grad_norm": 1.2022041950473747, |
| "learning_rate": 5.5789473684210526e-05, |
| "loss": 0.7075, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.3540258143822987, |
| "grad_norm": 1.6189179778780058, |
| "learning_rate": 5.68421052631579e-05, |
| "loss": 0.7105, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.3605818479819709, |
| "grad_norm": 1.2006446280190224, |
| "learning_rate": 5.789473684210527e-05, |
| "loss": 0.7151, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.3671378815816431, |
| "grad_norm": 1.6321920047861564, |
| "learning_rate": 5.8947368421052634e-05, |
| "loss": 0.7136, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.3736939151813153, |
| "grad_norm": 1.5287305591592921, |
| "learning_rate": 6.000000000000001e-05, |
| "loss": 0.7107, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.3802499487809875, |
| "grad_norm": 1.3257983915749183, |
| "learning_rate": 6.105263157894738e-05, |
| "loss": 0.7183, |
| "step": 58 |
| }, |
| { |
| "epoch": 0.3868059823806597, |
| "grad_norm": 0.8173893638911733, |
| "learning_rate": 6.210526315789474e-05, |
| "loss": 0.7087, |
| "step": 59 |
| }, |
| { |
| "epoch": 0.3933620159803319, |
| "grad_norm": 1.5027100216095572, |
| "learning_rate": 6.315789473684212e-05, |
| "loss": 0.7099, |
| "step": 60 |
| }, |
| { |
| "epoch": 0.3999180495800041, |
| "grad_norm": 0.9118886385672123, |
| "learning_rate": 6.421052631578948e-05, |
| "loss": 0.7056, |
| "step": 61 |
| }, |
| { |
| "epoch": 0.4064740831796763, |
| "grad_norm": 8.129014604966502, |
| "learning_rate": 6.526315789473685e-05, |
| "loss": 0.7359, |
| "step": 62 |
| }, |
| { |
| "epoch": 0.4130301167793485, |
| "grad_norm": 3.8240665538108343, |
| "learning_rate": 6.631578947368421e-05, |
| "loss": 0.7738, |
| "step": 63 |
| }, |
| { |
| "epoch": 0.4195861503790207, |
| "grad_norm": 3.710700934507178, |
| "learning_rate": 6.736842105263159e-05, |
| "loss": 0.7309, |
| "step": 64 |
| }, |
| { |
| "epoch": 0.4261421839786929, |
| "grad_norm": 1.7575755137455045, |
| "learning_rate": 6.842105263157895e-05, |
| "loss": 0.7373, |
| "step": 65 |
| }, |
| { |
| "epoch": 0.4326982175783651, |
| "grad_norm": 1.29443521888165, |
| "learning_rate": 6.947368421052632e-05, |
| "loss": 0.7149, |
| "step": 66 |
| }, |
| { |
| "epoch": 0.4392542511780373, |
| "grad_norm": 1.9730328727947009, |
| "learning_rate": 7.052631578947368e-05, |
| "loss": 0.7309, |
| "step": 67 |
| }, |
| { |
| "epoch": 0.4458102847777095, |
| "grad_norm": 1.5434748912530512, |
| "learning_rate": 7.157894736842105e-05, |
| "loss": 0.7307, |
| "step": 68 |
| }, |
| { |
| "epoch": 0.45236631837738167, |
| "grad_norm": 1.3098144472083486, |
| "learning_rate": 7.263157894736843e-05, |
| "loss": 0.7226, |
| "step": 69 |
| }, |
| { |
| "epoch": 0.4589223519770539, |
| "grad_norm": 1.4763764747327572, |
| "learning_rate": 7.368421052631579e-05, |
| "loss": 0.7122, |
| "step": 70 |
| }, |
| { |
| "epoch": 0.4654783855767261, |
| "grad_norm": 1.782399664190018, |
| "learning_rate": 7.473684210526316e-05, |
| "loss": 0.7105, |
| "step": 71 |
| }, |
| { |
| "epoch": 0.47203441917639827, |
| "grad_norm": 0.8645965374178922, |
| "learning_rate": 7.578947368421052e-05, |
| "loss": 0.7189, |
| "step": 72 |
| }, |
| { |
| "epoch": 0.4785904527760705, |
| "grad_norm": 1.1783141500345207, |
| "learning_rate": 7.68421052631579e-05, |
| "loss": 0.7155, |
| "step": 73 |
| }, |
| { |
| "epoch": 0.4851464863757427, |
| "grad_norm": 1.631405575969008, |
| "learning_rate": 7.789473684210527e-05, |
| "loss": 0.7163, |
| "step": 74 |
| }, |
| { |
| "epoch": 0.49170251997541486, |
| "grad_norm": 1.799584170211315, |
| "learning_rate": 7.894736842105263e-05, |
| "loss": 0.724, |
| "step": 75 |
| }, |
| { |
| "epoch": 0.4982585535750871, |
| "grad_norm": 131.28302876683892, |
| "learning_rate": 8e-05, |
| "loss": 0.9347, |
| "step": 76 |
| }, |
| { |
| "epoch": 0.5048145871747592, |
| "grad_norm": 2.879295120261663, |
| "learning_rate": 7.999957809295807e-05, |
| "loss": 0.7474, |
| "step": 77 |
| }, |
| { |
| "epoch": 0.5113706207744315, |
| "grad_norm": 0.9410659862896121, |
| "learning_rate": 7.99983123807325e-05, |
| "loss": 0.7124, |
| "step": 78 |
| }, |
| { |
| "epoch": 0.5179266543741037, |
| "grad_norm": 2.4498983574935487, |
| "learning_rate": 7.999620289002397e-05, |
| "loss": 0.7294, |
| "step": 79 |
| }, |
| { |
| "epoch": 0.5244826879737758, |
| "grad_norm": 1.3312297085268832, |
| "learning_rate": 7.999324966533291e-05, |
| "loss": 0.7155, |
| "step": 80 |
| }, |
| { |
| "epoch": 0.5310387215734481, |
| "grad_norm": 1.9333142275612973, |
| "learning_rate": 7.998945276895866e-05, |
| "loss": 0.7249, |
| "step": 81 |
| }, |
| { |
| "epoch": 0.5375947551731203, |
| "grad_norm": 1.5284340416931426, |
| "learning_rate": 7.998481228099806e-05, |
| "loss": 0.7199, |
| "step": 82 |
| }, |
| { |
| "epoch": 0.5441507887727924, |
| "grad_norm": 1.5374661844560404, |
| "learning_rate": 7.997932829934386e-05, |
| "loss": 0.7174, |
| "step": 83 |
| }, |
| { |
| "epoch": 0.5507068223724647, |
| "grad_norm": 1.099517635025565, |
| "learning_rate": 7.997300093968255e-05, |
| "loss": 0.7106, |
| "step": 84 |
| }, |
| { |
| "epoch": 0.5572628559721369, |
| "grad_norm": 1.4642952447149862, |
| "learning_rate": 7.996583033549204e-05, |
| "loss": 0.7087, |
| "step": 85 |
| }, |
| { |
| "epoch": 0.563818889571809, |
| "grad_norm": 1.4883363499473765, |
| "learning_rate": 7.995781663803876e-05, |
| "loss": 0.724, |
| "step": 86 |
| }, |
| { |
| "epoch": 0.5703749231714813, |
| "grad_norm": 0.8588229153701384, |
| "learning_rate": 7.994896001637443e-05, |
| "loss": 0.7052, |
| "step": 87 |
| }, |
| { |
| "epoch": 0.5769309567711535, |
| "grad_norm": 1.1329548393299773, |
| "learning_rate": 7.993926065733265e-05, |
| "loss": 0.7092, |
| "step": 88 |
| }, |
| { |
| "epoch": 0.5834869903708256, |
| "grad_norm": 1.9381285339000986, |
| "learning_rate": 7.99287187655248e-05, |
| "loss": 0.7075, |
| "step": 89 |
| }, |
| { |
| "epoch": 0.5900430239704979, |
| "grad_norm": 1.0298360769847832, |
| "learning_rate": 7.991733456333579e-05, |
| "loss": 0.7088, |
| "step": 90 |
| }, |
| { |
| "epoch": 0.5965990575701701, |
| "grad_norm": 1.5913971706858665, |
| "learning_rate": 7.990510829091938e-05, |
| "loss": 0.7044, |
| "step": 91 |
| }, |
| { |
| "epoch": 0.6031550911698422, |
| "grad_norm": 1.0752295835923937, |
| "learning_rate": 7.98920402061931e-05, |
| "loss": 0.698, |
| "step": 92 |
| }, |
| { |
| "epoch": 0.6097111247695144, |
| "grad_norm": 1.3411827568812442, |
| "learning_rate": 7.987813058483278e-05, |
| "loss": 0.6897, |
| "step": 93 |
| }, |
| { |
| "epoch": 0.6162671583691867, |
| "grad_norm": 1.0994663107424685, |
| "learning_rate": 7.98633797202668e-05, |
| "loss": 0.7009, |
| "step": 94 |
| }, |
| { |
| "epoch": 0.6228231919688588, |
| "grad_norm": 1.5572209303699358, |
| "learning_rate": 7.984778792366983e-05, |
| "loss": 0.7001, |
| "step": 95 |
| }, |
| { |
| "epoch": 0.629379225568531, |
| "grad_norm": 1.2622939681539915, |
| "learning_rate": 7.98313555239563e-05, |
| "loss": 0.7027, |
| "step": 96 |
| }, |
| { |
| "epoch": 0.6359352591682033, |
| "grad_norm": 1.5287178198020728, |
| "learning_rate": 7.98140828677735e-05, |
| "loss": 0.6977, |
| "step": 97 |
| }, |
| { |
| "epoch": 0.6424912927678754, |
| "grad_norm": 1.1853944350439, |
| "learning_rate": 7.979597031949415e-05, |
| "loss": 0.6943, |
| "step": 98 |
| }, |
| { |
| "epoch": 0.6490473263675476, |
| "grad_norm": 1.1219402341610256, |
| "learning_rate": 7.977701826120888e-05, |
| "loss": 0.6984, |
| "step": 99 |
| }, |
| { |
| "epoch": 0.6556033599672199, |
| "grad_norm": 1.8026221843535832, |
| "learning_rate": 7.975722709271799e-05, |
| "loss": 0.6955, |
| "step": 100 |
| }, |
| { |
| "epoch": 0.662159393566892, |
| "grad_norm": 1.0073031848270164, |
| "learning_rate": 7.973659723152317e-05, |
| "loss": 0.6942, |
| "step": 101 |
| }, |
| { |
| "epoch": 0.6687154271665642, |
| "grad_norm": 1.4709635304665, |
| "learning_rate": 7.97151291128186e-05, |
| "loss": 0.7024, |
| "step": 102 |
| }, |
| { |
| "epoch": 0.6752714607662365, |
| "grad_norm": 1.5710635409318867, |
| "learning_rate": 7.96928231894818e-05, |
| "loss": 0.6917, |
| "step": 103 |
| }, |
| { |
| "epoch": 0.6818274943659086, |
| "grad_norm": 0.8520728596618132, |
| "learning_rate": 7.96696799320641e-05, |
| "loss": 0.6821, |
| "step": 104 |
| }, |
| { |
| "epoch": 0.6883835279655808, |
| "grad_norm": 1.3459049216152963, |
| "learning_rate": 7.964569982878063e-05, |
| "loss": 0.6916, |
| "step": 105 |
| }, |
| { |
| "epoch": 0.694939561565253, |
| "grad_norm": 0.9826433336489447, |
| "learning_rate": 7.962088338550013e-05, |
| "loss": 0.6894, |
| "step": 106 |
| }, |
| { |
| "epoch": 0.7014955951649252, |
| "grad_norm": 1.2896938337982984, |
| "learning_rate": 7.959523112573422e-05, |
| "loss": 0.6933, |
| "step": 107 |
| }, |
| { |
| "epoch": 0.7080516287645974, |
| "grad_norm": 0.9867952374397386, |
| "learning_rate": 7.956874359062632e-05, |
| "loss": 0.697, |
| "step": 108 |
| }, |
| { |
| "epoch": 0.7146076623642696, |
| "grad_norm": 1.2639647577953248, |
| "learning_rate": 7.954142133894033e-05, |
| "loss": 0.6894, |
| "step": 109 |
| }, |
| { |
| "epoch": 0.7211636959639418, |
| "grad_norm": 1.357158506896498, |
| "learning_rate": 7.951326494704878e-05, |
| "loss": 0.691, |
| "step": 110 |
| }, |
| { |
| "epoch": 0.727719729563614, |
| "grad_norm": 0.8936367138602801, |
| "learning_rate": 7.948427500892065e-05, |
| "loss": 0.6887, |
| "step": 111 |
| }, |
| { |
| "epoch": 0.7342757631632862, |
| "grad_norm": 1.0046258611386971, |
| "learning_rate": 7.94544521361089e-05, |
| "loss": 0.6866, |
| "step": 112 |
| }, |
| { |
| "epoch": 0.7408317967629584, |
| "grad_norm": 1.0742250060740215, |
| "learning_rate": 7.942379695773753e-05, |
| "loss": 0.6888, |
| "step": 113 |
| }, |
| { |
| "epoch": 0.7473878303626306, |
| "grad_norm": 0.7284144457213086, |
| "learning_rate": 7.939231012048833e-05, |
| "loss": 0.6811, |
| "step": 114 |
| }, |
| { |
| "epoch": 0.7539438639623028, |
| "grad_norm": 0.7052427321018496, |
| "learning_rate": 7.93599922885872e-05, |
| "loss": 0.6763, |
| "step": 115 |
| }, |
| { |
| "epoch": 0.760499897561975, |
| "grad_norm": 0.798707901395817, |
| "learning_rate": 7.932684414379021e-05, |
| "loss": 0.692, |
| "step": 116 |
| }, |
| { |
| "epoch": 0.7670559311616472, |
| "grad_norm": 1.5879889624779187, |
| "learning_rate": 7.929286638536913e-05, |
| "loss": 0.6908, |
| "step": 117 |
| }, |
| { |
| "epoch": 0.7736119647613194, |
| "grad_norm": 0.8333519158322553, |
| "learning_rate": 7.925805973009672e-05, |
| "loss": 0.6734, |
| "step": 118 |
| }, |
| { |
| "epoch": 0.7801679983609916, |
| "grad_norm": 1.0218779201091506, |
| "learning_rate": 7.922242491223167e-05, |
| "loss": 0.684, |
| "step": 119 |
| }, |
| { |
| "epoch": 0.7867240319606638, |
| "grad_norm": 1.2082599250889219, |
| "learning_rate": 7.918596268350296e-05, |
| "loss": 0.6765, |
| "step": 120 |
| }, |
| { |
| "epoch": 0.793280065560336, |
| "grad_norm": 1.179921751372486, |
| "learning_rate": 7.914867381309418e-05, |
| "loss": 0.6868, |
| "step": 121 |
| }, |
| { |
| "epoch": 0.7998360991600082, |
| "grad_norm": 0.9486398114671882, |
| "learning_rate": 7.911055908762718e-05, |
| "loss": 0.6749, |
| "step": 122 |
| }, |
| { |
| "epoch": 0.8063921327596804, |
| "grad_norm": 0.8000482601006661, |
| "learning_rate": 7.90716193111455e-05, |
| "loss": 0.6747, |
| "step": 123 |
| }, |
| { |
| "epoch": 0.8129481663593526, |
| "grad_norm": 0.9517101245313929, |
| "learning_rate": 7.903185530509743e-05, |
| "loss": 0.6794, |
| "step": 124 |
| }, |
| { |
| "epoch": 0.8195041999590248, |
| "grad_norm": 0.9986534633574224, |
| "learning_rate": 7.899126790831869e-05, |
| "loss": 0.6774, |
| "step": 125 |
| }, |
| { |
| "epoch": 0.826060233558697, |
| "grad_norm": 1.1257859217497668, |
| "learning_rate": 7.894985797701472e-05, |
| "loss": 0.6793, |
| "step": 126 |
| }, |
| { |
| "epoch": 0.8326162671583692, |
| "grad_norm": 1.0292173321571256, |
| "learning_rate": 7.890762638474256e-05, |
| "loss": 0.6826, |
| "step": 127 |
| }, |
| { |
| "epoch": 0.8391723007580414, |
| "grad_norm": 0.7557871567857994, |
| "learning_rate": 7.886457402239256e-05, |
| "loss": 0.6792, |
| "step": 128 |
| }, |
| { |
| "epoch": 0.8457283343577136, |
| "grad_norm": 1.10326991797423, |
| "learning_rate": 7.882070179816944e-05, |
| "loss": 0.6786, |
| "step": 129 |
| }, |
| { |
| "epoch": 0.8522843679573858, |
| "grad_norm": 0.6776535159011668, |
| "learning_rate": 7.877601063757323e-05, |
| "loss": 0.6769, |
| "step": 130 |
| }, |
| { |
| "epoch": 0.858840401557058, |
| "grad_norm": 0.8010200105501427, |
| "learning_rate": 7.873050148337967e-05, |
| "loss": 0.6748, |
| "step": 131 |
| }, |
| { |
| "epoch": 0.8653964351567301, |
| "grad_norm": 0.9126529329815608, |
| "learning_rate": 7.868417529562043e-05, |
| "loss": 0.6632, |
| "step": 132 |
| }, |
| { |
| "epoch": 0.8719524687564024, |
| "grad_norm": 1.145222649075524, |
| "learning_rate": 7.863703305156273e-05, |
| "loss": 0.6756, |
| "step": 133 |
| }, |
| { |
| "epoch": 0.8785085023560746, |
| "grad_norm": 1.082669097422374, |
| "learning_rate": 7.858907574568882e-05, |
| "loss": 0.6765, |
| "step": 134 |
| }, |
| { |
| "epoch": 0.8850645359557467, |
| "grad_norm": 1.0310987601522028, |
| "learning_rate": 7.854030438967494e-05, |
| "loss": 0.6738, |
| "step": 135 |
| }, |
| { |
| "epoch": 0.891620569555419, |
| "grad_norm": 1.0160706638367318, |
| "learning_rate": 7.849072001237001e-05, |
| "loss": 0.6778, |
| "step": 136 |
| }, |
| { |
| "epoch": 0.8981766031550912, |
| "grad_norm": 0.9374366244899177, |
| "learning_rate": 7.844032365977396e-05, |
| "loss": 0.6736, |
| "step": 137 |
| }, |
| { |
| "epoch": 0.9047326367547633, |
| "grad_norm": 0.7008738978429749, |
| "learning_rate": 7.838911639501557e-05, |
| "loss": 0.6781, |
| "step": 138 |
| }, |
| { |
| "epoch": 0.9112886703544356, |
| "grad_norm": 0.4532137724870558, |
| "learning_rate": 7.833709929833012e-05, |
| "loss": 0.6686, |
| "step": 139 |
| }, |
| { |
| "epoch": 0.9178447039541078, |
| "grad_norm": 0.5139137753219752, |
| "learning_rate": 7.828427346703657e-05, |
| "loss": 0.6672, |
| "step": 140 |
| }, |
| { |
| "epoch": 0.9244007375537799, |
| "grad_norm": 0.47106346244337904, |
| "learning_rate": 7.823064001551445e-05, |
| "loss": 0.6621, |
| "step": 141 |
| }, |
| { |
| "epoch": 0.9309567711534522, |
| "grad_norm": 0.3844200151143341, |
| "learning_rate": 7.81762000751803e-05, |
| "loss": 0.6669, |
| "step": 142 |
| }, |
| { |
| "epoch": 0.9375128047531244, |
| "grad_norm": 0.4844251889310036, |
| "learning_rate": 7.812095479446383e-05, |
| "loss": 0.6606, |
| "step": 143 |
| }, |
| { |
| "epoch": 0.9440688383527965, |
| "grad_norm": 0.5614156709455322, |
| "learning_rate": 7.806490533878368e-05, |
| "loss": 0.6627, |
| "step": 144 |
| }, |
| { |
| "epoch": 0.9506248719524688, |
| "grad_norm": 0.6295196907491479, |
| "learning_rate": 7.800805289052286e-05, |
| "loss": 0.6652, |
| "step": 145 |
| }, |
| { |
| "epoch": 0.957180905552141, |
| "grad_norm": 0.7172605017331375, |
| "learning_rate": 7.795039864900378e-05, |
| "loss": 0.6632, |
| "step": 146 |
| }, |
| { |
| "epoch": 0.9637369391518131, |
| "grad_norm": 0.9441139561703984, |
| "learning_rate": 7.789194383046295e-05, |
| "loss": 0.6681, |
| "step": 147 |
| }, |
| { |
| "epoch": 0.9702929727514854, |
| "grad_norm": 1.3151350146737344, |
| "learning_rate": 7.783268966802539e-05, |
| "loss": 0.6703, |
| "step": 148 |
| }, |
| { |
| "epoch": 0.9768490063511576, |
| "grad_norm": 0.5115168470062825, |
| "learning_rate": 7.777263741167849e-05, |
| "loss": 0.6647, |
| "step": 149 |
| }, |
| { |
| "epoch": 0.9834050399508297, |
| "grad_norm": 1.0163356112728301, |
| "learning_rate": 7.771178832824573e-05, |
| "loss": 0.666, |
| "step": 150 |
| }, |
| { |
| "epoch": 0.9899610735505019, |
| "grad_norm": 1.1732972717592203, |
| "learning_rate": 7.765014370135999e-05, |
| "loss": 0.6592, |
| "step": 151 |
| }, |
| { |
| "epoch": 0.9965171071501742, |
| "grad_norm": 0.7851604401937097, |
| "learning_rate": 7.758770483143634e-05, |
| "loss": 0.6592, |
| "step": 152 |
| }, |
| { |
| "epoch": 1.0055316533497234, |
| "grad_norm": 1.6420611019668498, |
| "learning_rate": 7.752447303564475e-05, |
| "loss": 1.1949, |
| "step": 153 |
| }, |
| { |
| "epoch": 1.0120876869493955, |
| "grad_norm": 0.9808444668187579, |
| "learning_rate": 7.74604496478822e-05, |
| "loss": 0.6428, |
| "step": 154 |
| }, |
| { |
| "epoch": 1.0186437205490677, |
| "grad_norm": 1.1426869091876994, |
| "learning_rate": 7.73956360187446e-05, |
| "loss": 0.6346, |
| "step": 155 |
| }, |
| { |
| "epoch": 1.02519975414874, |
| "grad_norm": 1.147145143696543, |
| "learning_rate": 7.733003351549829e-05, |
| "loss": 0.6388, |
| "step": 156 |
| }, |
| { |
| "epoch": 1.0317557877484123, |
| "grad_norm": 0.7817012939106086, |
| "learning_rate": 7.726364352205117e-05, |
| "loss": 0.6408, |
| "step": 157 |
| }, |
| { |
| "epoch": 1.0383118213480844, |
| "grad_norm": 0.694581928412185, |
| "learning_rate": 7.719646743892352e-05, |
| "loss": 0.6284, |
| "step": 158 |
| }, |
| { |
| "epoch": 1.0448678549477566, |
| "grad_norm": 0.6070626955744831, |
| "learning_rate": 7.712850668321846e-05, |
| "loss": 0.6405, |
| "step": 159 |
| }, |
| { |
| "epoch": 1.0514238885474287, |
| "grad_norm": 0.6514623348225942, |
| "learning_rate": 7.705976268859207e-05, |
| "loss": 0.6395, |
| "step": 160 |
| }, |
| { |
| "epoch": 1.057979922147101, |
| "grad_norm": 0.5619592930813068, |
| "learning_rate": 7.699023690522315e-05, |
| "loss": 0.629, |
| "step": 161 |
| }, |
| { |
| "epoch": 1.0645359557467733, |
| "grad_norm": 0.5769129425178826, |
| "learning_rate": 7.691993079978252e-05, |
| "loss": 0.6381, |
| "step": 162 |
| }, |
| { |
| "epoch": 1.0710919893464454, |
| "grad_norm": 0.8198780194642444, |
| "learning_rate": 7.684884585540227e-05, |
| "loss": 0.6363, |
| "step": 163 |
| }, |
| { |
| "epoch": 1.0776480229461176, |
| "grad_norm": 0.7916477488592911, |
| "learning_rate": 7.677698357164431e-05, |
| "loss": 0.635, |
| "step": 164 |
| }, |
| { |
| "epoch": 1.0842040565457898, |
| "grad_norm": 0.5931606825910369, |
| "learning_rate": 7.670434546446886e-05, |
| "loss": 0.6333, |
| "step": 165 |
| }, |
| { |
| "epoch": 1.090760090145462, |
| "grad_norm": 0.5877139627697113, |
| "learning_rate": 7.663093306620231e-05, |
| "loss": 0.629, |
| "step": 166 |
| }, |
| { |
| "epoch": 1.097316123745134, |
| "grad_norm": 0.852370483441162, |
| "learning_rate": 7.655674792550507e-05, |
| "loss": 0.6395, |
| "step": 167 |
| }, |
| { |
| "epoch": 1.1038721573448065, |
| "grad_norm": 1.00327271618352, |
| "learning_rate": 7.648179160733883e-05, |
| "loss": 0.6478, |
| "step": 168 |
| }, |
| { |
| "epoch": 1.1104281909444786, |
| "grad_norm": 1.0959509941262375, |
| "learning_rate": 7.640606569293347e-05, |
| "loss": 0.6415, |
| "step": 169 |
| }, |
| { |
| "epoch": 1.1169842245441508, |
| "grad_norm": 0.7671153051839551, |
| "learning_rate": 7.632957177975387e-05, |
| "loss": 0.6401, |
| "step": 170 |
| }, |
| { |
| "epoch": 1.123540258143823, |
| "grad_norm": 0.7544679819781906, |
| "learning_rate": 7.625231148146601e-05, |
| "loss": 0.6365, |
| "step": 171 |
| }, |
| { |
| "epoch": 1.1300962917434951, |
| "grad_norm": 0.7948382138503196, |
| "learning_rate": 7.61742864279031e-05, |
| "loss": 0.6363, |
| "step": 172 |
| }, |
| { |
| "epoch": 1.1366523253431673, |
| "grad_norm": 0.7047533773597212, |
| "learning_rate": 7.609549826503115e-05, |
| "loss": 0.6421, |
| "step": 173 |
| }, |
| { |
| "epoch": 1.1432083589428397, |
| "grad_norm": 0.7374204351284568, |
| "learning_rate": 7.601594865491414e-05, |
| "loss": 0.6407, |
| "step": 174 |
| }, |
| { |
| "epoch": 1.1497643925425118, |
| "grad_norm": 0.6911928835033512, |
| "learning_rate": 7.593563927567916e-05, |
| "loss": 0.625, |
| "step": 175 |
| }, |
| { |
| "epoch": 1.156320426142184, |
| "grad_norm": 0.4803947112250499, |
| "learning_rate": 7.585457182148081e-05, |
| "loss": 0.628, |
| "step": 176 |
| }, |
| { |
| "epoch": 1.1628764597418562, |
| "grad_norm": 0.3950061404414093, |
| "learning_rate": 7.577274800246558e-05, |
| "loss": 0.6357, |
| "step": 177 |
| }, |
| { |
| "epoch": 1.1694324933415283, |
| "grad_norm": 0.4280389547638801, |
| "learning_rate": 7.569016954473577e-05, |
| "loss": 0.6434, |
| "step": 178 |
| }, |
| { |
| "epoch": 1.1759885269412005, |
| "grad_norm": 0.5103954479058104, |
| "learning_rate": 7.560683819031298e-05, |
| "loss": 0.6325, |
| "step": 179 |
| }, |
| { |
| "epoch": 1.1825445605408729, |
| "grad_norm": 0.48405163068225376, |
| "learning_rate": 7.552275569710152e-05, |
| "loss": 0.621, |
| "step": 180 |
| }, |
| { |
| "epoch": 1.189100594140545, |
| "grad_norm": 0.49505546120752403, |
| "learning_rate": 7.543792383885113e-05, |
| "loss": 0.6335, |
| "step": 181 |
| }, |
| { |
| "epoch": 1.1956566277402172, |
| "grad_norm": 0.5549014381017361, |
| "learning_rate": 7.535234440511979e-05, |
| "loss": 0.6401, |
| "step": 182 |
| }, |
| { |
| "epoch": 1.2022126613398894, |
| "grad_norm": 0.5679718866243005, |
| "learning_rate": 7.526601920123574e-05, |
| "loss": 0.6339, |
| "step": 183 |
| }, |
| { |
| "epoch": 1.2087686949395615, |
| "grad_norm": 0.5837881333403254, |
| "learning_rate": 7.517895004825956e-05, |
| "loss": 0.6201, |
| "step": 184 |
| }, |
| { |
| "epoch": 1.2153247285392337, |
| "grad_norm": 0.5275232093842687, |
| "learning_rate": 7.509113878294572e-05, |
| "loss": 0.6313, |
| "step": 185 |
| }, |
| { |
| "epoch": 1.2218807621389058, |
| "grad_norm": 0.4187731695243854, |
| "learning_rate": 7.500258725770375e-05, |
| "loss": 0.6297, |
| "step": 186 |
| }, |
| { |
| "epoch": 1.2284367957385782, |
| "grad_norm": 0.3496822319709094, |
| "learning_rate": 7.491329734055926e-05, |
| "loss": 0.6341, |
| "step": 187 |
| }, |
| { |
| "epoch": 1.2349928293382504, |
| "grad_norm": 0.2892298691529975, |
| "learning_rate": 7.48232709151145e-05, |
| "loss": 0.6289, |
| "step": 188 |
| }, |
| { |
| "epoch": 1.2415488629379225, |
| "grad_norm": 0.3970392684873568, |
| "learning_rate": 7.473250988050861e-05, |
| "loss": 0.6327, |
| "step": 189 |
| }, |
| { |
| "epoch": 1.2481048965375947, |
| "grad_norm": 0.5366046878238813, |
| "learning_rate": 7.464101615137756e-05, |
| "loss": 0.6234, |
| "step": 190 |
| }, |
| { |
| "epoch": 1.2546609301372669, |
| "grad_norm": 0.6365895325361283, |
| "learning_rate": 7.454879165781379e-05, |
| "loss": 0.6322, |
| "step": 191 |
| }, |
| { |
| "epoch": 1.2612169637369393, |
| "grad_norm": 0.6390242865748291, |
| "learning_rate": 7.445583834532546e-05, |
| "loss": 0.6252, |
| "step": 192 |
| }, |
| { |
| "epoch": 1.2677729973366114, |
| "grad_norm": 0.7130289120451936, |
| "learning_rate": 7.436215817479541e-05, |
| "loss": 0.6369, |
| "step": 193 |
| }, |
| { |
| "epoch": 1.2743290309362836, |
| "grad_norm": 0.8001337986119181, |
| "learning_rate": 7.426775312243986e-05, |
| "loss": 0.6341, |
| "step": 194 |
| }, |
| { |
| "epoch": 1.2808850645359557, |
| "grad_norm": 0.9421645194421999, |
| "learning_rate": 7.41726251797666e-05, |
| "loss": 0.629, |
| "step": 195 |
| }, |
| { |
| "epoch": 1.287441098135628, |
| "grad_norm": 1.0651209082990578, |
| "learning_rate": 7.407677635353308e-05, |
| "loss": 0.6274, |
| "step": 196 |
| }, |
| { |
| "epoch": 1.2939971317353, |
| "grad_norm": 0.8478545298742448, |
| "learning_rate": 7.398020866570407e-05, |
| "loss": 0.6373, |
| "step": 197 |
| }, |
| { |
| "epoch": 1.3005531653349722, |
| "grad_norm": 0.5323333743534195, |
| "learning_rate": 7.388292415340888e-05, |
| "loss": 0.631, |
| "step": 198 |
| }, |
| { |
| "epoch": 1.3071091989346446, |
| "grad_norm": 0.37990592076356283, |
| "learning_rate": 7.37849248688986e-05, |
| "loss": 0.631, |
| "step": 199 |
| }, |
| { |
| "epoch": 1.3136652325343168, |
| "grad_norm": 0.4732271711472085, |
| "learning_rate": 7.368621287950264e-05, |
| "loss": 0.6328, |
| "step": 200 |
| }, |
| { |
| "epoch": 1.320221266133989, |
| "grad_norm": 0.5259491499827896, |
| "learning_rate": 7.358679026758515e-05, |
| "loss": 0.6283, |
| "step": 201 |
| }, |
| { |
| "epoch": 1.326777299733661, |
| "grad_norm": 0.5000279931190099, |
| "learning_rate": 7.348665913050115e-05, |
| "loss": 0.6208, |
| "step": 202 |
| }, |
| { |
| "epoch": 1.3333333333333333, |
| "grad_norm": 0.3912490196632365, |
| "learning_rate": 7.338582158055224e-05, |
| "loss": 0.6251, |
| "step": 203 |
| }, |
| { |
| "epoch": 1.3398893669330056, |
| "grad_norm": 0.38905604815288286, |
| "learning_rate": 7.328427974494201e-05, |
| "loss": 0.6179, |
| "step": 204 |
| }, |
| { |
| "epoch": 1.3464454005326778, |
| "grad_norm": 0.4906878706390755, |
| "learning_rate": 7.318203576573126e-05, |
| "loss": 0.6271, |
| "step": 205 |
| }, |
| { |
| "epoch": 1.35300143413235, |
| "grad_norm": 0.4146373419913017, |
| "learning_rate": 7.307909179979274e-05, |
| "loss": 0.6256, |
| "step": 206 |
| }, |
| { |
| "epoch": 1.3595574677320221, |
| "grad_norm": 0.29246282870192875, |
| "learning_rate": 7.297545001876563e-05, |
| "loss": 0.6219, |
| "step": 207 |
| }, |
| { |
| "epoch": 1.3661135013316943, |
| "grad_norm": 0.27459922699544653, |
| "learning_rate": 7.28711126090098e-05, |
| "loss": 0.6346, |
| "step": 208 |
| }, |
| { |
| "epoch": 1.3726695349313665, |
| "grad_norm": 0.3594793278431249, |
| "learning_rate": 7.276608177155968e-05, |
| "loss": 0.6234, |
| "step": 209 |
| }, |
| { |
| "epoch": 1.3792255685310386, |
| "grad_norm": 0.46199653506152194, |
| "learning_rate": 7.266035972207773e-05, |
| "loss": 0.6328, |
| "step": 210 |
| }, |
| { |
| "epoch": 1.385781602130711, |
| "grad_norm": 0.5084391547285126, |
| "learning_rate": 7.25539486908078e-05, |
| "loss": 0.6287, |
| "step": 211 |
| }, |
| { |
| "epoch": 1.3923376357303832, |
| "grad_norm": 0.4970090588981773, |
| "learning_rate": 7.24468509225281e-05, |
| "loss": 0.6338, |
| "step": 212 |
| }, |
| { |
| "epoch": 1.3988936693300553, |
| "grad_norm": 0.4297079110960391, |
| "learning_rate": 7.233906867650373e-05, |
| "loss": 0.6246, |
| "step": 213 |
| }, |
| { |
| "epoch": 1.4054497029297275, |
| "grad_norm": 0.4277415770961674, |
| "learning_rate": 7.223060422643914e-05, |
| "loss": 0.6235, |
| "step": 214 |
| }, |
| { |
| "epoch": 1.4120057365293996, |
| "grad_norm": 0.5211718034258063, |
| "learning_rate": 7.212145986043007e-05, |
| "loss": 0.626, |
| "step": 215 |
| }, |
| { |
| "epoch": 1.418561770129072, |
| "grad_norm": 0.7113602181910502, |
| "learning_rate": 7.201163788091536e-05, |
| "loss": 0.626, |
| "step": 216 |
| }, |
| { |
| "epoch": 1.4251178037287442, |
| "grad_norm": 0.8702986528721038, |
| "learning_rate": 7.190114060462837e-05, |
| "loss": 0.6285, |
| "step": 217 |
| }, |
| { |
| "epoch": 1.4316738373284164, |
| "grad_norm": 0.88781618603766, |
| "learning_rate": 7.178997036254799e-05, |
| "loss": 0.625, |
| "step": 218 |
| }, |
| { |
| "epoch": 1.4382298709280885, |
| "grad_norm": 0.8543987595137604, |
| "learning_rate": 7.167812949984966e-05, |
| "loss": 0.6369, |
| "step": 219 |
| }, |
| { |
| "epoch": 1.4447859045277607, |
| "grad_norm": 0.6591694598278867, |
| "learning_rate": 7.156562037585576e-05, |
| "loss": 0.6309, |
| "step": 220 |
| }, |
| { |
| "epoch": 1.4513419381274328, |
| "grad_norm": 0.435460183285974, |
| "learning_rate": 7.145244536398584e-05, |
| "loss": 0.6337, |
| "step": 221 |
| }, |
| { |
| "epoch": 1.457897971727105, |
| "grad_norm": 0.5394064377655016, |
| "learning_rate": 7.133860685170665e-05, |
| "loss": 0.6272, |
| "step": 222 |
| }, |
| { |
| "epoch": 1.4644540053267772, |
| "grad_norm": 0.6560512889140389, |
| "learning_rate": 7.12241072404817e-05, |
| "loss": 0.6317, |
| "step": 223 |
| }, |
| { |
| "epoch": 1.4710100389264495, |
| "grad_norm": 0.5108872994150431, |
| "learning_rate": 7.110894894572056e-05, |
| "loss": 0.6266, |
| "step": 224 |
| }, |
| { |
| "epoch": 1.4775660725261217, |
| "grad_norm": 0.4508746314491801, |
| "learning_rate": 7.099313439672806e-05, |
| "loss": 0.6222, |
| "step": 225 |
| }, |
| { |
| "epoch": 1.4841221061257939, |
| "grad_norm": 0.5932762714408137, |
| "learning_rate": 7.087666603665284e-05, |
| "loss": 0.6174, |
| "step": 226 |
| }, |
| { |
| "epoch": 1.490678139725466, |
| "grad_norm": 0.5703671444864514, |
| "learning_rate": 7.0759546322436e-05, |
| "loss": 0.6286, |
| "step": 227 |
| }, |
| { |
| "epoch": 1.4972341733251384, |
| "grad_norm": 0.4219759927229514, |
| "learning_rate": 7.064177772475912e-05, |
| "loss": 0.621, |
| "step": 228 |
| }, |
| { |
| "epoch": 1.5037902069248106, |
| "grad_norm": 0.3899488374409842, |
| "learning_rate": 7.052336272799227e-05, |
| "loss": 0.6368, |
| "step": 229 |
| }, |
| { |
| "epoch": 1.5103462405244827, |
| "grad_norm": 0.3352814743100072, |
| "learning_rate": 7.040430383014146e-05, |
| "loss": 0.6273, |
| "step": 230 |
| }, |
| { |
| "epoch": 1.516902274124155, |
| "grad_norm": 0.49330992250119887, |
| "learning_rate": 7.02846035427961e-05, |
| "loss": 0.6268, |
| "step": 231 |
| }, |
| { |
| "epoch": 1.523458307723827, |
| "grad_norm": 0.7275390777504656, |
| "learning_rate": 7.016426439107586e-05, |
| "loss": 0.6198, |
| "step": 232 |
| }, |
| { |
| "epoch": 1.5300143413234992, |
| "grad_norm": 0.5944986948743242, |
| "learning_rate": 7.004328891357753e-05, |
| "loss": 0.6321, |
| "step": 233 |
| }, |
| { |
| "epoch": 1.5365703749231714, |
| "grad_norm": 0.40136071416301794, |
| "learning_rate": 6.992167966232143e-05, |
| "loss": 0.6205, |
| "step": 234 |
| }, |
| { |
| "epoch": 1.5431264085228436, |
| "grad_norm": 0.42683320761735183, |
| "learning_rate": 6.979943920269749e-05, |
| "loss": 0.6282, |
| "step": 235 |
| }, |
| { |
| "epoch": 1.5496824421225157, |
| "grad_norm": 0.37233786072575653, |
| "learning_rate": 6.967657011341126e-05, |
| "loss": 0.6216, |
| "step": 236 |
| }, |
| { |
| "epoch": 1.556238475722188, |
| "grad_norm": 0.367274159532209, |
| "learning_rate": 6.955307498642948e-05, |
| "loss": 0.6224, |
| "step": 237 |
| }, |
| { |
| "epoch": 1.5627945093218603, |
| "grad_norm": 0.3660447730807658, |
| "learning_rate": 6.942895642692527e-05, |
| "loss": 0.6202, |
| "step": 238 |
| }, |
| { |
| "epoch": 1.5693505429215324, |
| "grad_norm": 0.38882305238122433, |
| "learning_rate": 6.930421705322339e-05, |
| "loss": 0.6195, |
| "step": 239 |
| }, |
| { |
| "epoch": 1.5759065765212048, |
| "grad_norm": 0.5684648344288802, |
| "learning_rate": 6.917885949674483e-05, |
| "loss": 0.6228, |
| "step": 240 |
| }, |
| { |
| "epoch": 1.582462610120877, |
| "grad_norm": 0.7431328174808526, |
| "learning_rate": 6.905288640195141e-05, |
| "loss": 0.626, |
| "step": 241 |
| }, |
| { |
| "epoch": 1.5890186437205491, |
| "grad_norm": 0.7410518651663662, |
| "learning_rate": 6.892630042628988e-05, |
| "loss": 0.6248, |
| "step": 242 |
| }, |
| { |
| "epoch": 1.5955746773202213, |
| "grad_norm": 0.6219837555078648, |
| "learning_rate": 6.879910424013599e-05, |
| "loss": 0.6295, |
| "step": 243 |
| }, |
| { |
| "epoch": 1.6021307109198935, |
| "grad_norm": 0.5396699034196659, |
| "learning_rate": 6.867130052673806e-05, |
| "loss": 0.6231, |
| "step": 244 |
| }, |
| { |
| "epoch": 1.6086867445195656, |
| "grad_norm": 0.4531223244262461, |
| "learning_rate": 6.854289198216042e-05, |
| "loss": 0.6346, |
| "step": 245 |
| }, |
| { |
| "epoch": 1.6152427781192378, |
| "grad_norm": 0.3242170520225258, |
| "learning_rate": 6.841388131522656e-05, |
| "loss": 0.6292, |
| "step": 246 |
| }, |
| { |
| "epoch": 1.62179881171891, |
| "grad_norm": 0.34710026048248854, |
| "learning_rate": 6.828427124746191e-05, |
| "loss": 0.6153, |
| "step": 247 |
| }, |
| { |
| "epoch": 1.628354845318582, |
| "grad_norm": 0.44267957645174644, |
| "learning_rate": 6.815406451303647e-05, |
| "loss": 0.6205, |
| "step": 248 |
| }, |
| { |
| "epoch": 1.6349108789182545, |
| "grad_norm": 0.5545408679102592, |
| "learning_rate": 6.802326385870715e-05, |
| "loss": 0.621, |
| "step": 249 |
| }, |
| { |
| "epoch": 1.6414669125179266, |
| "grad_norm": 0.5443247431153629, |
| "learning_rate": 6.789187204375981e-05, |
| "loss": 0.622, |
| "step": 250 |
| }, |
| { |
| "epoch": 1.6480229461175988, |
| "grad_norm": 0.4301197066284, |
| "learning_rate": 6.775989183995108e-05, |
| "loss": 0.6135, |
| "step": 251 |
| }, |
| { |
| "epoch": 1.6545789797172712, |
| "grad_norm": 0.396698152583572, |
| "learning_rate": 6.762732603144978e-05, |
| "loss": 0.6216, |
| "step": 252 |
| }, |
| { |
| "epoch": 1.6611350133169434, |
| "grad_norm": 0.3515261284616528, |
| "learning_rate": 6.749417741477836e-05, |
| "loss": 0.6187, |
| "step": 253 |
| }, |
| { |
| "epoch": 1.6676910469166155, |
| "grad_norm": 0.35858581397311556, |
| "learning_rate": 6.736044879875373e-05, |
| "loss": 0.6185, |
| "step": 254 |
| }, |
| { |
| "epoch": 1.6742470805162877, |
| "grad_norm": 0.34392278311565416, |
| "learning_rate": 6.722614300442815e-05, |
| "loss": 0.6154, |
| "step": 255 |
| }, |
| { |
| "epoch": 1.6808031141159598, |
| "grad_norm": 0.36707892916862706, |
| "learning_rate": 6.709126286502965e-05, |
| "loss": 0.6179, |
| "step": 256 |
| }, |
| { |
| "epoch": 1.687359147715632, |
| "grad_norm": 0.441219103876971, |
| "learning_rate": 6.695581122590225e-05, |
| "loss": 0.6227, |
| "step": 257 |
| }, |
| { |
| "epoch": 1.6939151813153042, |
| "grad_norm": 0.3994835806954666, |
| "learning_rate": 6.681979094444596e-05, |
| "loss": 0.6192, |
| "step": 258 |
| }, |
| { |
| "epoch": 1.7004712149149763, |
| "grad_norm": 0.433716188444932, |
| "learning_rate": 6.668320489005654e-05, |
| "loss": 0.622, |
| "step": 259 |
| }, |
| { |
| "epoch": 1.7070272485146485, |
| "grad_norm": 0.47283003533467305, |
| "learning_rate": 6.654605594406486e-05, |
| "loss": 0.6253, |
| "step": 260 |
| }, |
| { |
| "epoch": 1.7135832821143209, |
| "grad_norm": 0.49013239605830206, |
| "learning_rate": 6.640834699967626e-05, |
| "loss": 0.6214, |
| "step": 261 |
| }, |
| { |
| "epoch": 1.720139315713993, |
| "grad_norm": 0.49324672918814405, |
| "learning_rate": 6.627008096190938e-05, |
| "loss": 0.6182, |
| "step": 262 |
| }, |
| { |
| "epoch": 1.7266953493136652, |
| "grad_norm": 0.5169572714827575, |
| "learning_rate": 6.6131260747535e-05, |
| "loss": 0.6259, |
| "step": 263 |
| }, |
| { |
| "epoch": 1.7332513829133376, |
| "grad_norm": 0.5873107078500235, |
| "learning_rate": 6.59918892850144e-05, |
| "loss": 0.6235, |
| "step": 264 |
| }, |
| { |
| "epoch": 1.7398074165130097, |
| "grad_norm": 0.5020487886809576, |
| "learning_rate": 6.585196951443763e-05, |
| "loss": 0.6221, |
| "step": 265 |
| }, |
| { |
| "epoch": 1.746363450112682, |
| "grad_norm": 0.329791614699069, |
| "learning_rate": 6.571150438746157e-05, |
| "loss": 0.6265, |
| "step": 266 |
| }, |
| { |
| "epoch": 1.752919483712354, |
| "grad_norm": 0.2861139606661766, |
| "learning_rate": 6.557049686724751e-05, |
| "loss": 0.624, |
| "step": 267 |
| }, |
| { |
| "epoch": 1.7594755173120262, |
| "grad_norm": 0.46750896287166216, |
| "learning_rate": 6.542894992839873e-05, |
| "loss": 0.6241, |
| "step": 268 |
| }, |
| { |
| "epoch": 1.7660315509116984, |
| "grad_norm": 0.5684934196993104, |
| "learning_rate": 6.528686655689774e-05, |
| "loss": 0.6152, |
| "step": 269 |
| }, |
| { |
| "epoch": 1.7725875845113706, |
| "grad_norm": 0.48490900499130857, |
| "learning_rate": 6.514424975004329e-05, |
| "loss": 0.6111, |
| "step": 270 |
| }, |
| { |
| "epoch": 1.7791436181110427, |
| "grad_norm": 0.5506105973703118, |
| "learning_rate": 6.500110251638715e-05, |
| "loss": 0.6152, |
| "step": 271 |
| }, |
| { |
| "epoch": 1.7856996517107149, |
| "grad_norm": 0.4539294761333573, |
| "learning_rate": 6.48574278756706e-05, |
| "loss": 0.6273, |
| "step": 272 |
| }, |
| { |
| "epoch": 1.7922556853103873, |
| "grad_norm": 0.34705648209401146, |
| "learning_rate": 6.471322885876077e-05, |
| "loss": 0.6204, |
| "step": 273 |
| }, |
| { |
| "epoch": 1.7988117189100594, |
| "grad_norm": 0.2427469168923526, |
| "learning_rate": 6.456850850758673e-05, |
| "loss": 0.6214, |
| "step": 274 |
| }, |
| { |
| "epoch": 1.8053677525097316, |
| "grad_norm": 0.3004384124304078, |
| "learning_rate": 6.44232698750752e-05, |
| "loss": 0.6135, |
| "step": 275 |
| }, |
| { |
| "epoch": 1.811923786109404, |
| "grad_norm": 0.29782974812920376, |
| "learning_rate": 6.427751602508628e-05, |
| "loss": 0.6104, |
| "step": 276 |
| }, |
| { |
| "epoch": 1.8184798197090761, |
| "grad_norm": 0.29267694689244445, |
| "learning_rate": 6.413125003234876e-05, |
| "loss": 0.624, |
| "step": 277 |
| }, |
| { |
| "epoch": 1.8250358533087483, |
| "grad_norm": 0.3462332305273243, |
| "learning_rate": 6.398447498239527e-05, |
| "loss": 0.612, |
| "step": 278 |
| }, |
| { |
| "epoch": 1.8315918869084205, |
| "grad_norm": 0.3949978565357491, |
| "learning_rate": 6.383719397149715e-05, |
| "loss": 0.622, |
| "step": 279 |
| }, |
| { |
| "epoch": 1.8381479205080926, |
| "grad_norm": 0.40536973102729335, |
| "learning_rate": 6.368941010659921e-05, |
| "loss": 0.6118, |
| "step": 280 |
| }, |
| { |
| "epoch": 1.8447039541077648, |
| "grad_norm": 0.3621578404050215, |
| "learning_rate": 6.354112650525407e-05, |
| "loss": 0.6149, |
| "step": 281 |
| }, |
| { |
| "epoch": 1.851259987707437, |
| "grad_norm": 0.3445107311397209, |
| "learning_rate": 6.339234629555655e-05, |
| "loss": 0.6196, |
| "step": 282 |
| }, |
| { |
| "epoch": 1.857816021307109, |
| "grad_norm": 0.30605392165712425, |
| "learning_rate": 6.324307261607754e-05, |
| "loss": 0.6239, |
| "step": 283 |
| }, |
| { |
| "epoch": 1.8643720549067813, |
| "grad_norm": 0.3080701040643038, |
| "learning_rate": 6.309330861579786e-05, |
| "loss": 0.629, |
| "step": 284 |
| }, |
| { |
| "epoch": 1.8709280885064536, |
| "grad_norm": 0.3349937492345958, |
| "learning_rate": 6.294305745404185e-05, |
| "loss": 0.6224, |
| "step": 285 |
| }, |
| { |
| "epoch": 1.8774841221061258, |
| "grad_norm": 0.3396848816828842, |
| "learning_rate": 6.279232230041065e-05, |
| "loss": 0.6182, |
| "step": 286 |
| }, |
| { |
| "epoch": 1.884040155705798, |
| "grad_norm": 0.3410093342010557, |
| "learning_rate": 6.26411063347154e-05, |
| "loss": 0.6188, |
| "step": 287 |
| }, |
| { |
| "epoch": 1.8905961893054704, |
| "grad_norm": 0.314163527323835, |
| "learning_rate": 6.248941274691017e-05, |
| "loss": 0.6169, |
| "step": 288 |
| }, |
| { |
| "epoch": 1.8971522229051425, |
| "grad_norm": 0.3976946162007582, |
| "learning_rate": 6.233724473702457e-05, |
| "loss": 0.6195, |
| "step": 289 |
| }, |
| { |
| "epoch": 1.9037082565048147, |
| "grad_norm": 0.5601353024746372, |
| "learning_rate": 6.218460551509636e-05, |
| "loss": 0.6206, |
| "step": 290 |
| }, |
| { |
| "epoch": 1.9102642901044868, |
| "grad_norm": 0.6132983016524085, |
| "learning_rate": 6.203149830110367e-05, |
| "loss": 0.6138, |
| "step": 291 |
| }, |
| { |
| "epoch": 1.916820323704159, |
| "grad_norm": 0.5021407478552538, |
| "learning_rate": 6.18779263248971e-05, |
| "loss": 0.6219, |
| "step": 292 |
| }, |
| { |
| "epoch": 1.9233763573038312, |
| "grad_norm": 0.3890448472238098, |
| "learning_rate": 6.172389282613151e-05, |
| "loss": 0.6187, |
| "step": 293 |
| }, |
| { |
| "epoch": 1.9299323909035033, |
| "grad_norm": 0.3361740572901459, |
| "learning_rate": 6.156940105419785e-05, |
| "loss": 0.6218, |
| "step": 294 |
| }, |
| { |
| "epoch": 1.9364884245031755, |
| "grad_norm": 0.2907876046977178, |
| "learning_rate": 6.141445426815443e-05, |
| "loss": 0.6166, |
| "step": 295 |
| }, |
| { |
| "epoch": 1.9430444581028476, |
| "grad_norm": 0.2632987500846716, |
| "learning_rate": 6.125905573665824e-05, |
| "loss": 0.6232, |
| "step": 296 |
| }, |
| { |
| "epoch": 1.9496004917025198, |
| "grad_norm": 0.32120831457165155, |
| "learning_rate": 6.110320873789604e-05, |
| "loss": 0.6242, |
| "step": 297 |
| }, |
| { |
| "epoch": 1.9561565253021922, |
| "grad_norm": 0.32170026271883145, |
| "learning_rate": 6.094691655951512e-05, |
| "loss": 0.6094, |
| "step": 298 |
| }, |
| { |
| "epoch": 1.9627125589018644, |
| "grad_norm": 0.27126547953075625, |
| "learning_rate": 6.079018249855402e-05, |
| "loss": 0.622, |
| "step": 299 |
| }, |
| { |
| "epoch": 1.9692685925015365, |
| "grad_norm": 0.25823866111956445, |
| "learning_rate": 6.063300986137297e-05, |
| "loss": 0.6195, |
| "step": 300 |
| }, |
| { |
| "epoch": 1.975824626101209, |
| "grad_norm": 0.2777978831118673, |
| "learning_rate": 6.047540196358405e-05, |
| "loss": 0.6188, |
| "step": 301 |
| }, |
| { |
| "epoch": 1.982380659700881, |
| "grad_norm": 0.3097245776135163, |
| "learning_rate": 6.0317362129981375e-05, |
| "loss": 0.6161, |
| "step": 302 |
| }, |
| { |
| "epoch": 1.9889366933005532, |
| "grad_norm": 0.3108082557836966, |
| "learning_rate": 6.015889369447088e-05, |
| "loss": 0.6166, |
| "step": 303 |
| }, |
| { |
| "epoch": 1.9954927269002254, |
| "grad_norm": 0.2659644339645489, |
| "learning_rate": 6.000000000000001e-05, |
| "loss": 0.6143, |
| "step": 304 |
| }, |
| { |
| "epoch": 2.006556033599672, |
| "grad_norm": 0.3001511379848704, |
| "learning_rate": 5.9840684398487186e-05, |
| "loss": 0.5867, |
| "step": 305 |
| }, |
| { |
| "epoch": 2.0131120671993443, |
| "grad_norm": 0.3404835526439102, |
| "learning_rate": 5.968095025075114e-05, |
| "loss": 0.5745, |
| "step": 306 |
| }, |
| { |
| "epoch": 2.0196681007990165, |
| "grad_norm": 0.5738437895797627, |
| "learning_rate": 5.952080092643993e-05, |
| "loss": 0.5828, |
| "step": 307 |
| }, |
| { |
| "epoch": 2.0262241343986886, |
| "grad_norm": 0.7841436327218428, |
| "learning_rate": 5.936023980395997e-05, |
| "loss": 0.583, |
| "step": 308 |
| }, |
| { |
| "epoch": 2.032780167998361, |
| "grad_norm": 0.7382363346356913, |
| "learning_rate": 5.919927027040463e-05, |
| "loss": 0.5795, |
| "step": 309 |
| }, |
| { |
| "epoch": 2.039336201598033, |
| "grad_norm": 0.6545848867535631, |
| "learning_rate": 5.903789572148295e-05, |
| "loss": 0.5835, |
| "step": 310 |
| }, |
| { |
| "epoch": 2.0458922351977056, |
| "grad_norm": 0.516762214742333, |
| "learning_rate": 5.887611956144782e-05, |
| "loss": 0.5787, |
| "step": 311 |
| }, |
| { |
| "epoch": 2.0524482687973777, |
| "grad_norm": 0.5327353778668117, |
| "learning_rate": 5.871394520302432e-05, |
| "loss": 0.5742, |
| "step": 312 |
| }, |
| { |
| "epoch": 2.05900430239705, |
| "grad_norm": 0.5080742962254841, |
| "learning_rate": 5.8551376067337626e-05, |
| "loss": 0.5737, |
| "step": 313 |
| }, |
| { |
| "epoch": 2.065560335996722, |
| "grad_norm": 0.5649012551531187, |
| "learning_rate": 5.838841558384091e-05, |
| "loss": 0.5764, |
| "step": 314 |
| }, |
| { |
| "epoch": 2.0721163695963942, |
| "grad_norm": 0.5820350754778492, |
| "learning_rate": 5.8225067190242925e-05, |
| "loss": 0.5716, |
| "step": 315 |
| }, |
| { |
| "epoch": 2.0786724031960664, |
| "grad_norm": 0.46076842055118317, |
| "learning_rate": 5.806133433243558e-05, |
| "loss": 0.5753, |
| "step": 316 |
| }, |
| { |
| "epoch": 2.0852284367957385, |
| "grad_norm": 0.47933858405340773, |
| "learning_rate": 5.789722046442114e-05, |
| "loss": 0.575, |
| "step": 317 |
| }, |
| { |
| "epoch": 2.0917844703954107, |
| "grad_norm": 0.5168151561479127, |
| "learning_rate": 5.7732729048239444e-05, |
| "loss": 0.5749, |
| "step": 318 |
| }, |
| { |
| "epoch": 2.098340503995083, |
| "grad_norm": 0.40840899989073715, |
| "learning_rate": 5.756786355389482e-05, |
| "loss": 0.5802, |
| "step": 319 |
| }, |
| { |
| "epoch": 2.104896537594755, |
| "grad_norm": 0.44997063706673934, |
| "learning_rate": 5.740262745928293e-05, |
| "loss": 0.5717, |
| "step": 320 |
| }, |
| { |
| "epoch": 2.111452571194427, |
| "grad_norm": 0.5369629713906064, |
| "learning_rate": 5.723702425011738e-05, |
| "loss": 0.5751, |
| "step": 321 |
| }, |
| { |
| "epoch": 2.1180086047940994, |
| "grad_norm": 0.4475818760429054, |
| "learning_rate": 5.707105741985615e-05, |
| "loss": 0.5765, |
| "step": 322 |
| }, |
| { |
| "epoch": 2.124564638393772, |
| "grad_norm": 0.30932774384801376, |
| "learning_rate": 5.6904730469627985e-05, |
| "loss": 0.5785, |
| "step": 323 |
| }, |
| { |
| "epoch": 2.131120671993444, |
| "grad_norm": 0.33040348186144713, |
| "learning_rate": 5.673804690815845e-05, |
| "loss": 0.5768, |
| "step": 324 |
| }, |
| { |
| "epoch": 2.1376767055931163, |
| "grad_norm": 0.3576099667078179, |
| "learning_rate": 5.6571010251695954e-05, |
| "loss": 0.5815, |
| "step": 325 |
| }, |
| { |
| "epoch": 2.1442327391927885, |
| "grad_norm": 0.4237003547581017, |
| "learning_rate": 5.6403624023937614e-05, |
| "loss": 0.5747, |
| "step": 326 |
| }, |
| { |
| "epoch": 2.1507887727924606, |
| "grad_norm": 0.4700100373051473, |
| "learning_rate": 5.62358917559548e-05, |
| "loss": 0.5751, |
| "step": 327 |
| }, |
| { |
| "epoch": 2.1573448063921328, |
| "grad_norm": 0.39617955284785017, |
| "learning_rate": 5.606781698611879e-05, |
| "loss": 0.5798, |
| "step": 328 |
| }, |
| { |
| "epoch": 2.163900839991805, |
| "grad_norm": 0.28548306735776896, |
| "learning_rate": 5.5899403260026006e-05, |
| "loss": 0.5724, |
| "step": 329 |
| }, |
| { |
| "epoch": 2.170456873591477, |
| "grad_norm": 0.3315286342160672, |
| "learning_rate": 5.573065413042333e-05, |
| "loss": 0.5721, |
| "step": 330 |
| }, |
| { |
| "epoch": 2.1770129071911493, |
| "grad_norm": 0.3343441243180511, |
| "learning_rate": 5.556157315713305e-05, |
| "loss": 0.5783, |
| "step": 331 |
| }, |
| { |
| "epoch": 2.1835689407908214, |
| "grad_norm": 0.29129418213987884, |
| "learning_rate": 5.5392163906977835e-05, |
| "loss": 0.577, |
| "step": 332 |
| }, |
| { |
| "epoch": 2.1901249743904936, |
| "grad_norm": 0.25286013518278194, |
| "learning_rate": 5.522242995370545e-05, |
| "loss": 0.5698, |
| "step": 333 |
| }, |
| { |
| "epoch": 2.1966810079901657, |
| "grad_norm": 0.5001313540973934, |
| "learning_rate": 5.505237487791343e-05, |
| "loss": 0.5915, |
| "step": 334 |
| }, |
| { |
| "epoch": 2.2032370415898384, |
| "grad_norm": 0.2322461169124981, |
| "learning_rate": 5.488200226697345e-05, |
| "loss": 0.5707, |
| "step": 335 |
| }, |
| { |
| "epoch": 2.2097930751895105, |
| "grad_norm": 0.24729941417632514, |
| "learning_rate": 5.471131571495574e-05, |
| "loss": 0.5688, |
| "step": 336 |
| }, |
| { |
| "epoch": 2.2163491087891827, |
| "grad_norm": 0.25095902176236934, |
| "learning_rate": 5.454031882255319e-05, |
| "loss": 0.5804, |
| "step": 337 |
| }, |
| { |
| "epoch": 2.222905142388855, |
| "grad_norm": 0.26393993849887953, |
| "learning_rate": 5.4369015197005506e-05, |
| "loss": 0.5741, |
| "step": 338 |
| }, |
| { |
| "epoch": 2.229461175988527, |
| "grad_norm": 0.32581537912637687, |
| "learning_rate": 5.419740845202292e-05, |
| "loss": 0.5841, |
| "step": 339 |
| }, |
| { |
| "epoch": 2.236017209588199, |
| "grad_norm": 0.2928430349319568, |
| "learning_rate": 5.4025502207710184e-05, |
| "loss": 0.5763, |
| "step": 340 |
| }, |
| { |
| "epoch": 2.2425732431878713, |
| "grad_norm": 0.28344192622339104, |
| "learning_rate": 5.385330009049003e-05, |
| "loss": 0.5748, |
| "step": 341 |
| }, |
| { |
| "epoch": 2.2491292767875435, |
| "grad_norm": 0.2108350778995995, |
| "learning_rate": 5.368080573302676e-05, |
| "loss": 0.5677, |
| "step": 342 |
| }, |
| { |
| "epoch": 2.2556853103872156, |
| "grad_norm": 0.2654969873255356, |
| "learning_rate": 5.3508022774149574e-05, |
| "loss": 0.5759, |
| "step": 343 |
| }, |
| { |
| "epoch": 2.262241343986888, |
| "grad_norm": 0.26893594961047856, |
| "learning_rate": 5.333495485877583e-05, |
| "loss": 0.5713, |
| "step": 344 |
| }, |
| { |
| "epoch": 2.26879737758656, |
| "grad_norm": 0.2359706290249212, |
| "learning_rate": 5.3161605637834135e-05, |
| "loss": 0.5826, |
| "step": 345 |
| }, |
| { |
| "epoch": 2.2753534111862326, |
| "grad_norm": 0.2001826448668912, |
| "learning_rate": 5.298797876818735e-05, |
| "loss": 0.5828, |
| "step": 346 |
| }, |
| { |
| "epoch": 2.2819094447859047, |
| "grad_norm": 0.18687800048623612, |
| "learning_rate": 5.2814077912555415e-05, |
| "loss": 0.5674, |
| "step": 347 |
| }, |
| { |
| "epoch": 2.288465478385577, |
| "grad_norm": 0.18869821797720956, |
| "learning_rate": 5.263990673943811e-05, |
| "loss": 0.5795, |
| "step": 348 |
| }, |
| { |
| "epoch": 2.295021511985249, |
| "grad_norm": 0.18002946746844742, |
| "learning_rate": 5.246546892303766e-05, |
| "loss": 0.5766, |
| "step": 349 |
| }, |
| { |
| "epoch": 2.3015775455849212, |
| "grad_norm": 0.2532930596607861, |
| "learning_rate": 5.229076814318122e-05, |
| "loss": 0.5742, |
| "step": 350 |
| }, |
| { |
| "epoch": 2.3081335791845934, |
| "grad_norm": 0.3187321065178083, |
| "learning_rate": 5.211580808524325e-05, |
| "loss": 0.5739, |
| "step": 351 |
| }, |
| { |
| "epoch": 2.3146896127842655, |
| "grad_norm": 0.3465322974483502, |
| "learning_rate": 5.194059244006779e-05, |
| "loss": 0.5699, |
| "step": 352 |
| }, |
| { |
| "epoch": 2.3212456463839377, |
| "grad_norm": 0.48722607810944585, |
| "learning_rate": 5.176512490389055e-05, |
| "loss": 0.5756, |
| "step": 353 |
| }, |
| { |
| "epoch": 2.32780167998361, |
| "grad_norm": 1.1074199903854496, |
| "learning_rate": 5.158940917826099e-05, |
| "loss": 0.5832, |
| "step": 354 |
| }, |
| { |
| "epoch": 2.334357713583282, |
| "grad_norm": 0.3988094901700725, |
| "learning_rate": 5.141344896996422e-05, |
| "loss": 0.5871, |
| "step": 355 |
| }, |
| { |
| "epoch": 2.340913747182954, |
| "grad_norm": 0.3381422893134184, |
| "learning_rate": 5.123724799094279e-05, |
| "loss": 0.5908, |
| "step": 356 |
| }, |
| { |
| "epoch": 2.3474697807826264, |
| "grad_norm": 0.30952489856697063, |
| "learning_rate": 5.106080995821836e-05, |
| "loss": 0.5763, |
| "step": 357 |
| }, |
| { |
| "epoch": 2.3540258143822985, |
| "grad_norm": 0.3302426550471252, |
| "learning_rate": 5.088413859381341e-05, |
| "loss": 0.5796, |
| "step": 358 |
| }, |
| { |
| "epoch": 2.3605818479819707, |
| "grad_norm": 0.2513297320397786, |
| "learning_rate": 5.070723762467254e-05, |
| "loss": 0.5749, |
| "step": 359 |
| }, |
| { |
| "epoch": 2.3671378815816433, |
| "grad_norm": 0.26942873256331556, |
| "learning_rate": 5.053011078258397e-05, |
| "loss": 0.5782, |
| "step": 360 |
| }, |
| { |
| "epoch": 2.3736939151813155, |
| "grad_norm": 0.26811935982451485, |
| "learning_rate": 5.0352761804100835e-05, |
| "loss": 0.5893, |
| "step": 361 |
| }, |
| { |
| "epoch": 2.3802499487809876, |
| "grad_norm": 0.24961381636090246, |
| "learning_rate": 5.017519443046226e-05, |
| "loss": 0.5752, |
| "step": 362 |
| }, |
| { |
| "epoch": 2.3868059823806598, |
| "grad_norm": 0.2789479751757785, |
| "learning_rate": 4.999741240751451e-05, |
| "loss": 0.5819, |
| "step": 363 |
| }, |
| { |
| "epoch": 2.393362015980332, |
| "grad_norm": 0.23968714486879794, |
| "learning_rate": 4.981941948563197e-05, |
| "loss": 0.5864, |
| "step": 364 |
| }, |
| { |
| "epoch": 2.399918049580004, |
| "grad_norm": 0.3892339244230853, |
| "learning_rate": 4.9641219419637985e-05, |
| "loss": 0.589, |
| "step": 365 |
| }, |
| { |
| "epoch": 2.4064740831796763, |
| "grad_norm": 0.19471837418765098, |
| "learning_rate": 4.94628159687257e-05, |
| "loss": 0.5841, |
| "step": 366 |
| }, |
| { |
| "epoch": 2.4130301167793484, |
| "grad_norm": 0.19750205892208292, |
| "learning_rate": 4.928421289637871e-05, |
| "loss": 0.5735, |
| "step": 367 |
| }, |
| { |
| "epoch": 2.4195861503790206, |
| "grad_norm": 0.2849687278419682, |
| "learning_rate": 4.9105413970291747e-05, |
| "loss": 0.5806, |
| "step": 368 |
| }, |
| { |
| "epoch": 2.4261421839786927, |
| "grad_norm": 0.4039957657566193, |
| "learning_rate": 4.892642296229107e-05, |
| "loss": 0.5802, |
| "step": 369 |
| }, |
| { |
| "epoch": 2.432698217578365, |
| "grad_norm": 0.339476402567376, |
| "learning_rate": 4.874724364825504e-05, |
| "loss": 0.583, |
| "step": 370 |
| }, |
| { |
| "epoch": 2.4392542511780375, |
| "grad_norm": 0.2352273125693621, |
| "learning_rate": 4.856787980803437e-05, |
| "loss": 0.5822, |
| "step": 371 |
| }, |
| { |
| "epoch": 2.4458102847777097, |
| "grad_norm": 0.20048904331921943, |
| "learning_rate": 4.8388335225372416e-05, |
| "loss": 0.5865, |
| "step": 372 |
| }, |
| { |
| "epoch": 2.452366318377382, |
| "grad_norm": 0.20167869688700724, |
| "learning_rate": 4.820861368782537e-05, |
| "loss": 0.5753, |
| "step": 373 |
| }, |
| { |
| "epoch": 2.458922351977054, |
| "grad_norm": 0.30771705944153094, |
| "learning_rate": 4.802871898668237e-05, |
| "loss": 0.5888, |
| "step": 374 |
| }, |
| { |
| "epoch": 2.465478385576726, |
| "grad_norm": 0.38232619534443235, |
| "learning_rate": 4.7848654916885446e-05, |
| "loss": 0.5729, |
| "step": 375 |
| }, |
| { |
| "epoch": 2.4720344191763983, |
| "grad_norm": 0.2297372798599205, |
| "learning_rate": 4.7668425276949546e-05, |
| "loss": 0.5806, |
| "step": 376 |
| }, |
| { |
| "epoch": 2.4785904527760705, |
| "grad_norm": 0.22126295840938795, |
| "learning_rate": 4.74880338688824e-05, |
| "loss": 0.5761, |
| "step": 377 |
| }, |
| { |
| "epoch": 2.4851464863757426, |
| "grad_norm": 0.2882635600513661, |
| "learning_rate": 4.730748449810429e-05, |
| "loss": 0.577, |
| "step": 378 |
| }, |
| { |
| "epoch": 2.491702519975415, |
| "grad_norm": 0.306926273041758, |
| "learning_rate": 4.712678097336773e-05, |
| "loss": 0.5822, |
| "step": 379 |
| }, |
| { |
| "epoch": 2.498258553575087, |
| "grad_norm": 0.22726474030427726, |
| "learning_rate": 4.694592710667723e-05, |
| "loss": 0.5795, |
| "step": 380 |
| }, |
| { |
| "epoch": 2.504814587174759, |
| "grad_norm": 0.25322867792835857, |
| "learning_rate": 4.6764926713208756e-05, |
| "loss": 0.5762, |
| "step": 381 |
| }, |
| { |
| "epoch": 2.5113706207744313, |
| "grad_norm": 0.22298267407895275, |
| "learning_rate": 4.658378361122936e-05, |
| "loss": 0.5737, |
| "step": 382 |
| }, |
| { |
| "epoch": 2.5179266543741035, |
| "grad_norm": 0.20258291898015557, |
| "learning_rate": 4.640250162201656e-05, |
| "loss": 0.5787, |
| "step": 383 |
| }, |
| { |
| "epoch": 2.5244826879737756, |
| "grad_norm": 0.22752302168713118, |
| "learning_rate": 4.622108456977773e-05, |
| "loss": 0.58, |
| "step": 384 |
| }, |
| { |
| "epoch": 2.5310387215734482, |
| "grad_norm": 0.7757782276758474, |
| "learning_rate": 4.6039536281569476e-05, |
| "loss": 0.5823, |
| "step": 385 |
| }, |
| { |
| "epoch": 2.5375947551731204, |
| "grad_norm": 0.2825695671491106, |
| "learning_rate": 4.585786058721687e-05, |
| "loss": 0.5617, |
| "step": 386 |
| }, |
| { |
| "epoch": 2.5441507887727925, |
| "grad_norm": 0.18948706553634664, |
| "learning_rate": 4.567606131923263e-05, |
| "loss": 0.5747, |
| "step": 387 |
| }, |
| { |
| "epoch": 2.5507068223724647, |
| "grad_norm": 0.21721693917710894, |
| "learning_rate": 4.549414231273633e-05, |
| "loss": 0.5829, |
| "step": 388 |
| }, |
| { |
| "epoch": 2.557262855972137, |
| "grad_norm": 0.2383552245325273, |
| "learning_rate": 4.531210740537347e-05, |
| "loss": 0.5845, |
| "step": 389 |
| }, |
| { |
| "epoch": 2.563818889571809, |
| "grad_norm": 0.30583160784152247, |
| "learning_rate": 4.512996043723453e-05, |
| "loss": 0.5865, |
| "step": 390 |
| }, |
| { |
| "epoch": 2.570374923171481, |
| "grad_norm": 0.2812991313238803, |
| "learning_rate": 4.494770525077392e-05, |
| "loss": 0.5886, |
| "step": 391 |
| }, |
| { |
| "epoch": 2.5769309567711534, |
| "grad_norm": 0.2573334610716088, |
| "learning_rate": 4.476534569072895e-05, |
| "loss": 0.5812, |
| "step": 392 |
| }, |
| { |
| "epoch": 2.5834869903708255, |
| "grad_norm": 0.18065438415446586, |
| "learning_rate": 4.458288560403878e-05, |
| "loss": 0.5745, |
| "step": 393 |
| }, |
| { |
| "epoch": 2.590043023970498, |
| "grad_norm": 0.2944192743209924, |
| "learning_rate": 4.440032883976318e-05, |
| "loss": 0.5771, |
| "step": 394 |
| }, |
| { |
| "epoch": 2.5965990575701703, |
| "grad_norm": 0.23346564334137798, |
| "learning_rate": 4.421767924900136e-05, |
| "loss": 0.5773, |
| "step": 395 |
| }, |
| { |
| "epoch": 2.6031550911698424, |
| "grad_norm": 0.2274103844648281, |
| "learning_rate": 4.403494068481074e-05, |
| "loss": 0.5792, |
| "step": 396 |
| }, |
| { |
| "epoch": 2.6097111247695146, |
| "grad_norm": 0.18276656304603, |
| "learning_rate": 4.385211700212567e-05, |
| "loss": 0.5858, |
| "step": 397 |
| }, |
| { |
| "epoch": 2.6162671583691868, |
| "grad_norm": 0.18285901538118296, |
| "learning_rate": 4.3669212057676145e-05, |
| "loss": 0.5799, |
| "step": 398 |
| }, |
| { |
| "epoch": 2.622823191968859, |
| "grad_norm": 0.18097190512177241, |
| "learning_rate": 4.348622970990634e-05, |
| "loss": 0.5739, |
| "step": 399 |
| }, |
| { |
| "epoch": 2.629379225568531, |
| "grad_norm": 0.1930953632606967, |
| "learning_rate": 4.33031738188933e-05, |
| "loss": 0.5753, |
| "step": 400 |
| }, |
| { |
| "epoch": 2.6359352591682033, |
| "grad_norm": 0.17356931537284068, |
| "learning_rate": 4.312004824626551e-05, |
| "loss": 0.5777, |
| "step": 401 |
| }, |
| { |
| "epoch": 2.6424912927678754, |
| "grad_norm": 0.1734706404199846, |
| "learning_rate": 4.293685685512142e-05, |
| "loss": 0.5752, |
| "step": 402 |
| }, |
| { |
| "epoch": 2.6490473263675476, |
| "grad_norm": 0.19379812038615699, |
| "learning_rate": 4.275360350994791e-05, |
| "loss": 0.5795, |
| "step": 403 |
| }, |
| { |
| "epoch": 2.6556033599672197, |
| "grad_norm": 0.1660813082194996, |
| "learning_rate": 4.257029207653881e-05, |
| "loss": 0.5826, |
| "step": 404 |
| }, |
| { |
| "epoch": 2.662159393566892, |
| "grad_norm": 0.19954108364376924, |
| "learning_rate": 4.238692642191336e-05, |
| "loss": 0.5825, |
| "step": 405 |
| }, |
| { |
| "epoch": 2.668715427166564, |
| "grad_norm": 0.19322396699850317, |
| "learning_rate": 4.220351041423462e-05, |
| "loss": 0.5841, |
| "step": 406 |
| }, |
| { |
| "epoch": 2.6752714607662362, |
| "grad_norm": 0.17735120440062302, |
| "learning_rate": 4.202004792272785e-05, |
| "loss": 0.5798, |
| "step": 407 |
| }, |
| { |
| "epoch": 2.6818274943659084, |
| "grad_norm": 0.17855572539839387, |
| "learning_rate": 4.183654281759888e-05, |
| "loss": 0.5785, |
| "step": 408 |
| }, |
| { |
| "epoch": 2.6883835279655806, |
| "grad_norm": 0.1998138629729567, |
| "learning_rate": 4.165299896995253e-05, |
| "loss": 0.5782, |
| "step": 409 |
| }, |
| { |
| "epoch": 2.694939561565253, |
| "grad_norm": 0.19733364770879444, |
| "learning_rate": 4.1469420251710905e-05, |
| "loss": 0.5738, |
| "step": 410 |
| }, |
| { |
| "epoch": 2.7014955951649253, |
| "grad_norm": 0.17455960566932077, |
| "learning_rate": 4.128581053553169e-05, |
| "loss": 0.5804, |
| "step": 411 |
| }, |
| { |
| "epoch": 2.7080516287645975, |
| "grad_norm": 0.2277904402914904, |
| "learning_rate": 4.110217369472649e-05, |
| "loss": 0.5823, |
| "step": 412 |
| }, |
| { |
| "epoch": 2.7146076623642696, |
| "grad_norm": 0.17727829284042818, |
| "learning_rate": 4.091851360317912e-05, |
| "loss": 0.5806, |
| "step": 413 |
| }, |
| { |
| "epoch": 2.721163695963942, |
| "grad_norm": 0.20033803661764574, |
| "learning_rate": 4.07348341352639e-05, |
| "loss": 0.5737, |
| "step": 414 |
| }, |
| { |
| "epoch": 2.727719729563614, |
| "grad_norm": 0.28585140789050134, |
| "learning_rate": 4.055113916576386e-05, |
| "loss": 0.5922, |
| "step": 415 |
| }, |
| { |
| "epoch": 2.734275763163286, |
| "grad_norm": 0.1770226397602663, |
| "learning_rate": 4.0367432569789065e-05, |
| "loss": 0.5708, |
| "step": 416 |
| }, |
| { |
| "epoch": 2.7408317967629583, |
| "grad_norm": 0.19810748724107435, |
| "learning_rate": 4.0183718222694823e-05, |
| "loss": 0.5822, |
| "step": 417 |
| }, |
| { |
| "epoch": 2.7473878303626305, |
| "grad_norm": 0.16916185476319684, |
| "learning_rate": 4e-05, |
| "loss": 0.5781, |
| "step": 418 |
| }, |
| { |
| "epoch": 2.753943863962303, |
| "grad_norm": 0.1832943105404772, |
| "learning_rate": 3.9816281777305176e-05, |
| "loss": 0.5812, |
| "step": 419 |
| }, |
| { |
| "epoch": 2.7604998975619752, |
| "grad_norm": 0.1502272563296327, |
| "learning_rate": 3.963256743021095e-05, |
| "loss": 0.5716, |
| "step": 420 |
| }, |
| { |
| "epoch": 2.7670559311616474, |
| "grad_norm": 0.21339511828926633, |
| "learning_rate": 3.944886083423615e-05, |
| "loss": 0.5801, |
| "step": 421 |
| }, |
| { |
| "epoch": 2.7736119647613195, |
| "grad_norm": 0.177457878421534, |
| "learning_rate": 3.92651658647361e-05, |
| "loss": 0.5747, |
| "step": 422 |
| }, |
| { |
| "epoch": 2.7801679983609917, |
| "grad_norm": 0.1839395596929418, |
| "learning_rate": 3.908148639682089e-05, |
| "loss": 0.5843, |
| "step": 423 |
| }, |
| { |
| "epoch": 2.786724031960664, |
| "grad_norm": 0.19753047943291085, |
| "learning_rate": 3.889782630527353e-05, |
| "loss": 0.574, |
| "step": 424 |
| }, |
| { |
| "epoch": 2.793280065560336, |
| "grad_norm": 0.14640408330908058, |
| "learning_rate": 3.8714189464468334e-05, |
| "loss": 0.5786, |
| "step": 425 |
| }, |
| { |
| "epoch": 2.799836099160008, |
| "grad_norm": 0.19103262838391485, |
| "learning_rate": 3.853057974828911e-05, |
| "loss": 0.5798, |
| "step": 426 |
| }, |
| { |
| "epoch": 2.8063921327596804, |
| "grad_norm": 0.1673182640044429, |
| "learning_rate": 3.834700103004747e-05, |
| "loss": 0.5729, |
| "step": 427 |
| }, |
| { |
| "epoch": 2.8129481663593525, |
| "grad_norm": 0.20105657869010776, |
| "learning_rate": 3.816345718240113e-05, |
| "loss": 0.5744, |
| "step": 428 |
| }, |
| { |
| "epoch": 2.8195041999590247, |
| "grad_norm": 0.13814925979553985, |
| "learning_rate": 3.797995207727217e-05, |
| "loss": 0.5694, |
| "step": 429 |
| }, |
| { |
| "epoch": 2.826060233558697, |
| "grad_norm": 0.16253212459987165, |
| "learning_rate": 3.779648958576538e-05, |
| "loss": 0.5732, |
| "step": 430 |
| }, |
| { |
| "epoch": 2.832616267158369, |
| "grad_norm": 0.14306544183315967, |
| "learning_rate": 3.7613073578086644e-05, |
| "loss": 0.5737, |
| "step": 431 |
| }, |
| { |
| "epoch": 2.839172300758041, |
| "grad_norm": 0.16148301199590984, |
| "learning_rate": 3.74297079234612e-05, |
| "loss": 0.5696, |
| "step": 432 |
| }, |
| { |
| "epoch": 2.8457283343577133, |
| "grad_norm": 0.1591738101867343, |
| "learning_rate": 3.7246396490052117e-05, |
| "loss": 0.5752, |
| "step": 433 |
| }, |
| { |
| "epoch": 2.852284367957386, |
| "grad_norm": 0.14229281054940401, |
| "learning_rate": 3.706314314487859e-05, |
| "loss": 0.5778, |
| "step": 434 |
| }, |
| { |
| "epoch": 2.858840401557058, |
| "grad_norm": 0.18401979626969228, |
| "learning_rate": 3.687995175373449e-05, |
| "loss": 0.5785, |
| "step": 435 |
| }, |
| { |
| "epoch": 2.8653964351567303, |
| "grad_norm": 0.14532042872787476, |
| "learning_rate": 3.669682618110671e-05, |
| "loss": 0.5717, |
| "step": 436 |
| }, |
| { |
| "epoch": 2.8719524687564024, |
| "grad_norm": 0.14280350348324158, |
| "learning_rate": 3.6513770290093674e-05, |
| "loss": 0.5737, |
| "step": 437 |
| }, |
| { |
| "epoch": 2.8785085023560746, |
| "grad_norm": 0.17828968160150296, |
| "learning_rate": 3.6330787942323855e-05, |
| "loss": 0.5739, |
| "step": 438 |
| }, |
| { |
| "epoch": 2.8850645359557467, |
| "grad_norm": 0.1680340039645173, |
| "learning_rate": 3.614788299787434e-05, |
| "loss": 0.5719, |
| "step": 439 |
| }, |
| { |
| "epoch": 2.891620569555419, |
| "grad_norm": 0.16773501838978377, |
| "learning_rate": 3.5965059315189274e-05, |
| "loss": 0.5861, |
| "step": 440 |
| }, |
| { |
| "epoch": 2.898176603155091, |
| "grad_norm": 0.1466119984148444, |
| "learning_rate": 3.578232075099866e-05, |
| "loss": 0.5719, |
| "step": 441 |
| }, |
| { |
| "epoch": 2.9047326367547632, |
| "grad_norm": 0.18860006392370388, |
| "learning_rate": 3.559967116023683e-05, |
| "loss": 0.573, |
| "step": 442 |
| }, |
| { |
| "epoch": 2.911288670354436, |
| "grad_norm": 0.17180023728534166, |
| "learning_rate": 3.541711439596122e-05, |
| "loss": 0.5715, |
| "step": 443 |
| }, |
| { |
| "epoch": 2.917844703954108, |
| "grad_norm": 0.16863095150334853, |
| "learning_rate": 3.523465430927106e-05, |
| "loss": 0.587, |
| "step": 444 |
| }, |
| { |
| "epoch": 2.92440073755378, |
| "grad_norm": 0.18865259295148487, |
| "learning_rate": 3.5052294749226094e-05, |
| "loss": 0.5705, |
| "step": 445 |
| }, |
| { |
| "epoch": 2.9309567711534523, |
| "grad_norm": 0.3900404738786457, |
| "learning_rate": 3.4870039562765475e-05, |
| "loss": 0.5808, |
| "step": 446 |
| }, |
| { |
| "epoch": 2.9375128047531245, |
| "grad_norm": 0.1577690647524753, |
| "learning_rate": 3.4687892594626536e-05, |
| "loss": 0.5767, |
| "step": 447 |
| }, |
| { |
| "epoch": 2.9440688383527966, |
| "grad_norm": 0.17930301136018978, |
| "learning_rate": 3.4505857687263675e-05, |
| "loss": 0.5794, |
| "step": 448 |
| }, |
| { |
| "epoch": 2.950624871952469, |
| "grad_norm": 0.21442963322771297, |
| "learning_rate": 3.432393868076739e-05, |
| "loss": 0.5813, |
| "step": 449 |
| }, |
| { |
| "epoch": 2.957180905552141, |
| "grad_norm": 0.5263365227065998, |
| "learning_rate": 3.414213941278314e-05, |
| "loss": 0.5712, |
| "step": 450 |
| }, |
| { |
| "epoch": 2.963736939151813, |
| "grad_norm": 0.24486881857008144, |
| "learning_rate": 3.396046371843052e-05, |
| "loss": 0.5862, |
| "step": 451 |
| }, |
| { |
| "epoch": 2.9702929727514853, |
| "grad_norm": 0.14980108475089587, |
| "learning_rate": 3.377891543022229e-05, |
| "loss": 0.5801, |
| "step": 452 |
| }, |
| { |
| "epoch": 2.9768490063511575, |
| "grad_norm": 0.20536856095424458, |
| "learning_rate": 3.3597498377983444e-05, |
| "loss": 0.5802, |
| "step": 453 |
| }, |
| { |
| "epoch": 2.9834050399508296, |
| "grad_norm": 0.18782303046647886, |
| "learning_rate": 3.341621638877064e-05, |
| "loss": 0.5784, |
| "step": 454 |
| }, |
| { |
| "epoch": 2.989961073550502, |
| "grad_norm": 0.14515446375387317, |
| "learning_rate": 3.3235073286791264e-05, |
| "loss": 0.5754, |
| "step": 455 |
| }, |
| { |
| "epoch": 2.996517107150174, |
| "grad_norm": 0.16161434398106808, |
| "learning_rate": 3.305407289332279e-05, |
| "loss": 0.5756, |
| "step": 456 |
| }, |
| { |
| "epoch": 3.0055316533497236, |
| "grad_norm": 0.3533323244538683, |
| "learning_rate": 3.287321902663229e-05, |
| "loss": 1.0491, |
| "step": 457 |
| }, |
| { |
| "epoch": 3.0120876869493958, |
| "grad_norm": 0.3045514737546258, |
| "learning_rate": 3.269251550189573e-05, |
| "loss": 0.5446, |
| "step": 458 |
| }, |
| { |
| "epoch": 3.018643720549068, |
| "grad_norm": 0.25472049022532234, |
| "learning_rate": 3.251196613111761e-05, |
| "loss": 0.5407, |
| "step": 459 |
| }, |
| { |
| "epoch": 3.02519975414874, |
| "grad_norm": 0.25510858573604617, |
| "learning_rate": 3.2331574723050474e-05, |
| "loss": 0.5411, |
| "step": 460 |
| }, |
| { |
| "epoch": 3.0317557877484123, |
| "grad_norm": 0.38483951060765287, |
| "learning_rate": 3.2151345083114574e-05, |
| "loss": 0.5396, |
| "step": 461 |
| }, |
| { |
| "epoch": 3.0383118213480844, |
| "grad_norm": 0.2558618267692145, |
| "learning_rate": 3.197128101331764e-05, |
| "loss": 0.5349, |
| "step": 462 |
| }, |
| { |
| "epoch": 3.0448678549477566, |
| "grad_norm": 0.23801295747592094, |
| "learning_rate": 3.179138631217463e-05, |
| "loss": 0.5439, |
| "step": 463 |
| }, |
| { |
| "epoch": 3.0514238885474287, |
| "grad_norm": 0.2746658371274729, |
| "learning_rate": 3.161166477462759e-05, |
| "loss": 0.5463, |
| "step": 464 |
| }, |
| { |
| "epoch": 3.057979922147101, |
| "grad_norm": 0.23307565195751595, |
| "learning_rate": 3.1432120191965647e-05, |
| "loss": 0.538, |
| "step": 465 |
| }, |
| { |
| "epoch": 3.064535955746773, |
| "grad_norm": 0.23849808366960823, |
| "learning_rate": 3.125275635174497e-05, |
| "loss": 0.5431, |
| "step": 466 |
| }, |
| { |
| "epoch": 3.0710919893464452, |
| "grad_norm": 0.22477045443908192, |
| "learning_rate": 3.1073577037708935e-05, |
| "loss": 0.5422, |
| "step": 467 |
| }, |
| { |
| "epoch": 3.0776480229461174, |
| "grad_norm": 0.2060382177048335, |
| "learning_rate": 3.089458602970828e-05, |
| "loss": 0.5446, |
| "step": 468 |
| }, |
| { |
| "epoch": 3.08420405654579, |
| "grad_norm": 0.22200054079896087, |
| "learning_rate": 3.0715787103621294e-05, |
| "loss": 0.5479, |
| "step": 469 |
| }, |
| { |
| "epoch": 3.090760090145462, |
| "grad_norm": 0.17039781849738478, |
| "learning_rate": 3.0537184031274306e-05, |
| "loss": 0.5305, |
| "step": 470 |
| }, |
| { |
| "epoch": 3.0973161237451343, |
| "grad_norm": 0.16941181980633563, |
| "learning_rate": 3.0358780580362025e-05, |
| "loss": 0.5261, |
| "step": 471 |
| }, |
| { |
| "epoch": 3.1038721573448065, |
| "grad_norm": 0.17660138294537184, |
| "learning_rate": 3.0180580514368037e-05, |
| "loss": 0.5312, |
| "step": 472 |
| }, |
| { |
| "epoch": 3.1104281909444786, |
| "grad_norm": 0.18073996525501446, |
| "learning_rate": 3.0002587592485497e-05, |
| "loss": 0.5371, |
| "step": 473 |
| }, |
| { |
| "epoch": 3.116984224544151, |
| "grad_norm": 0.3988440151798082, |
| "learning_rate": 2.9824805569537747e-05, |
| "loss": 0.5513, |
| "step": 474 |
| }, |
| { |
| "epoch": 3.123540258143823, |
| "grad_norm": 0.20423237269747027, |
| "learning_rate": 2.9647238195899168e-05, |
| "loss": 0.5375, |
| "step": 475 |
| }, |
| { |
| "epoch": 3.130096291743495, |
| "grad_norm": 0.17213193598614746, |
| "learning_rate": 2.9469889217416045e-05, |
| "loss": 0.5437, |
| "step": 476 |
| }, |
| { |
| "epoch": 3.1366523253431673, |
| "grad_norm": 0.21207611299854595, |
| "learning_rate": 2.9292762375327483e-05, |
| "loss": 0.5339, |
| "step": 477 |
| }, |
| { |
| "epoch": 3.1432083589428395, |
| "grad_norm": 0.15310274061749113, |
| "learning_rate": 2.9115861406186593e-05, |
| "loss": 0.5303, |
| "step": 478 |
| }, |
| { |
| "epoch": 3.1497643925425116, |
| "grad_norm": 0.1877335953144904, |
| "learning_rate": 2.8939190041781647e-05, |
| "loss": 0.5319, |
| "step": 479 |
| }, |
| { |
| "epoch": 3.1563204261421838, |
| "grad_norm": 0.1458292108542802, |
| "learning_rate": 2.8762752009057232e-05, |
| "loss": 0.54, |
| "step": 480 |
| }, |
| { |
| "epoch": 3.1628764597418564, |
| "grad_norm": 0.17442833234197713, |
| "learning_rate": 2.85865510300358e-05, |
| "loss": 0.5377, |
| "step": 481 |
| }, |
| { |
| "epoch": 3.1694324933415285, |
| "grad_norm": 0.15457635774978962, |
| "learning_rate": 2.841059082173902e-05, |
| "loss": 0.5389, |
| "step": 482 |
| }, |
| { |
| "epoch": 3.1759885269412007, |
| "grad_norm": 0.16823476302413515, |
| "learning_rate": 2.823487509610946e-05, |
| "loss": 0.5435, |
| "step": 483 |
| }, |
| { |
| "epoch": 3.182544560540873, |
| "grad_norm": 0.14604985181804828, |
| "learning_rate": 2.805940755993223e-05, |
| "loss": 0.5377, |
| "step": 484 |
| }, |
| { |
| "epoch": 3.189100594140545, |
| "grad_norm": 0.14560052298572645, |
| "learning_rate": 2.7884191914756757e-05, |
| "loss": 0.5409, |
| "step": 485 |
| }, |
| { |
| "epoch": 3.195656627740217, |
| "grad_norm": 0.1472508058071573, |
| "learning_rate": 2.770923185681878e-05, |
| "loss": 0.5455, |
| "step": 486 |
| }, |
| { |
| "epoch": 3.2022126613398894, |
| "grad_norm": 0.13624234976069768, |
| "learning_rate": 2.7534531076962356e-05, |
| "loss": 0.5433, |
| "step": 487 |
| }, |
| { |
| "epoch": 3.2087686949395615, |
| "grad_norm": 0.15264871662264248, |
| "learning_rate": 2.7360093260561904e-05, |
| "loss": 0.5372, |
| "step": 488 |
| }, |
| { |
| "epoch": 3.2153247285392337, |
| "grad_norm": 0.1462659967641793, |
| "learning_rate": 2.7185922087444602e-05, |
| "loss": 0.538, |
| "step": 489 |
| }, |
| { |
| "epoch": 3.221880762138906, |
| "grad_norm": 0.16303592219574492, |
| "learning_rate": 2.7012021231812666e-05, |
| "loss": 0.542, |
| "step": 490 |
| }, |
| { |
| "epoch": 3.228436795738578, |
| "grad_norm": 0.14173407966037618, |
| "learning_rate": 2.6838394362165875e-05, |
| "loss": 0.5387, |
| "step": 491 |
| }, |
| { |
| "epoch": 3.23499282933825, |
| "grad_norm": 0.15432095686939976, |
| "learning_rate": 2.6665045141224193e-05, |
| "loss": 0.5377, |
| "step": 492 |
| }, |
| { |
| "epoch": 3.2415488629379228, |
| "grad_norm": 0.1482561893469378, |
| "learning_rate": 2.6491977225850446e-05, |
| "loss": 0.5371, |
| "step": 493 |
| }, |
| { |
| "epoch": 3.248104896537595, |
| "grad_norm": 0.16624560752182815, |
| "learning_rate": 2.6319194266973256e-05, |
| "loss": 0.5359, |
| "step": 494 |
| }, |
| { |
| "epoch": 3.254660930137267, |
| "grad_norm": 0.15016468812997047, |
| "learning_rate": 2.6146699909509984e-05, |
| "loss": 0.5411, |
| "step": 495 |
| }, |
| { |
| "epoch": 3.2612169637369393, |
| "grad_norm": 0.2450944209511445, |
| "learning_rate": 2.597449779228983e-05, |
| "loss": 0.5429, |
| "step": 496 |
| }, |
| { |
| "epoch": 3.2677729973366114, |
| "grad_norm": 0.2412941457078939, |
| "learning_rate": 2.580259154797709e-05, |
| "loss": 0.5343, |
| "step": 497 |
| }, |
| { |
| "epoch": 3.2743290309362836, |
| "grad_norm": 0.14493021044588159, |
| "learning_rate": 2.563098480299451e-05, |
| "loss": 0.5409, |
| "step": 498 |
| }, |
| { |
| "epoch": 3.2808850645359557, |
| "grad_norm": 0.16960807003233563, |
| "learning_rate": 2.5459681177446803e-05, |
| "loss": 0.5389, |
| "step": 499 |
| }, |
| { |
| "epoch": 3.287441098135628, |
| "grad_norm": 0.14869909968414688, |
| "learning_rate": 2.5288684285044283e-05, |
| "loss": 0.5353, |
| "step": 500 |
| }, |
| { |
| "epoch": 3.2939971317353, |
| "grad_norm": 0.14777225005600433, |
| "learning_rate": 2.5117997733026566e-05, |
| "loss": 0.5393, |
| "step": 501 |
| }, |
| { |
| "epoch": 3.3005531653349722, |
| "grad_norm": 0.15139629935458157, |
| "learning_rate": 2.4947625122086585e-05, |
| "loss": 0.5435, |
| "step": 502 |
| }, |
| { |
| "epoch": 3.3071091989346444, |
| "grad_norm": 0.1392849654786053, |
| "learning_rate": 2.477757004629456e-05, |
| "loss": 0.5375, |
| "step": 503 |
| }, |
| { |
| "epoch": 3.3136652325343166, |
| "grad_norm": 0.1367733939357107, |
| "learning_rate": 2.460783609302218e-05, |
| "loss": 0.5426, |
| "step": 504 |
| }, |
| { |
| "epoch": 3.3202212661339887, |
| "grad_norm": 0.12699654005182315, |
| "learning_rate": 2.4438426842866966e-05, |
| "loss": 0.5353, |
| "step": 505 |
| }, |
| { |
| "epoch": 3.3267772997336613, |
| "grad_norm": 0.1390969751788505, |
| "learning_rate": 2.4269345869576676e-05, |
| "loss": 0.539, |
| "step": 506 |
| }, |
| { |
| "epoch": 3.3333333333333335, |
| "grad_norm": 0.12248688460845146, |
| "learning_rate": 2.4100596739973993e-05, |
| "loss": 0.5392, |
| "step": 507 |
| }, |
| { |
| "epoch": 3.3398893669330056, |
| "grad_norm": 0.13805750150926843, |
| "learning_rate": 2.393218301388123e-05, |
| "loss": 0.5401, |
| "step": 508 |
| }, |
| { |
| "epoch": 3.346445400532678, |
| "grad_norm": 0.12180259855826225, |
| "learning_rate": 2.3764108244045212e-05, |
| "loss": 0.5465, |
| "step": 509 |
| }, |
| { |
| "epoch": 3.35300143413235, |
| "grad_norm": 0.12245696463081471, |
| "learning_rate": 2.35963759760624e-05, |
| "loss": 0.5311, |
| "step": 510 |
| }, |
| { |
| "epoch": 3.359557467732022, |
| "grad_norm": 0.12896584095171493, |
| "learning_rate": 2.342898974830405e-05, |
| "loss": 0.5391, |
| "step": 511 |
| }, |
| { |
| "epoch": 3.3661135013316943, |
| "grad_norm": 0.12281667944621683, |
| "learning_rate": 2.3261953091841553e-05, |
| "loss": 0.5335, |
| "step": 512 |
| }, |
| { |
| "epoch": 3.3726695349313665, |
| "grad_norm": 0.11829751242108535, |
| "learning_rate": 2.3095269530372032e-05, |
| "loss": 0.5447, |
| "step": 513 |
| }, |
| { |
| "epoch": 3.3792255685310386, |
| "grad_norm": 0.13317495284187636, |
| "learning_rate": 2.2928942580143855e-05, |
| "loss": 0.5438, |
| "step": 514 |
| }, |
| { |
| "epoch": 3.3857816021307108, |
| "grad_norm": 0.12369993983516651, |
| "learning_rate": 2.276297574988263e-05, |
| "loss": 0.5433, |
| "step": 515 |
| }, |
| { |
| "epoch": 3.392337635730383, |
| "grad_norm": 0.1246149968311202, |
| "learning_rate": 2.2597372540717083e-05, |
| "loss": 0.5412, |
| "step": 516 |
| }, |
| { |
| "epoch": 3.3988936693300555, |
| "grad_norm": 0.12919720585139893, |
| "learning_rate": 2.2432136446105192e-05, |
| "loss": 0.5442, |
| "step": 517 |
| }, |
| { |
| "epoch": 3.4054497029297277, |
| "grad_norm": 0.12089802144087428, |
| "learning_rate": 2.226727095176057e-05, |
| "loss": 0.539, |
| "step": 518 |
| }, |
| { |
| "epoch": 3.4120057365294, |
| "grad_norm": 0.13804336756907107, |
| "learning_rate": 2.210277953557888e-05, |
| "loss": 0.5462, |
| "step": 519 |
| }, |
| { |
| "epoch": 3.418561770129072, |
| "grad_norm": 0.12982437086238163, |
| "learning_rate": 2.1938665667564435e-05, |
| "loss": 0.5344, |
| "step": 520 |
| }, |
| { |
| "epoch": 3.425117803728744, |
| "grad_norm": 0.12345004982954612, |
| "learning_rate": 2.177493280975708e-05, |
| "loss": 0.5465, |
| "step": 521 |
| }, |
| { |
| "epoch": 3.4316738373284164, |
| "grad_norm": 0.14573537922186017, |
| "learning_rate": 2.1611584416159106e-05, |
| "loss": 0.5323, |
| "step": 522 |
| }, |
| { |
| "epoch": 3.4382298709280885, |
| "grad_norm": 0.1097589192258888, |
| "learning_rate": 2.1448623932662377e-05, |
| "loss": 0.5437, |
| "step": 523 |
| }, |
| { |
| "epoch": 3.4447859045277607, |
| "grad_norm": 0.1317797633987106, |
| "learning_rate": 2.1286054796975696e-05, |
| "loss": 0.5377, |
| "step": 524 |
| }, |
| { |
| "epoch": 3.451341938127433, |
| "grad_norm": 0.13321471541724095, |
| "learning_rate": 2.1123880438552187e-05, |
| "loss": 0.5518, |
| "step": 525 |
| }, |
| { |
| "epoch": 3.457897971727105, |
| "grad_norm": 0.1352447131480959, |
| "learning_rate": 2.096210427851706e-05, |
| "loss": 0.532, |
| "step": 526 |
| }, |
| { |
| "epoch": 3.464454005326777, |
| "grad_norm": 0.13508733978142928, |
| "learning_rate": 2.0800729729595385e-05, |
| "loss": 0.5314, |
| "step": 527 |
| }, |
| { |
| "epoch": 3.4710100389264493, |
| "grad_norm": 0.1380525297606027, |
| "learning_rate": 2.063976019604006e-05, |
| "loss": 0.5438, |
| "step": 528 |
| }, |
| { |
| "epoch": 3.4775660725261215, |
| "grad_norm": 0.13529984461100034, |
| "learning_rate": 2.0479199073560084e-05, |
| "loss": 0.5451, |
| "step": 529 |
| }, |
| { |
| "epoch": 3.4841221061257936, |
| "grad_norm": 0.13383345341162217, |
| "learning_rate": 2.0319049749248876e-05, |
| "loss": 0.5335, |
| "step": 530 |
| }, |
| { |
| "epoch": 3.4906781397254663, |
| "grad_norm": 0.12966276679499278, |
| "learning_rate": 2.0159315601512817e-05, |
| "loss": 0.5404, |
| "step": 531 |
| }, |
| { |
| "epoch": 3.4972341733251384, |
| "grad_norm": 0.12661938479816495, |
| "learning_rate": 2.0000000000000012e-05, |
| "loss": 0.5361, |
| "step": 532 |
| }, |
| { |
| "epoch": 3.5037902069248106, |
| "grad_norm": 0.25950697427544217, |
| "learning_rate": 1.9841106305529133e-05, |
| "loss": 0.5475, |
| "step": 533 |
| }, |
| { |
| "epoch": 3.5103462405244827, |
| "grad_norm": 0.12388736409274281, |
| "learning_rate": 1.9682637870018638e-05, |
| "loss": 0.5433, |
| "step": 534 |
| }, |
| { |
| "epoch": 3.516902274124155, |
| "grad_norm": 0.13692640136699888, |
| "learning_rate": 1.9524598036415973e-05, |
| "loss": 0.5379, |
| "step": 535 |
| }, |
| { |
| "epoch": 3.523458307723827, |
| "grad_norm": 0.22971792027818427, |
| "learning_rate": 1.9366990138627054e-05, |
| "loss": 0.5354, |
| "step": 536 |
| }, |
| { |
| "epoch": 3.5300143413234992, |
| "grad_norm": 0.13477440715558586, |
| "learning_rate": 1.9209817501445978e-05, |
| "loss": 0.5408, |
| "step": 537 |
| }, |
| { |
| "epoch": 3.5365703749231714, |
| "grad_norm": 0.12000818819921334, |
| "learning_rate": 1.9053083440484887e-05, |
| "loss": 0.539, |
| "step": 538 |
| }, |
| { |
| "epoch": 3.5431264085228436, |
| "grad_norm": 0.18018036064763043, |
| "learning_rate": 1.889679126210397e-05, |
| "loss": 0.54, |
| "step": 539 |
| }, |
| { |
| "epoch": 3.5496824421225157, |
| "grad_norm": 0.13187429164193115, |
| "learning_rate": 1.8740944263341773e-05, |
| "loss": 0.5336, |
| "step": 540 |
| }, |
| { |
| "epoch": 3.5562384757221883, |
| "grad_norm": 0.1115344100782877, |
| "learning_rate": 1.8585545731845584e-05, |
| "loss": 0.5457, |
| "step": 541 |
| }, |
| { |
| "epoch": 3.5627945093218605, |
| "grad_norm": 0.4821099792222361, |
| "learning_rate": 1.8430598945802156e-05, |
| "loss": 0.5429, |
| "step": 542 |
| }, |
| { |
| "epoch": 3.5693505429215326, |
| "grad_norm": 0.13726412619312758, |
| "learning_rate": 1.8276107173868503e-05, |
| "loss": 0.551, |
| "step": 543 |
| }, |
| { |
| "epoch": 3.575906576521205, |
| "grad_norm": 0.11727031408692318, |
| "learning_rate": 1.8122073675102935e-05, |
| "loss": 0.5315, |
| "step": 544 |
| }, |
| { |
| "epoch": 3.582462610120877, |
| "grad_norm": 0.12442269217723008, |
| "learning_rate": 1.7968501698896346e-05, |
| "loss": 0.5338, |
| "step": 545 |
| }, |
| { |
| "epoch": 3.589018643720549, |
| "grad_norm": 0.11894782465239562, |
| "learning_rate": 1.781539448490365e-05, |
| "loss": 0.5365, |
| "step": 546 |
| }, |
| { |
| "epoch": 3.5955746773202213, |
| "grad_norm": 0.11002626026600501, |
| "learning_rate": 1.7662755262975432e-05, |
| "loss": 0.5404, |
| "step": 547 |
| }, |
| { |
| "epoch": 3.6021307109198935, |
| "grad_norm": 0.13251223380465485, |
| "learning_rate": 1.7510587253089842e-05, |
| "loss": 0.543, |
| "step": 548 |
| }, |
| { |
| "epoch": 3.6086867445195656, |
| "grad_norm": 0.11425745596834401, |
| "learning_rate": 1.7358893665284595e-05, |
| "loss": 0.5345, |
| "step": 549 |
| }, |
| { |
| "epoch": 3.6152427781192378, |
| "grad_norm": 0.11826825954302657, |
| "learning_rate": 1.7207677699589355e-05, |
| "loss": 0.5463, |
| "step": 550 |
| }, |
| { |
| "epoch": 3.62179881171891, |
| "grad_norm": 0.10903975348972718, |
| "learning_rate": 1.7056942545958167e-05, |
| "loss": 0.5396, |
| "step": 551 |
| }, |
| { |
| "epoch": 3.628354845318582, |
| "grad_norm": 0.11693940603496059, |
| "learning_rate": 1.690669138420215e-05, |
| "loss": 0.5305, |
| "step": 552 |
| }, |
| { |
| "epoch": 3.6349108789182543, |
| "grad_norm": 0.1142560549805311, |
| "learning_rate": 1.6756927383922473e-05, |
| "loss": 0.5391, |
| "step": 553 |
| }, |
| { |
| "epoch": 3.6414669125179264, |
| "grad_norm": 0.10839349884507322, |
| "learning_rate": 1.6607653704443457e-05, |
| "loss": 0.542, |
| "step": 554 |
| }, |
| { |
| "epoch": 3.6480229461175986, |
| "grad_norm": 0.11225360540051456, |
| "learning_rate": 1.6458873494745926e-05, |
| "loss": 0.5388, |
| "step": 555 |
| }, |
| { |
| "epoch": 3.654578979717271, |
| "grad_norm": 0.11188297317863019, |
| "learning_rate": 1.6310589893400804e-05, |
| "loss": 0.5381, |
| "step": 556 |
| }, |
| { |
| "epoch": 3.6611350133169434, |
| "grad_norm": 0.7105498519897432, |
| "learning_rate": 1.6162806028502852e-05, |
| "loss": 0.5411, |
| "step": 557 |
| }, |
| { |
| "epoch": 3.6676910469166155, |
| "grad_norm": 0.11475316554461859, |
| "learning_rate": 1.601552501760473e-05, |
| "loss": 0.5416, |
| "step": 558 |
| }, |
| { |
| "epoch": 3.6742470805162877, |
| "grad_norm": 0.12194606250202103, |
| "learning_rate": 1.5868749967651252e-05, |
| "loss": 0.5396, |
| "step": 559 |
| }, |
| { |
| "epoch": 3.68080311411596, |
| "grad_norm": 0.12661866626370497, |
| "learning_rate": 1.5722483974913737e-05, |
| "loss": 0.531, |
| "step": 560 |
| }, |
| { |
| "epoch": 3.687359147715632, |
| "grad_norm": 0.11256332358317793, |
| "learning_rate": 1.5576730124924822e-05, |
| "loss": 0.5418, |
| "step": 561 |
| }, |
| { |
| "epoch": 3.693915181315304, |
| "grad_norm": 0.11015606275564016, |
| "learning_rate": 1.5431491492413288e-05, |
| "loss": 0.5411, |
| "step": 562 |
| }, |
| { |
| "epoch": 3.7004712149149763, |
| "grad_norm": 0.12074411265164606, |
| "learning_rate": 1.528677114123923e-05, |
| "loss": 0.53, |
| "step": 563 |
| }, |
| { |
| "epoch": 3.7070272485146485, |
| "grad_norm": 0.10893671793883805, |
| "learning_rate": 1.5142572124329418e-05, |
| "loss": 0.5397, |
| "step": 564 |
| }, |
| { |
| "epoch": 3.713583282114321, |
| "grad_norm": 0.11746065306540288, |
| "learning_rate": 1.4998897483612865e-05, |
| "loss": 0.5423, |
| "step": 565 |
| }, |
| { |
| "epoch": 3.7201393157139933, |
| "grad_norm": 0.1126528745871064, |
| "learning_rate": 1.4855750249956718e-05, |
| "loss": 0.5426, |
| "step": 566 |
| }, |
| { |
| "epoch": 3.7266953493136654, |
| "grad_norm": 0.3165608272055915, |
| "learning_rate": 1.4713133443102283e-05, |
| "loss": 0.5428, |
| "step": 567 |
| }, |
| { |
| "epoch": 3.7332513829133376, |
| "grad_norm": 0.11800782137431694, |
| "learning_rate": 1.457105007160129e-05, |
| "loss": 0.5443, |
| "step": 568 |
| }, |
| { |
| "epoch": 3.7398074165130097, |
| "grad_norm": 0.11122501505766147, |
| "learning_rate": 1.44295031327525e-05, |
| "loss": 0.5397, |
| "step": 569 |
| }, |
| { |
| "epoch": 3.746363450112682, |
| "grad_norm": 0.12798481597310407, |
| "learning_rate": 1.4288495612538427e-05, |
| "loss": 0.5457, |
| "step": 570 |
| }, |
| { |
| "epoch": 3.752919483712354, |
| "grad_norm": 0.10441587675286511, |
| "learning_rate": 1.4148030485562362e-05, |
| "loss": 0.5371, |
| "step": 571 |
| }, |
| { |
| "epoch": 3.7594755173120262, |
| "grad_norm": 0.12743357362001942, |
| "learning_rate": 1.4008110714985623e-05, |
| "loss": 0.5392, |
| "step": 572 |
| }, |
| { |
| "epoch": 3.7660315509116984, |
| "grad_norm": 0.1114224145991442, |
| "learning_rate": 1.3868739252465017e-05, |
| "loss": 0.5368, |
| "step": 573 |
| }, |
| { |
| "epoch": 3.7725875845113706, |
| "grad_norm": 0.10668882720776016, |
| "learning_rate": 1.3729919038090627e-05, |
| "loss": 0.5386, |
| "step": 574 |
| }, |
| { |
| "epoch": 3.7791436181110427, |
| "grad_norm": 0.11554119750738572, |
| "learning_rate": 1.3591653000323764e-05, |
| "loss": 0.5336, |
| "step": 575 |
| }, |
| { |
| "epoch": 3.785699651710715, |
| "grad_norm": 0.11456186328555522, |
| "learning_rate": 1.3453944055935151e-05, |
| "loss": 0.5447, |
| "step": 576 |
| }, |
| { |
| "epoch": 3.792255685310387, |
| "grad_norm": 0.12119615336966778, |
| "learning_rate": 1.3316795109943476e-05, |
| "loss": 0.5333, |
| "step": 577 |
| }, |
| { |
| "epoch": 3.798811718910059, |
| "grad_norm": 0.1097823442621975, |
| "learning_rate": 1.3180209055554043e-05, |
| "loss": 0.5353, |
| "step": 578 |
| }, |
| { |
| "epoch": 3.8053677525097314, |
| "grad_norm": 0.12113771390183198, |
| "learning_rate": 1.3044188774097757e-05, |
| "loss": 0.5406, |
| "step": 579 |
| }, |
| { |
| "epoch": 3.811923786109404, |
| "grad_norm": 0.1076103282360743, |
| "learning_rate": 1.2908737134970367e-05, |
| "loss": 0.5428, |
| "step": 580 |
| }, |
| { |
| "epoch": 3.818479819709076, |
| "grad_norm": 0.10778481828407349, |
| "learning_rate": 1.2773856995571858e-05, |
| "loss": 0.5346, |
| "step": 581 |
| }, |
| { |
| "epoch": 3.8250358533087483, |
| "grad_norm": 0.10605992762588787, |
| "learning_rate": 1.2639551201246278e-05, |
| "loss": 0.5394, |
| "step": 582 |
| }, |
| { |
| "epoch": 3.8315918869084205, |
| "grad_norm": 0.10492552888703595, |
| "learning_rate": 1.2505822585221665e-05, |
| "loss": 0.5402, |
| "step": 583 |
| }, |
| { |
| "epoch": 3.8381479205080926, |
| "grad_norm": 0.10835031468844004, |
| "learning_rate": 1.2372673968550229e-05, |
| "loss": 0.5449, |
| "step": 584 |
| }, |
| { |
| "epoch": 3.8447039541077648, |
| "grad_norm": 0.10610757084756944, |
| "learning_rate": 1.2240108160048934e-05, |
| "loss": 0.5366, |
| "step": 585 |
| }, |
| { |
| "epoch": 3.851259987707437, |
| "grad_norm": 0.10287232050693294, |
| "learning_rate": 1.2108127956240186e-05, |
| "loss": 0.5367, |
| "step": 586 |
| }, |
| { |
| "epoch": 3.857816021307109, |
| "grad_norm": 0.10929614703844563, |
| "learning_rate": 1.1976736141292853e-05, |
| "loss": 0.5335, |
| "step": 587 |
| }, |
| { |
| "epoch": 3.8643720549067813, |
| "grad_norm": 0.1204731707453792, |
| "learning_rate": 1.1845935486963546e-05, |
| "loss": 0.5404, |
| "step": 588 |
| }, |
| { |
| "epoch": 3.870928088506454, |
| "grad_norm": 0.10037277898866681, |
| "learning_rate": 1.1715728752538103e-05, |
| "loss": 0.5386, |
| "step": 589 |
| }, |
| { |
| "epoch": 3.877484122106126, |
| "grad_norm": 0.10755482937489974, |
| "learning_rate": 1.158611868477344e-05, |
| "loss": 0.5396, |
| "step": 590 |
| }, |
| { |
| "epoch": 3.884040155705798, |
| "grad_norm": 0.1125655323891997, |
| "learning_rate": 1.1457108017839587e-05, |
| "loss": 0.5522, |
| "step": 591 |
| }, |
| { |
| "epoch": 3.8905961893054704, |
| "grad_norm": 0.09908512763395612, |
| "learning_rate": 1.1328699473261957e-05, |
| "loss": 0.5389, |
| "step": 592 |
| }, |
| { |
| "epoch": 3.8971522229051425, |
| "grad_norm": 0.10011512841325966, |
| "learning_rate": 1.1200895759864027e-05, |
| "loss": 0.5402, |
| "step": 593 |
| }, |
| { |
| "epoch": 3.9037082565048147, |
| "grad_norm": 0.11190228493508152, |
| "learning_rate": 1.107369957371013e-05, |
| "loss": 0.5402, |
| "step": 594 |
| }, |
| { |
| "epoch": 3.910264290104487, |
| "grad_norm": 0.0997294109524961, |
| "learning_rate": 1.09471135980486e-05, |
| "loss": 0.5356, |
| "step": 595 |
| }, |
| { |
| "epoch": 3.916820323704159, |
| "grad_norm": 0.10926000465284759, |
| "learning_rate": 1.0821140503255174e-05, |
| "loss": 0.5459, |
| "step": 596 |
| }, |
| { |
| "epoch": 3.923376357303831, |
| "grad_norm": 0.10485038006440338, |
| "learning_rate": 1.0695782946776619e-05, |
| "loss": 0.5407, |
| "step": 597 |
| }, |
| { |
| "epoch": 3.9299323909035033, |
| "grad_norm": 0.10291841117205842, |
| "learning_rate": 1.0571043573074737e-05, |
| "loss": 0.5373, |
| "step": 598 |
| }, |
| { |
| "epoch": 3.9364884245031755, |
| "grad_norm": 0.10153144309813497, |
| "learning_rate": 1.0446925013570545e-05, |
| "loss": 0.5408, |
| "step": 599 |
| }, |
| { |
| "epoch": 3.9430444581028476, |
| "grad_norm": 0.10605125481156527, |
| "learning_rate": 1.0323429886588743e-05, |
| "loss": 0.5411, |
| "step": 600 |
| }, |
| { |
| "epoch": 3.94960049170252, |
| "grad_norm": 0.10215451404049354, |
| "learning_rate": 1.020056079730252e-05, |
| "loss": 0.5428, |
| "step": 601 |
| }, |
| { |
| "epoch": 3.956156525302192, |
| "grad_norm": 0.10818454985255951, |
| "learning_rate": 1.0078320337678584e-05, |
| "loss": 0.5396, |
| "step": 602 |
| }, |
| { |
| "epoch": 3.962712558901864, |
| "grad_norm": 0.10188342720575627, |
| "learning_rate": 9.956711086422471e-06, |
| "loss": 0.5486, |
| "step": 603 |
| }, |
| { |
| "epoch": 3.9692685925015363, |
| "grad_norm": 0.10148315246310981, |
| "learning_rate": 9.835735608924155e-06, |
| "loss": 0.5341, |
| "step": 604 |
| }, |
| { |
| "epoch": 3.975824626101209, |
| "grad_norm": 0.09241997326882184, |
| "learning_rate": 9.715396457203918e-06, |
| "loss": 0.5389, |
| "step": 605 |
| }, |
| { |
| "epoch": 3.982380659700881, |
| "grad_norm": 0.102047541160635, |
| "learning_rate": 9.595696169858542e-06, |
| "loss": 0.5343, |
| "step": 606 |
| }, |
| { |
| "epoch": 3.9889366933005532, |
| "grad_norm": 0.16329932937909056, |
| "learning_rate": 9.476637272007748e-06, |
| "loss": 0.5414, |
| "step": 607 |
| }, |
| { |
| "epoch": 3.9954927269002254, |
| "grad_norm": 0.10040695165913859, |
| "learning_rate": 9.358222275240884e-06, |
| "loss": 0.5365, |
| "step": 608 |
| }, |
| { |
| "epoch": 4.006556033599672, |
| "grad_norm": 0.16186764492729486, |
| "learning_rate": 9.24045367756401e-06, |
| "loss": 0.5171, |
| "step": 609 |
| }, |
| { |
| "epoch": 4.013112067199344, |
| "grad_norm": 0.1322033843585489, |
| "learning_rate": 9.123333963347166e-06, |
| "loss": 0.5209, |
| "step": 610 |
| }, |
| { |
| "epoch": 4.0196681007990165, |
| "grad_norm": 0.12245478327966862, |
| "learning_rate": 9.006865603271952e-06, |
| "loss": 0.5112, |
| "step": 611 |
| }, |
| { |
| "epoch": 4.026224134398689, |
| "grad_norm": 0.117065406177187, |
| "learning_rate": 8.89105105427945e-06, |
| "loss": 0.5089, |
| "step": 612 |
| }, |
| { |
| "epoch": 4.032780167998361, |
| "grad_norm": 0.1358673718015058, |
| "learning_rate": 8.775892759518321e-06, |
| "loss": 0.5245, |
| "step": 613 |
| }, |
| { |
| "epoch": 4.039336201598033, |
| "grad_norm": 0.1436375868810231, |
| "learning_rate": 8.661393148293355e-06, |
| "loss": 0.523, |
| "step": 614 |
| }, |
| { |
| "epoch": 4.045892235197705, |
| "grad_norm": 0.11924450152050672, |
| "learning_rate": 8.547554636014177e-06, |
| "loss": 0.506, |
| "step": 615 |
| }, |
| { |
| "epoch": 4.052448268797377, |
| "grad_norm": 0.11988650232489285, |
| "learning_rate": 8.434379624144261e-06, |
| "loss": 0.515, |
| "step": 616 |
| }, |
| { |
| "epoch": 4.0590043023970495, |
| "grad_norm": 0.13738230782732197, |
| "learning_rate": 8.321870500150347e-06, |
| "loss": 0.511, |
| "step": 617 |
| }, |
| { |
| "epoch": 4.065560335996722, |
| "grad_norm": 0.1371441628587204, |
| "learning_rate": 8.210029637452016e-06, |
| "loss": 0.5139, |
| "step": 618 |
| }, |
| { |
| "epoch": 4.072116369596394, |
| "grad_norm": 0.3628915461879963, |
| "learning_rate": 8.098859395371641e-06, |
| "loss": 0.5165, |
| "step": 619 |
| }, |
| { |
| "epoch": 4.078672403196066, |
| "grad_norm": 0.11429063516858927, |
| "learning_rate": 7.988362119084642e-06, |
| "loss": 0.5136, |
| "step": 620 |
| }, |
| { |
| "epoch": 4.085228436795739, |
| "grad_norm": 0.13045931327904303, |
| "learning_rate": 7.87854013956994e-06, |
| "loss": 0.5097, |
| "step": 621 |
| }, |
| { |
| "epoch": 4.091784470395411, |
| "grad_norm": 0.11629419650039642, |
| "learning_rate": 7.769395773560874e-06, |
| "loss": 0.5157, |
| "step": 622 |
| }, |
| { |
| "epoch": 4.098340503995083, |
| "grad_norm": 0.11362363257154101, |
| "learning_rate": 7.660931323496283e-06, |
| "loss": 0.5188, |
| "step": 623 |
| }, |
| { |
| "epoch": 4.1048965375947555, |
| "grad_norm": 0.10992983384699095, |
| "learning_rate": 7.553149077471915e-06, |
| "loss": 0.5124, |
| "step": 624 |
| }, |
| { |
| "epoch": 4.111452571194428, |
| "grad_norm": 0.10426021171852758, |
| "learning_rate": 7.446051309192204e-06, |
| "loss": 0.5142, |
| "step": 625 |
| }, |
| { |
| "epoch": 4.1180086047941, |
| "grad_norm": 0.10863479923168164, |
| "learning_rate": 7.3396402779222845e-06, |
| "loss": 0.5127, |
| "step": 626 |
| }, |
| { |
| "epoch": 4.124564638393772, |
| "grad_norm": 0.10635132221465905, |
| "learning_rate": 7.233918228440324e-06, |
| "loss": 0.5162, |
| "step": 627 |
| }, |
| { |
| "epoch": 4.131120671993444, |
| "grad_norm": 0.09912102469646186, |
| "learning_rate": 7.128887390990198e-06, |
| "loss": 0.5224, |
| "step": 628 |
| }, |
| { |
| "epoch": 4.137676705593116, |
| "grad_norm": 0.10290328673729038, |
| "learning_rate": 7.024549981234377e-06, |
| "loss": 0.5217, |
| "step": 629 |
| }, |
| { |
| "epoch": 4.1442327391927885, |
| "grad_norm": 0.1038941435526061, |
| "learning_rate": 6.9209082002072725e-06, |
| "loss": 0.5133, |
| "step": 630 |
| }, |
| { |
| "epoch": 4.150788772792461, |
| "grad_norm": 0.09603936143418888, |
| "learning_rate": 6.817964234268748e-06, |
| "loss": 0.5176, |
| "step": 631 |
| }, |
| { |
| "epoch": 4.157344806392133, |
| "grad_norm": 0.11813529879188972, |
| "learning_rate": 6.715720255058e-06, |
| "loss": 0.5152, |
| "step": 632 |
| }, |
| { |
| "epoch": 4.163900839991805, |
| "grad_norm": 0.09790007822885208, |
| "learning_rate": 6.614178419447781e-06, |
| "loss": 0.5158, |
| "step": 633 |
| }, |
| { |
| "epoch": 4.170456873591477, |
| "grad_norm": 0.10792344691656378, |
| "learning_rate": 6.513340869498859e-06, |
| "loss": 0.5122, |
| "step": 634 |
| }, |
| { |
| "epoch": 4.177012907191149, |
| "grad_norm": 0.24107917853631503, |
| "learning_rate": 6.4132097324148556e-06, |
| "loss": 0.5161, |
| "step": 635 |
| }, |
| { |
| "epoch": 4.183568940790821, |
| "grad_norm": 0.09669004690900093, |
| "learning_rate": 6.313787120497376e-06, |
| "loss": 0.507, |
| "step": 636 |
| }, |
| { |
| "epoch": 4.190124974390494, |
| "grad_norm": 0.10731913975620888, |
| "learning_rate": 6.215075131101405e-06, |
| "loss": 0.5087, |
| "step": 637 |
| }, |
| { |
| "epoch": 4.196681007990166, |
| "grad_norm": 0.0997103006367049, |
| "learning_rate": 6.117075846591123e-06, |
| "loss": 0.5201, |
| "step": 638 |
| }, |
| { |
| "epoch": 4.203237041589838, |
| "grad_norm": 0.09955705599544924, |
| "learning_rate": 6.019791334295955e-06, |
| "loss": 0.5076, |
| "step": 639 |
| }, |
| { |
| "epoch": 4.20979307518951, |
| "grad_norm": 0.09540357195970695, |
| "learning_rate": 5.923223646466923e-06, |
| "loss": 0.5181, |
| "step": 640 |
| }, |
| { |
| "epoch": 4.216349108789182, |
| "grad_norm": 0.0952939694361504, |
| "learning_rate": 5.827374820233407e-06, |
| "loss": 0.5195, |
| "step": 641 |
| }, |
| { |
| "epoch": 4.222905142388854, |
| "grad_norm": 0.09942874484427339, |
| "learning_rate": 5.732246877560146e-06, |
| "loss": 0.5171, |
| "step": 642 |
| }, |
| { |
| "epoch": 4.2294611759885266, |
| "grad_norm": 0.09564484595006637, |
| "learning_rate": 5.637841825204588e-06, |
| "loss": 0.5131, |
| "step": 643 |
| }, |
| { |
| "epoch": 4.236017209588199, |
| "grad_norm": 0.1056325087529548, |
| "learning_rate": 5.5441616546745646e-06, |
| "loss": 0.5095, |
| "step": 644 |
| }, |
| { |
| "epoch": 4.242573243187872, |
| "grad_norm": 0.09442156372590707, |
| "learning_rate": 5.451208342186229e-06, |
| "loss": 0.5139, |
| "step": 645 |
| }, |
| { |
| "epoch": 4.249129276787544, |
| "grad_norm": 0.08963880451358089, |
| "learning_rate": 5.358983848622452e-06, |
| "loss": 0.5192, |
| "step": 646 |
| }, |
| { |
| "epoch": 4.255685310387216, |
| "grad_norm": 0.0917773531851403, |
| "learning_rate": 5.26749011949141e-06, |
| "loss": 0.5112, |
| "step": 647 |
| }, |
| { |
| "epoch": 4.262241343986888, |
| "grad_norm": 0.09650192405895454, |
| "learning_rate": 5.176729084885508e-06, |
| "loss": 0.5156, |
| "step": 648 |
| }, |
| { |
| "epoch": 4.26879737758656, |
| "grad_norm": 0.09638202151344397, |
| "learning_rate": 5.086702659440743e-06, |
| "loss": 0.5132, |
| "step": 649 |
| }, |
| { |
| "epoch": 4.275353411186233, |
| "grad_norm": 0.09515733205388084, |
| "learning_rate": 4.99741274229625e-06, |
| "loss": 0.5107, |
| "step": 650 |
| }, |
| { |
| "epoch": 4.281909444785905, |
| "grad_norm": 0.08694276833101254, |
| "learning_rate": 4.908861217054281e-06, |
| "loss": 0.511, |
| "step": 651 |
| }, |
| { |
| "epoch": 4.288465478385577, |
| "grad_norm": 0.09365787216266765, |
| "learning_rate": 4.821049951740442e-06, |
| "loss": 0.5196, |
| "step": 652 |
| }, |
| { |
| "epoch": 4.295021511985249, |
| "grad_norm": 0.09875335551778978, |
| "learning_rate": 4.733980798764273e-06, |
| "loss": 0.5139, |
| "step": 653 |
| }, |
| { |
| "epoch": 4.301577545584921, |
| "grad_norm": 0.0984165654822695, |
| "learning_rate": 4.647655594880225e-06, |
| "loss": 0.5194, |
| "step": 654 |
| }, |
| { |
| "epoch": 4.308133579184593, |
| "grad_norm": 0.08981515555040814, |
| "learning_rate": 4.562076161148881e-06, |
| "loss": 0.5159, |
| "step": 655 |
| }, |
| { |
| "epoch": 4.3146896127842655, |
| "grad_norm": 0.08699993576971486, |
| "learning_rate": 4.4772443028985004e-06, |
| "loss": 0.5107, |
| "step": 656 |
| }, |
| { |
| "epoch": 4.321245646383938, |
| "grad_norm": 0.09531375447276737, |
| "learning_rate": 4.393161809687021e-06, |
| "loss": 0.5211, |
| "step": 657 |
| }, |
| { |
| "epoch": 4.32780167998361, |
| "grad_norm": 0.16067767992274057, |
| "learning_rate": 4.3098304552642385e-06, |
| "loss": 0.5163, |
| "step": 658 |
| }, |
| { |
| "epoch": 4.334357713583282, |
| "grad_norm": 0.09069517004519909, |
| "learning_rate": 4.227251997534416e-06, |
| "loss": 0.5107, |
| "step": 659 |
| }, |
| { |
| "epoch": 4.340913747182954, |
| "grad_norm": 0.08688482436943581, |
| "learning_rate": 4.1454281785191995e-06, |
| "loss": 0.5118, |
| "step": 660 |
| }, |
| { |
| "epoch": 4.347469780782626, |
| "grad_norm": 0.09109521686027529, |
| "learning_rate": 4.064360724320846e-06, |
| "loss": 0.5114, |
| "step": 661 |
| }, |
| { |
| "epoch": 4.3540258143822985, |
| "grad_norm": 0.0885789818679224, |
| "learning_rate": 3.984051345085855e-06, |
| "loss": 0.5126, |
| "step": 662 |
| }, |
| { |
| "epoch": 4.360581847981971, |
| "grad_norm": 0.09379559333937212, |
| "learning_rate": 3.90450173496887e-06, |
| "loss": 0.5207, |
| "step": 663 |
| }, |
| { |
| "epoch": 4.367137881581643, |
| "grad_norm": 0.08937679173054215, |
| "learning_rate": 3.825713572096903e-06, |
| "loss": 0.5175, |
| "step": 664 |
| }, |
| { |
| "epoch": 4.373693915181315, |
| "grad_norm": 0.08924834510849128, |
| "learning_rate": 3.747688518534003e-06, |
| "loss": 0.5186, |
| "step": 665 |
| }, |
| { |
| "epoch": 4.380249948780987, |
| "grad_norm": 0.09109376121051856, |
| "learning_rate": 3.6704282202461515e-06, |
| "loss": 0.5161, |
| "step": 666 |
| }, |
| { |
| "epoch": 4.386805982380659, |
| "grad_norm": 0.10119588536048828, |
| "learning_rate": 3.5939343070665243e-06, |
| "loss": 0.5032, |
| "step": 667 |
| }, |
| { |
| "epoch": 4.3933620159803315, |
| "grad_norm": 0.09644206288351137, |
| "learning_rate": 3.518208392661184e-06, |
| "loss": 0.5188, |
| "step": 668 |
| }, |
| { |
| "epoch": 4.399918049580004, |
| "grad_norm": 0.08778658376932175, |
| "learning_rate": 3.4432520744949317e-06, |
| "loss": 0.5133, |
| "step": 669 |
| }, |
| { |
| "epoch": 4.406474083179677, |
| "grad_norm": 0.0941601011641409, |
| "learning_rate": 3.3690669337977e-06, |
| "loss": 0.5061, |
| "step": 670 |
| }, |
| { |
| "epoch": 4.413030116779349, |
| "grad_norm": 0.09791658312405495, |
| "learning_rate": 3.295654535531161e-06, |
| "loss": 0.5193, |
| "step": 671 |
| }, |
| { |
| "epoch": 4.419586150379021, |
| "grad_norm": 0.09354050139551012, |
| "learning_rate": 3.2230164283556918e-06, |
| "loss": 0.5172, |
| "step": 672 |
| }, |
| { |
| "epoch": 4.426142183978693, |
| "grad_norm": 0.08893737381738466, |
| "learning_rate": 3.151154144597741e-06, |
| "loss": 0.5111, |
| "step": 673 |
| }, |
| { |
| "epoch": 4.432698217578365, |
| "grad_norm": 0.08445332878405215, |
| "learning_rate": 3.080069200217497e-06, |
| "loss": 0.5097, |
| "step": 674 |
| }, |
| { |
| "epoch": 4.4392542511780375, |
| "grad_norm": 0.08801285298434917, |
| "learning_rate": 3.0097630947768695e-06, |
| "loss": 0.5135, |
| "step": 675 |
| }, |
| { |
| "epoch": 4.44581028477771, |
| "grad_norm": 0.09402581699959618, |
| "learning_rate": 2.9402373114079295e-06, |
| "loss": 0.5086, |
| "step": 676 |
| }, |
| { |
| "epoch": 4.452366318377382, |
| "grad_norm": 0.0905312100821144, |
| "learning_rate": 2.871493316781546e-06, |
| "loss": 0.515, |
| "step": 677 |
| }, |
| { |
| "epoch": 4.458922351977054, |
| "grad_norm": 0.09633255689013458, |
| "learning_rate": 2.803532561076492e-06, |
| "loss": 0.5123, |
| "step": 678 |
| }, |
| { |
| "epoch": 4.465478385576726, |
| "grad_norm": 0.08803341472121784, |
| "learning_rate": 2.7363564779488448e-06, |
| "loss": 0.5135, |
| "step": 679 |
| }, |
| { |
| "epoch": 4.472034419176398, |
| "grad_norm": 0.08797758072438489, |
| "learning_rate": 2.669966484501716e-06, |
| "loss": 0.5199, |
| "step": 680 |
| }, |
| { |
| "epoch": 4.4785904527760705, |
| "grad_norm": 0.09112213281433264, |
| "learning_rate": 2.6043639812554043e-06, |
| "loss": 0.5205, |
| "step": 681 |
| }, |
| { |
| "epoch": 4.485146486375743, |
| "grad_norm": 0.09544750033107532, |
| "learning_rate": 2.5395503521178143e-06, |
| "loss": 0.5253, |
| "step": 682 |
| }, |
| { |
| "epoch": 4.491702519975415, |
| "grad_norm": 0.09376536821854904, |
| "learning_rate": 2.4755269643552594e-06, |
| "loss": 0.5164, |
| "step": 683 |
| }, |
| { |
| "epoch": 4.498258553575087, |
| "grad_norm": 0.09454989320113857, |
| "learning_rate": 2.4122951685636674e-06, |
| "loss": 0.5187, |
| "step": 684 |
| }, |
| { |
| "epoch": 4.504814587174759, |
| "grad_norm": 0.08421901449177807, |
| "learning_rate": 2.3498562986400187e-06, |
| "loss": 0.5154, |
| "step": 685 |
| }, |
| { |
| "epoch": 4.511370620774431, |
| "grad_norm": 0.08753626044860159, |
| "learning_rate": 2.2882116717542634e-06, |
| "loss": 0.5176, |
| "step": 686 |
| }, |
| { |
| "epoch": 4.5179266543741035, |
| "grad_norm": 0.0907417635028892, |
| "learning_rate": 2.22736258832152e-06, |
| "loss": 0.5244, |
| "step": 687 |
| }, |
| { |
| "epoch": 4.524482687973776, |
| "grad_norm": 0.08866763487206057, |
| "learning_rate": 2.1673103319746146e-06, |
| "loss": 0.5188, |
| "step": 688 |
| }, |
| { |
| "epoch": 4.531038721573448, |
| "grad_norm": 0.08406194293605605, |
| "learning_rate": 2.1080561695370425e-06, |
| "loss": 0.5117, |
| "step": 689 |
| }, |
| { |
| "epoch": 4.53759475517312, |
| "grad_norm": 0.08781306185528281, |
| "learning_rate": 2.049601350996233e-06, |
| "loss": 0.5105, |
| "step": 690 |
| }, |
| { |
| "epoch": 4.544150788772792, |
| "grad_norm": 0.0862714841925732, |
| "learning_rate": 1.9919471094771523e-06, |
| "loss": 0.513, |
| "step": 691 |
| }, |
| { |
| "epoch": 4.550706822372465, |
| "grad_norm": 0.08316320143150852, |
| "learning_rate": 1.93509466121633e-06, |
| "loss": 0.5143, |
| "step": 692 |
| }, |
| { |
| "epoch": 4.557262855972137, |
| "grad_norm": 0.08459303029799788, |
| "learning_rate": 1.8790452055361764e-06, |
| "loss": 0.5117, |
| "step": 693 |
| }, |
| { |
| "epoch": 4.5638188895718095, |
| "grad_norm": 0.08646962847262735, |
| "learning_rate": 1.8237999248197002e-06, |
| "loss": 0.5122, |
| "step": 694 |
| }, |
| { |
| "epoch": 4.570374923171482, |
| "grad_norm": 0.18896054046080968, |
| "learning_rate": 1.7693599844855568e-06, |
| "loss": 0.5187, |
| "step": 695 |
| }, |
| { |
| "epoch": 4.576930956771154, |
| "grad_norm": 0.08583158851506041, |
| "learning_rate": 1.7157265329634354e-06, |
| "loss": 0.5206, |
| "step": 696 |
| }, |
| { |
| "epoch": 4.583486990370826, |
| "grad_norm": 0.0823819160891982, |
| "learning_rate": 1.6629007016698918e-06, |
| "loss": 0.5119, |
| "step": 697 |
| }, |
| { |
| "epoch": 4.590043023970498, |
| "grad_norm": 0.0856122092163256, |
| "learning_rate": 1.6108836049844434e-06, |
| "loss": 0.5192, |
| "step": 698 |
| }, |
| { |
| "epoch": 4.59659905757017, |
| "grad_norm": 0.08366071868253996, |
| "learning_rate": 1.5596763402260462e-06, |
| "loss": 0.5169, |
| "step": 699 |
| }, |
| { |
| "epoch": 4.6031550911698424, |
| "grad_norm": 0.08756847365451516, |
| "learning_rate": 1.5092799876299835e-06, |
| "loss": 0.5142, |
| "step": 700 |
| }, |
| { |
| "epoch": 4.609711124769515, |
| "grad_norm": 0.08472854571793725, |
| "learning_rate": 1.459695610325067e-06, |
| "loss": 0.5191, |
| "step": 701 |
| }, |
| { |
| "epoch": 4.616267158369187, |
| "grad_norm": 0.08619093020176126, |
| "learning_rate": 1.4109242543111834e-06, |
| "loss": 0.5166, |
| "step": 702 |
| }, |
| { |
| "epoch": 4.622823191968859, |
| "grad_norm": 0.08301363626424127, |
| "learning_rate": 1.3629669484372722e-06, |
| "loss": 0.517, |
| "step": 703 |
| }, |
| { |
| "epoch": 4.629379225568531, |
| "grad_norm": 0.08230801802672034, |
| "learning_rate": 1.3158247043795735e-06, |
| "loss": 0.5033, |
| "step": 704 |
| }, |
| { |
| "epoch": 4.635935259168203, |
| "grad_norm": 0.08247077823518248, |
| "learning_rate": 1.2694985166203311e-06, |
| "loss": 0.523, |
| "step": 705 |
| }, |
| { |
| "epoch": 4.642491292767875, |
| "grad_norm": 0.08077290324213217, |
| "learning_rate": 1.2239893624267852e-06, |
| "loss": 0.5218, |
| "step": 706 |
| }, |
| { |
| "epoch": 4.649047326367548, |
| "grad_norm": 0.08354463750730508, |
| "learning_rate": 1.1792982018305677e-06, |
| "loss": 0.5143, |
| "step": 707 |
| }, |
| { |
| "epoch": 4.65560335996722, |
| "grad_norm": 0.08393485609493388, |
| "learning_rate": 1.1354259776074472e-06, |
| "loss": 0.5151, |
| "step": 708 |
| }, |
| { |
| "epoch": 4.662159393566892, |
| "grad_norm": 0.0832402976978181, |
| "learning_rate": 1.0923736152574428e-06, |
| "loss": 0.5233, |
| "step": 709 |
| }, |
| { |
| "epoch": 4.668715427166564, |
| "grad_norm": 0.0835585475483202, |
| "learning_rate": 1.050142022985292e-06, |
| "loss": 0.5093, |
| "step": 710 |
| }, |
| { |
| "epoch": 4.675271460766236, |
| "grad_norm": 0.08176206607867041, |
| "learning_rate": 1.0087320916813127e-06, |
| "loss": 0.5225, |
| "step": 711 |
| }, |
| { |
| "epoch": 4.681827494365908, |
| "grad_norm": 0.08752232597844366, |
| "learning_rate": 9.681446949025752e-07, |
| "loss": 0.5156, |
| "step": 712 |
| }, |
| { |
| "epoch": 4.6883835279655806, |
| "grad_norm": 0.08108785189147111, |
| "learning_rate": 9.283806888545111e-07, |
| "loss": 0.5138, |
| "step": 713 |
| }, |
| { |
| "epoch": 4.694939561565253, |
| "grad_norm": 0.08223618174467849, |
| "learning_rate": 8.89440912372832e-07, |
| "loss": 0.5172, |
| "step": 714 |
| }, |
| { |
| "epoch": 4.701495595164925, |
| "grad_norm": 0.08351263590200621, |
| "learning_rate": 8.513261869058209e-07, |
| "loss": 0.5282, |
| "step": 715 |
| }, |
| { |
| "epoch": 4.708051628764597, |
| "grad_norm": 0.08286053605490928, |
| "learning_rate": 8.140373164970428e-07, |
| "loss": 0.5244, |
| "step": 716 |
| }, |
| { |
| "epoch": 4.714607662364269, |
| "grad_norm": 0.08518895529322469, |
| "learning_rate": 7.775750877683452e-07, |
| "loss": 0.5171, |
| "step": 717 |
| }, |
| { |
| "epoch": 4.721163695963941, |
| "grad_norm": 0.0813459247213106, |
| "learning_rate": 7.419402699032852e-07, |
| "loss": 0.5186, |
| "step": 718 |
| }, |
| { |
| "epoch": 4.7277197295636135, |
| "grad_norm": 0.0803887295459458, |
| "learning_rate": 7.071336146308883e-07, |
| "loss": 0.5059, |
| "step": 719 |
| }, |
| { |
| "epoch": 4.734275763163287, |
| "grad_norm": 0.08172953576855356, |
| "learning_rate": 6.731558562097995e-07, |
| "loss": 0.5224, |
| "step": 720 |
| }, |
| { |
| "epoch": 4.740831796762959, |
| "grad_norm": 0.08549688805268194, |
| "learning_rate": 6.400077114128023e-07, |
| "loss": 0.5165, |
| "step": 721 |
| }, |
| { |
| "epoch": 4.747387830362631, |
| "grad_norm": 0.08286626015188583, |
| "learning_rate": 6.076898795116792e-07, |
| "loss": 0.51, |
| "step": 722 |
| }, |
| { |
| "epoch": 4.753943863962303, |
| "grad_norm": 0.0823372680890407, |
| "learning_rate": 5.762030422624732e-07, |
| "loss": 0.5212, |
| "step": 723 |
| }, |
| { |
| "epoch": 4.760499897561975, |
| "grad_norm": 0.08161043456784334, |
| "learning_rate": 5.455478638911071e-07, |
| "loss": 0.5116, |
| "step": 724 |
| }, |
| { |
| "epoch": 4.767055931161647, |
| "grad_norm": 0.08090694406998772, |
| "learning_rate": 5.15724991079356e-07, |
| "loss": 0.5223, |
| "step": 725 |
| }, |
| { |
| "epoch": 4.7736119647613195, |
| "grad_norm": 0.08130626566721383, |
| "learning_rate": 4.867350529512261e-07, |
| "loss": 0.5104, |
| "step": 726 |
| }, |
| { |
| "epoch": 4.780167998360992, |
| "grad_norm": 0.08487388342418309, |
| "learning_rate": 4.5857866105966763e-07, |
| "loss": 0.5129, |
| "step": 727 |
| }, |
| { |
| "epoch": 4.786724031960664, |
| "grad_norm": 0.0816069016289385, |
| "learning_rate": 4.3125640937368373e-07, |
| "loss": 0.5123, |
| "step": 728 |
| }, |
| { |
| "epoch": 4.793280065560336, |
| "grad_norm": 0.08049465996969124, |
| "learning_rate": 4.047688742657885e-07, |
| "loss": 0.5063, |
| "step": 729 |
| }, |
| { |
| "epoch": 4.799836099160008, |
| "grad_norm": 0.0830575657951303, |
| "learning_rate": 3.791166144998704e-07, |
| "loss": 0.5121, |
| "step": 730 |
| }, |
| { |
| "epoch": 4.80639213275968, |
| "grad_norm": 0.08387321061054255, |
| "learning_rate": 3.54300171219375e-07, |
| "loss": 0.5146, |
| "step": 731 |
| }, |
| { |
| "epoch": 4.8129481663593525, |
| "grad_norm": 0.08224120804212026, |
| "learning_rate": 3.3032006793590977e-07, |
| "loss": 0.5206, |
| "step": 732 |
| }, |
| { |
| "epoch": 4.819504199959025, |
| "grad_norm": 0.08072502199031102, |
| "learning_rate": 3.0717681051819935e-07, |
| "loss": 0.522, |
| "step": 733 |
| }, |
| { |
| "epoch": 4.826060233558697, |
| "grad_norm": 0.07860877555923813, |
| "learning_rate": 2.848708871814054e-07, |
| "loss": 0.5188, |
| "step": 734 |
| }, |
| { |
| "epoch": 4.832616267158369, |
| "grad_norm": 0.08001404199090854, |
| "learning_rate": 2.634027684768414e-07, |
| "loss": 0.5124, |
| "step": 735 |
| }, |
| { |
| "epoch": 4.839172300758041, |
| "grad_norm": 0.08310618061839851, |
| "learning_rate": 2.4277290728202063e-07, |
| "loss": 0.5212, |
| "step": 736 |
| }, |
| { |
| "epoch": 4.845728334357713, |
| "grad_norm": 0.0797549571951669, |
| "learning_rate": 2.2298173879113481e-07, |
| "loss": 0.5168, |
| "step": 737 |
| }, |
| { |
| "epoch": 4.8522843679573855, |
| "grad_norm": 0.0824361700062477, |
| "learning_rate": 2.040296805058528e-07, |
| "loss": 0.5158, |
| "step": 738 |
| }, |
| { |
| "epoch": 4.858840401557058, |
| "grad_norm": 0.08076093089733337, |
| "learning_rate": 1.859171322265141e-07, |
| "loss": 0.5253, |
| "step": 739 |
| }, |
| { |
| "epoch": 4.86539643515673, |
| "grad_norm": 0.08043719703067892, |
| "learning_rate": 1.6864447604370004e-07, |
| "loss": 0.5244, |
| "step": 740 |
| }, |
| { |
| "epoch": 4.871952468756403, |
| "grad_norm": 0.08277960997697664, |
| "learning_rate": 1.522120763301782e-07, |
| "loss": 0.5085, |
| "step": 741 |
| }, |
| { |
| "epoch": 4.878508502356075, |
| "grad_norm": 0.08180092674268333, |
| "learning_rate": 1.3662027973320614e-07, |
| "loss": 0.5202, |
| "step": 742 |
| }, |
| { |
| "epoch": 4.885064535955747, |
| "grad_norm": 0.08183966123917269, |
| "learning_rate": 1.2186941516722173e-07, |
| "loss": 0.5231, |
| "step": 743 |
| }, |
| { |
| "epoch": 4.891620569555419, |
| "grad_norm": 0.07963446358425141, |
| "learning_rate": 1.0795979380690657e-07, |
| "loss": 0.5136, |
| "step": 744 |
| }, |
| { |
| "epoch": 4.8981766031550915, |
| "grad_norm": 0.07852420651092652, |
| "learning_rate": 9.489170908062228e-08, |
| "loss": 0.5204, |
| "step": 745 |
| }, |
| { |
| "epoch": 4.904732636754764, |
| "grad_norm": 0.12877804510272148, |
| "learning_rate": 8.266543666421544e-08, |
| "loss": 0.5185, |
| "step": 746 |
| }, |
| { |
| "epoch": 4.911288670354436, |
| "grad_norm": 0.0869160748356661, |
| "learning_rate": 7.128123447520452e-08, |
| "loss": 0.5234, |
| "step": 747 |
| }, |
| { |
| "epoch": 4.917844703954108, |
| "grad_norm": 0.08086821677688971, |
| "learning_rate": 6.073934266735303e-08, |
| "loss": 0.5129, |
| "step": 748 |
| }, |
| { |
| "epoch": 4.92440073755378, |
| "grad_norm": 0.07880134225937953, |
| "learning_rate": 5.10399836255715e-08, |
| "loss": 0.5064, |
| "step": 749 |
| }, |
| { |
| "epoch": 4.930956771153452, |
| "grad_norm": 0.07797418195960645, |
| "learning_rate": 4.218336196125439e-08, |
| "loss": 0.5106, |
| "step": 750 |
| }, |
| { |
| "epoch": 4.9375128047531245, |
| "grad_norm": 0.08006201262285005, |
| "learning_rate": 3.416966450795922e-08, |
| "loss": 0.5099, |
| "step": 751 |
| }, |
| { |
| "epoch": 4.944068838352797, |
| "grad_norm": 0.08059149745693742, |
| "learning_rate": 2.699906031745414e-08, |
| "loss": 0.5176, |
| "step": 752 |
| }, |
| { |
| "epoch": 4.950624871952469, |
| "grad_norm": 0.08251393526468398, |
| "learning_rate": 2.067170065615187e-08, |
| "loss": 0.5235, |
| "step": 753 |
| }, |
| { |
| "epoch": 4.957180905552141, |
| "grad_norm": 0.08234699138692923, |
| "learning_rate": 1.5187719001943378e-08, |
| "loss": 0.5122, |
| "step": 754 |
| }, |
| { |
| "epoch": 4.963736939151813, |
| "grad_norm": 0.08191620156821956, |
| "learning_rate": 1.0547231041346806e-08, |
| "loss": 0.5237, |
| "step": 755 |
| }, |
| { |
| "epoch": 4.970292972751485, |
| "grad_norm": 0.08084676979927179, |
| "learning_rate": 6.750334667091629e-09, |
| "loss": 0.5178, |
| "step": 756 |
| }, |
| { |
| "epoch": 4.9768490063511575, |
| "grad_norm": 0.08117469896416148, |
| "learning_rate": 3.797109976035884e-09, |
| "loss": 0.5192, |
| "step": 757 |
| }, |
| { |
| "epoch": 4.98340503995083, |
| "grad_norm": 0.0821326486604627, |
| "learning_rate": 1.6876192675052695e-09, |
| "loss": 0.519, |
| "step": 758 |
| }, |
| { |
| "epoch": 4.989961073550502, |
| "grad_norm": 0.08465924625593949, |
| "learning_rate": 4.219070419475557e-10, |
| "loss": 0.5078, |
| "step": 759 |
| }, |
| { |
| "epoch": 4.996517107150174, |
| "grad_norm": 0.08042205253823026, |
| "learning_rate": 0.0, |
| "loss": 0.5193, |
| "step": 760 |
| }, |
| { |
| "epoch": 4.996517107150174, |
| "step": 760, |
| "total_flos": 2.0225694010775175e+19, |
| "train_loss": 0.10313788385767686, |
| "train_runtime": 35636.9099, |
| "train_samples_per_second": 10.956, |
| "train_steps_per_second": 0.021 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 760, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 5, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 2.0225694010775175e+19, |
| "train_batch_size": 1, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|