| { | |
| "best_global_step": null, | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 6.668896321070234, | |
| "eval_steps": 500, | |
| "global_step": 500, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.013377926421404682, | |
| "grad_norm": 0.5197079181671143, | |
| "learning_rate": 0.0, | |
| "loss": 4.2636, | |
| "step": 1 | |
| }, | |
| { | |
| "epoch": 0.026755852842809364, | |
| "grad_norm": 0.5626901984214783, | |
| "learning_rate": 4e-05, | |
| "loss": 4.3971, | |
| "step": 2 | |
| }, | |
| { | |
| "epoch": 0.04013377926421405, | |
| "grad_norm": 0.5167903304100037, | |
| "learning_rate": 8e-05, | |
| "loss": 4.3249, | |
| "step": 3 | |
| }, | |
| { | |
| "epoch": 0.05351170568561873, | |
| "grad_norm": 0.4764951169490814, | |
| "learning_rate": 0.00012, | |
| "loss": 4.2031, | |
| "step": 4 | |
| }, | |
| { | |
| "epoch": 0.06688963210702341, | |
| "grad_norm": 0.45488491654396057, | |
| "learning_rate": 0.00016, | |
| "loss": 4.3914, | |
| "step": 5 | |
| }, | |
| { | |
| "epoch": 0.0802675585284281, | |
| "grad_norm": 0.568274736404419, | |
| "learning_rate": 0.0002, | |
| "loss": 4.2346, | |
| "step": 6 | |
| }, | |
| { | |
| "epoch": 0.09364548494983277, | |
| "grad_norm": 0.5974003076553345, | |
| "learning_rate": 0.0001999555061179088, | |
| "loss": 4.131, | |
| "step": 7 | |
| }, | |
| { | |
| "epoch": 0.10702341137123746, | |
| "grad_norm": 0.6204471588134766, | |
| "learning_rate": 0.00019991101223581757, | |
| "loss": 4.2256, | |
| "step": 8 | |
| }, | |
| { | |
| "epoch": 0.12040133779264214, | |
| "grad_norm": 0.7143808603286743, | |
| "learning_rate": 0.00019986651835372636, | |
| "loss": 3.8449, | |
| "step": 9 | |
| }, | |
| { | |
| "epoch": 0.13377926421404682, | |
| "grad_norm": 0.7799420356750488, | |
| "learning_rate": 0.00019982202447163517, | |
| "loss": 4.4301, | |
| "step": 10 | |
| }, | |
| { | |
| "epoch": 0.14715719063545152, | |
| "grad_norm": 0.8880407214164734, | |
| "learning_rate": 0.00019977753058954395, | |
| "loss": 4.2266, | |
| "step": 11 | |
| }, | |
| { | |
| "epoch": 0.1605351170568562, | |
| "grad_norm": 0.7776209712028503, | |
| "learning_rate": 0.00019973303670745273, | |
| "loss": 4.3208, | |
| "step": 12 | |
| }, | |
| { | |
| "epoch": 0.17391304347826086, | |
| "grad_norm": 0.9125858545303345, | |
| "learning_rate": 0.0001996885428253615, | |
| "loss": 4.4363, | |
| "step": 13 | |
| }, | |
| { | |
| "epoch": 0.18729096989966554, | |
| "grad_norm": 0.9000256657600403, | |
| "learning_rate": 0.00019964404894327032, | |
| "loss": 4.2917, | |
| "step": 14 | |
| }, | |
| { | |
| "epoch": 0.20066889632107024, | |
| "grad_norm": 0.9995108246803284, | |
| "learning_rate": 0.00019959955506117908, | |
| "loss": 4.1784, | |
| "step": 15 | |
| }, | |
| { | |
| "epoch": 0.2140468227424749, | |
| "grad_norm": 0.9209024310112, | |
| "learning_rate": 0.0001995550611790879, | |
| "loss": 4.7852, | |
| "step": 16 | |
| }, | |
| { | |
| "epoch": 0.22742474916387959, | |
| "grad_norm": 0.9421981573104858, | |
| "learning_rate": 0.00019951056729699667, | |
| "loss": 4.8501, | |
| "step": 17 | |
| }, | |
| { | |
| "epoch": 0.2408026755852843, | |
| "grad_norm": 0.9213201403617859, | |
| "learning_rate": 0.00019946607341490545, | |
| "loss": 4.7923, | |
| "step": 18 | |
| }, | |
| { | |
| "epoch": 0.25418060200668896, | |
| "grad_norm": 0.9378194212913513, | |
| "learning_rate": 0.00019942157953281423, | |
| "loss": 4.9593, | |
| "step": 19 | |
| }, | |
| { | |
| "epoch": 0.26755852842809363, | |
| "grad_norm": 1.0096492767333984, | |
| "learning_rate": 0.00019937708565072304, | |
| "loss": 4.7099, | |
| "step": 20 | |
| }, | |
| { | |
| "epoch": 0.2809364548494983, | |
| "grad_norm": 0.8903587460517883, | |
| "learning_rate": 0.00019933259176863183, | |
| "loss": 4.3746, | |
| "step": 21 | |
| }, | |
| { | |
| "epoch": 0.29431438127090304, | |
| "grad_norm": 0.7808490991592407, | |
| "learning_rate": 0.0001992880978865406, | |
| "loss": 4.5873, | |
| "step": 22 | |
| }, | |
| { | |
| "epoch": 0.3076923076923077, | |
| "grad_norm": 0.8145670294761658, | |
| "learning_rate": 0.0001992436040044494, | |
| "loss": 4.7924, | |
| "step": 23 | |
| }, | |
| { | |
| "epoch": 0.3210702341137124, | |
| "grad_norm": 0.7945849299430847, | |
| "learning_rate": 0.0001991991101223582, | |
| "loss": 4.8881, | |
| "step": 24 | |
| }, | |
| { | |
| "epoch": 0.33444816053511706, | |
| "grad_norm": 0.7871395349502563, | |
| "learning_rate": 0.00019915461624026696, | |
| "loss": 4.6922, | |
| "step": 25 | |
| }, | |
| { | |
| "epoch": 0.34782608695652173, | |
| "grad_norm": 0.9111238718032837, | |
| "learning_rate": 0.00019911012235817577, | |
| "loss": 4.9982, | |
| "step": 26 | |
| }, | |
| { | |
| "epoch": 0.3612040133779264, | |
| "grad_norm": 0.7121369242668152, | |
| "learning_rate": 0.00019906562847608455, | |
| "loss": 4.4756, | |
| "step": 27 | |
| }, | |
| { | |
| "epoch": 0.3745819397993311, | |
| "grad_norm": 0.7118422389030457, | |
| "learning_rate": 0.00019902113459399333, | |
| "loss": 5.1389, | |
| "step": 28 | |
| }, | |
| { | |
| "epoch": 0.3879598662207358, | |
| "grad_norm": 0.7100292444229126, | |
| "learning_rate": 0.0001989766407119021, | |
| "loss": 4.7691, | |
| "step": 29 | |
| }, | |
| { | |
| "epoch": 0.4013377926421405, | |
| "grad_norm": 0.708591639995575, | |
| "learning_rate": 0.00019893214682981092, | |
| "loss": 4.8721, | |
| "step": 30 | |
| }, | |
| { | |
| "epoch": 0.41471571906354515, | |
| "grad_norm": 0.6711616516113281, | |
| "learning_rate": 0.0001988876529477197, | |
| "loss": 4.9152, | |
| "step": 31 | |
| }, | |
| { | |
| "epoch": 0.4280936454849498, | |
| "grad_norm": 0.7158232927322388, | |
| "learning_rate": 0.0001988431590656285, | |
| "loss": 4.828, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.4414715719063545, | |
| "grad_norm": 0.6246087551116943, | |
| "learning_rate": 0.00019879866518353727, | |
| "loss": 4.8452, | |
| "step": 33 | |
| }, | |
| { | |
| "epoch": 0.45484949832775917, | |
| "grad_norm": 0.6088873744010925, | |
| "learning_rate": 0.00019875417130144608, | |
| "loss": 4.9702, | |
| "step": 34 | |
| }, | |
| { | |
| "epoch": 0.4682274247491639, | |
| "grad_norm": 0.5798126459121704, | |
| "learning_rate": 0.00019870967741935483, | |
| "loss": 4.9838, | |
| "step": 35 | |
| }, | |
| { | |
| "epoch": 0.4816053511705686, | |
| "grad_norm": 0.6268919706344604, | |
| "learning_rate": 0.00019866518353726364, | |
| "loss": 4.7636, | |
| "step": 36 | |
| }, | |
| { | |
| "epoch": 0.49498327759197325, | |
| "grad_norm": 0.5649904012680054, | |
| "learning_rate": 0.00019862068965517243, | |
| "loss": 4.506, | |
| "step": 37 | |
| }, | |
| { | |
| "epoch": 0.5083612040133779, | |
| "grad_norm": 0.5947792530059814, | |
| "learning_rate": 0.0001985761957730812, | |
| "loss": 4.8057, | |
| "step": 38 | |
| }, | |
| { | |
| "epoch": 0.5217391304347826, | |
| "grad_norm": 0.6204257011413574, | |
| "learning_rate": 0.00019853170189099, | |
| "loss": 5.0511, | |
| "step": 39 | |
| }, | |
| { | |
| "epoch": 0.5351170568561873, | |
| "grad_norm": 0.5972265601158142, | |
| "learning_rate": 0.0001984872080088988, | |
| "loss": 4.924, | |
| "step": 40 | |
| }, | |
| { | |
| "epoch": 0.5484949832775919, | |
| "grad_norm": 0.6117077469825745, | |
| "learning_rate": 0.00019844271412680758, | |
| "loss": 4.8729, | |
| "step": 41 | |
| }, | |
| { | |
| "epoch": 0.5618729096989966, | |
| "grad_norm": 0.5085508823394775, | |
| "learning_rate": 0.00019839822024471637, | |
| "loss": 4.3616, | |
| "step": 42 | |
| }, | |
| { | |
| "epoch": 0.5752508361204013, | |
| "grad_norm": 0.550647497177124, | |
| "learning_rate": 0.00019835372636262515, | |
| "loss": 5.2512, | |
| "step": 43 | |
| }, | |
| { | |
| "epoch": 0.5886287625418061, | |
| "grad_norm": 0.48329588770866394, | |
| "learning_rate": 0.00019830923248053396, | |
| "loss": 4.9501, | |
| "step": 44 | |
| }, | |
| { | |
| "epoch": 0.6020066889632107, | |
| "grad_norm": 0.6313246488571167, | |
| "learning_rate": 0.0001982647385984427, | |
| "loss": 4.5767, | |
| "step": 45 | |
| }, | |
| { | |
| "epoch": 0.6153846153846154, | |
| "grad_norm": 0.5111928582191467, | |
| "learning_rate": 0.00019822024471635152, | |
| "loss": 4.5586, | |
| "step": 46 | |
| }, | |
| { | |
| "epoch": 0.6287625418060201, | |
| "grad_norm": 0.5264492630958557, | |
| "learning_rate": 0.0001981757508342603, | |
| "loss": 4.7033, | |
| "step": 47 | |
| }, | |
| { | |
| "epoch": 0.6421404682274248, | |
| "grad_norm": 0.5058289170265198, | |
| "learning_rate": 0.0001981312569521691, | |
| "loss": 4.8396, | |
| "step": 48 | |
| }, | |
| { | |
| "epoch": 0.6555183946488294, | |
| "grad_norm": 0.5688439607620239, | |
| "learning_rate": 0.00019808676307007787, | |
| "loss": 5.1887, | |
| "step": 49 | |
| }, | |
| { | |
| "epoch": 0.6688963210702341, | |
| "grad_norm": 0.5488842129707336, | |
| "learning_rate": 0.00019804226918798665, | |
| "loss": 4.6075, | |
| "step": 50 | |
| }, | |
| { | |
| "epoch": 0.6822742474916388, | |
| "grad_norm": 0.5358632206916809, | |
| "learning_rate": 0.00019799777530589546, | |
| "loss": 5.0205, | |
| "step": 51 | |
| }, | |
| { | |
| "epoch": 0.6956521739130435, | |
| "grad_norm": 0.47869494557380676, | |
| "learning_rate": 0.00019795328142380422, | |
| "loss": 4.8495, | |
| "step": 52 | |
| }, | |
| { | |
| "epoch": 0.7090301003344481, | |
| "grad_norm": 0.49378660321235657, | |
| "learning_rate": 0.00019790878754171303, | |
| "loss": 4.6563, | |
| "step": 53 | |
| }, | |
| { | |
| "epoch": 0.7224080267558528, | |
| "grad_norm": 0.5167868733406067, | |
| "learning_rate": 0.0001978642936596218, | |
| "loss": 5.2558, | |
| "step": 54 | |
| }, | |
| { | |
| "epoch": 0.7357859531772575, | |
| "grad_norm": 0.5230040550231934, | |
| "learning_rate": 0.0001978197997775306, | |
| "loss": 4.7769, | |
| "step": 55 | |
| }, | |
| { | |
| "epoch": 0.7491638795986622, | |
| "grad_norm": 0.4822310507297516, | |
| "learning_rate": 0.00019777530589543937, | |
| "loss": 4.9282, | |
| "step": 56 | |
| }, | |
| { | |
| "epoch": 0.7625418060200669, | |
| "grad_norm": 0.500045895576477, | |
| "learning_rate": 0.00019773081201334818, | |
| "loss": 5.0399, | |
| "step": 57 | |
| }, | |
| { | |
| "epoch": 0.7759197324414716, | |
| "grad_norm": 0.4740642309188843, | |
| "learning_rate": 0.00019768631813125696, | |
| "loss": 4.8041, | |
| "step": 58 | |
| }, | |
| { | |
| "epoch": 0.7892976588628763, | |
| "grad_norm": 0.45918184518814087, | |
| "learning_rate": 0.00019764182424916575, | |
| "loss": 4.6304, | |
| "step": 59 | |
| }, | |
| { | |
| "epoch": 0.802675585284281, | |
| "grad_norm": 0.53122878074646, | |
| "learning_rate": 0.00019759733036707453, | |
| "loss": 4.8377, | |
| "step": 60 | |
| }, | |
| { | |
| "epoch": 0.8160535117056856, | |
| "grad_norm": 0.4925791919231415, | |
| "learning_rate": 0.00019755283648498334, | |
| "loss": 5.0919, | |
| "step": 61 | |
| }, | |
| { | |
| "epoch": 0.8294314381270903, | |
| "grad_norm": 0.4777262806892395, | |
| "learning_rate": 0.0001975083426028921, | |
| "loss": 4.8379, | |
| "step": 62 | |
| }, | |
| { | |
| "epoch": 0.842809364548495, | |
| "grad_norm": 0.49119675159454346, | |
| "learning_rate": 0.0001974638487208009, | |
| "loss": 5.0819, | |
| "step": 63 | |
| }, | |
| { | |
| "epoch": 0.8561872909698997, | |
| "grad_norm": 0.4732685089111328, | |
| "learning_rate": 0.00019741935483870969, | |
| "loss": 4.8948, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.8695652173913043, | |
| "grad_norm": 0.46269145607948303, | |
| "learning_rate": 0.00019737486095661847, | |
| "loss": 4.824, | |
| "step": 65 | |
| }, | |
| { | |
| "epoch": 0.882943143812709, | |
| "grad_norm": 0.49532708525657654, | |
| "learning_rate": 0.00019733036707452725, | |
| "loss": 4.8986, | |
| "step": 66 | |
| }, | |
| { | |
| "epoch": 0.8963210702341137, | |
| "grad_norm": 0.5253002643585205, | |
| "learning_rate": 0.00019728587319243606, | |
| "loss": 4.9073, | |
| "step": 67 | |
| }, | |
| { | |
| "epoch": 0.9096989966555183, | |
| "grad_norm": 0.5069419145584106, | |
| "learning_rate": 0.00019724137931034484, | |
| "loss": 4.8962, | |
| "step": 68 | |
| }, | |
| { | |
| "epoch": 0.9230769230769231, | |
| "grad_norm": 0.5038817524909973, | |
| "learning_rate": 0.00019719688542825363, | |
| "loss": 4.8711, | |
| "step": 69 | |
| }, | |
| { | |
| "epoch": 0.9364548494983278, | |
| "grad_norm": 0.4987100064754486, | |
| "learning_rate": 0.0001971523915461624, | |
| "loss": 4.8816, | |
| "step": 70 | |
| }, | |
| { | |
| "epoch": 0.9498327759197325, | |
| "grad_norm": 0.47370976209640503, | |
| "learning_rate": 0.00019710789766407122, | |
| "loss": 4.9675, | |
| "step": 71 | |
| }, | |
| { | |
| "epoch": 0.9632107023411371, | |
| "grad_norm": 0.5081727504730225, | |
| "learning_rate": 0.00019706340378197997, | |
| "loss": 4.2768, | |
| "step": 72 | |
| }, | |
| { | |
| "epoch": 0.9765886287625418, | |
| "grad_norm": 0.45571258664131165, | |
| "learning_rate": 0.00019701890989988878, | |
| "loss": 4.6182, | |
| "step": 73 | |
| }, | |
| { | |
| "epoch": 0.9899665551839465, | |
| "grad_norm": 0.5216127634048462, | |
| "learning_rate": 0.00019697441601779756, | |
| "loss": 4.7126, | |
| "step": 74 | |
| }, | |
| { | |
| "epoch": 1.0, | |
| "grad_norm": 0.5393329858779907, | |
| "learning_rate": 0.00019692992213570635, | |
| "loss": 4.4919, | |
| "step": 75 | |
| }, | |
| { | |
| "epoch": 1.0133779264214047, | |
| "grad_norm": 0.4506986737251282, | |
| "learning_rate": 0.00019688542825361513, | |
| "loss": 4.5089, | |
| "step": 76 | |
| }, | |
| { | |
| "epoch": 1.0267558528428093, | |
| "grad_norm": 0.4328899085521698, | |
| "learning_rate": 0.00019684093437152394, | |
| "loss": 4.7518, | |
| "step": 77 | |
| }, | |
| { | |
| "epoch": 1.040133779264214, | |
| "grad_norm": 0.4397362470626831, | |
| "learning_rate": 0.00019679644048943272, | |
| "loss": 4.5069, | |
| "step": 78 | |
| }, | |
| { | |
| "epoch": 1.0535117056856187, | |
| "grad_norm": 0.4604664146900177, | |
| "learning_rate": 0.0001967519466073415, | |
| "loss": 4.7054, | |
| "step": 79 | |
| }, | |
| { | |
| "epoch": 1.0668896321070234, | |
| "grad_norm": 0.4398234784603119, | |
| "learning_rate": 0.00019670745272525029, | |
| "loss": 4.2743, | |
| "step": 80 | |
| }, | |
| { | |
| "epoch": 1.080267558528428, | |
| "grad_norm": 0.4570735692977905, | |
| "learning_rate": 0.0001966629588431591, | |
| "loss": 4.8012, | |
| "step": 81 | |
| }, | |
| { | |
| "epoch": 1.0936454849498327, | |
| "grad_norm": 0.4814144968986511, | |
| "learning_rate": 0.00019661846496106785, | |
| "loss": 4.6449, | |
| "step": 82 | |
| }, | |
| { | |
| "epoch": 1.1070234113712374, | |
| "grad_norm": 0.4526231288909912, | |
| "learning_rate": 0.00019657397107897666, | |
| "loss": 4.5546, | |
| "step": 83 | |
| }, | |
| { | |
| "epoch": 1.120401337792642, | |
| "grad_norm": 0.4847906827926636, | |
| "learning_rate": 0.00019652947719688544, | |
| "loss": 4.4421, | |
| "step": 84 | |
| }, | |
| { | |
| "epoch": 1.1337792642140467, | |
| "grad_norm": 0.5136271715164185, | |
| "learning_rate": 0.00019648498331479422, | |
| "loss": 4.7136, | |
| "step": 85 | |
| }, | |
| { | |
| "epoch": 1.1471571906354514, | |
| "grad_norm": 0.49209895730018616, | |
| "learning_rate": 0.000196440489432703, | |
| "loss": 4.3145, | |
| "step": 86 | |
| }, | |
| { | |
| "epoch": 1.160535117056856, | |
| "grad_norm": 0.4972032904624939, | |
| "learning_rate": 0.00019639599555061182, | |
| "loss": 4.0408, | |
| "step": 87 | |
| }, | |
| { | |
| "epoch": 1.1739130434782608, | |
| "grad_norm": 0.5077862739562988, | |
| "learning_rate": 0.0001963515016685206, | |
| "loss": 4.4074, | |
| "step": 88 | |
| }, | |
| { | |
| "epoch": 1.1872909698996654, | |
| "grad_norm": 0.5293861031532288, | |
| "learning_rate": 0.00019630700778642935, | |
| "loss": 4.5385, | |
| "step": 89 | |
| }, | |
| { | |
| "epoch": 1.2006688963210703, | |
| "grad_norm": 0.5062645673751831, | |
| "learning_rate": 0.00019626251390433816, | |
| "loss": 4.5141, | |
| "step": 90 | |
| }, | |
| { | |
| "epoch": 1.214046822742475, | |
| "grad_norm": 0.49655866622924805, | |
| "learning_rate": 0.00019621802002224695, | |
| "loss": 4.4765, | |
| "step": 91 | |
| }, | |
| { | |
| "epoch": 1.2274247491638797, | |
| "grad_norm": 0.6059755086898804, | |
| "learning_rate": 0.00019617352614015573, | |
| "loss": 4.568, | |
| "step": 92 | |
| }, | |
| { | |
| "epoch": 1.2408026755852843, | |
| "grad_norm": 0.5442761778831482, | |
| "learning_rate": 0.0001961290322580645, | |
| "loss": 4.7724, | |
| "step": 93 | |
| }, | |
| { | |
| "epoch": 1.254180602006689, | |
| "grad_norm": 0.5426056385040283, | |
| "learning_rate": 0.00019608453837597332, | |
| "loss": 4.5308, | |
| "step": 94 | |
| }, | |
| { | |
| "epoch": 1.2675585284280937, | |
| "grad_norm": 0.525372326374054, | |
| "learning_rate": 0.0001960400444938821, | |
| "loss": 4.394, | |
| "step": 95 | |
| }, | |
| { | |
| "epoch": 1.2809364548494984, | |
| "grad_norm": 0.5407588481903076, | |
| "learning_rate": 0.00019599555061179089, | |
| "loss": 4.7347, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 1.294314381270903, | |
| "grad_norm": 0.5726659893989563, | |
| "learning_rate": 0.00019595105672969967, | |
| "loss": 4.9446, | |
| "step": 97 | |
| }, | |
| { | |
| "epoch": 1.3076923076923077, | |
| "grad_norm": 0.6211283206939697, | |
| "learning_rate": 0.00019590656284760848, | |
| "loss": 4.697, | |
| "step": 98 | |
| }, | |
| { | |
| "epoch": 1.3210702341137124, | |
| "grad_norm": 0.5627567172050476, | |
| "learning_rate": 0.00019586206896551723, | |
| "loss": 4.4892, | |
| "step": 99 | |
| }, | |
| { | |
| "epoch": 1.334448160535117, | |
| "grad_norm": 0.6174790859222412, | |
| "learning_rate": 0.00019581757508342604, | |
| "loss": 4.5686, | |
| "step": 100 | |
| }, | |
| { | |
| "epoch": 1.3478260869565217, | |
| "grad_norm": 0.5586990118026733, | |
| "learning_rate": 0.00019577308120133482, | |
| "loss": 4.3916, | |
| "step": 101 | |
| }, | |
| { | |
| "epoch": 1.3612040133779264, | |
| "grad_norm": 0.5655365586280823, | |
| "learning_rate": 0.0001957285873192436, | |
| "loss": 4.1789, | |
| "step": 102 | |
| }, | |
| { | |
| "epoch": 1.374581939799331, | |
| "grad_norm": 0.5834594964981079, | |
| "learning_rate": 0.0001956840934371524, | |
| "loss": 4.3316, | |
| "step": 103 | |
| }, | |
| { | |
| "epoch": 1.3879598662207357, | |
| "grad_norm": 0.6065447926521301, | |
| "learning_rate": 0.0001956395995550612, | |
| "loss": 4.5167, | |
| "step": 104 | |
| }, | |
| { | |
| "epoch": 1.4013377926421404, | |
| "grad_norm": 0.5250216722488403, | |
| "learning_rate": 0.00019559510567296998, | |
| "loss": 4.1718, | |
| "step": 105 | |
| }, | |
| { | |
| "epoch": 1.414715719063545, | |
| "grad_norm": 0.5861116051673889, | |
| "learning_rate": 0.00019555061179087876, | |
| "loss": 4.3077, | |
| "step": 106 | |
| }, | |
| { | |
| "epoch": 1.4280936454849498, | |
| "grad_norm": 0.6138104796409607, | |
| "learning_rate": 0.00019550611790878755, | |
| "loss": 4.4748, | |
| "step": 107 | |
| }, | |
| { | |
| "epoch": 1.4414715719063544, | |
| "grad_norm": 0.6742071509361267, | |
| "learning_rate": 0.00019546162402669636, | |
| "loss": 4.8769, | |
| "step": 108 | |
| }, | |
| { | |
| "epoch": 1.4548494983277591, | |
| "grad_norm": 0.6634951233863831, | |
| "learning_rate": 0.0001954171301446051, | |
| "loss": 4.6423, | |
| "step": 109 | |
| }, | |
| { | |
| "epoch": 1.468227424749164, | |
| "grad_norm": 0.626646876335144, | |
| "learning_rate": 0.00019537263626251392, | |
| "loss": 4.4654, | |
| "step": 110 | |
| }, | |
| { | |
| "epoch": 1.4816053511705687, | |
| "grad_norm": 0.6306963562965393, | |
| "learning_rate": 0.0001953281423804227, | |
| "loss": 4.7021, | |
| "step": 111 | |
| }, | |
| { | |
| "epoch": 1.4949832775919734, | |
| "grad_norm": 0.620370626449585, | |
| "learning_rate": 0.00019528364849833149, | |
| "loss": 4.587, | |
| "step": 112 | |
| }, | |
| { | |
| "epoch": 1.508361204013378, | |
| "grad_norm": 0.6410287618637085, | |
| "learning_rate": 0.00019523915461624027, | |
| "loss": 4.8089, | |
| "step": 113 | |
| }, | |
| { | |
| "epoch": 1.5217391304347827, | |
| "grad_norm": 0.676434338092804, | |
| "learning_rate": 0.00019519466073414908, | |
| "loss": 4.668, | |
| "step": 114 | |
| }, | |
| { | |
| "epoch": 1.5351170568561874, | |
| "grad_norm": 0.5756319761276245, | |
| "learning_rate": 0.00019515016685205786, | |
| "loss": 4.3223, | |
| "step": 115 | |
| }, | |
| { | |
| "epoch": 1.548494983277592, | |
| "grad_norm": 0.5850693583488464, | |
| "learning_rate": 0.00019510567296996664, | |
| "loss": 4.2343, | |
| "step": 116 | |
| }, | |
| { | |
| "epoch": 1.5618729096989967, | |
| "grad_norm": 0.6172360777854919, | |
| "learning_rate": 0.00019506117908787542, | |
| "loss": 4.6102, | |
| "step": 117 | |
| }, | |
| { | |
| "epoch": 1.5752508361204014, | |
| "grad_norm": 0.5887568593025208, | |
| "learning_rate": 0.00019501668520578423, | |
| "loss": 4.8097, | |
| "step": 118 | |
| }, | |
| { | |
| "epoch": 1.588628762541806, | |
| "grad_norm": 0.5763369798660278, | |
| "learning_rate": 0.000194972191323693, | |
| "loss": 4.2001, | |
| "step": 119 | |
| }, | |
| { | |
| "epoch": 1.6020066889632107, | |
| "grad_norm": 0.6158986687660217, | |
| "learning_rate": 0.0001949276974416018, | |
| "loss": 4.7075, | |
| "step": 120 | |
| }, | |
| { | |
| "epoch": 1.6153846153846154, | |
| "grad_norm": 0.5540957450866699, | |
| "learning_rate": 0.00019488320355951058, | |
| "loss": 4.452, | |
| "step": 121 | |
| }, | |
| { | |
| "epoch": 1.62876254180602, | |
| "grad_norm": 0.6193795204162598, | |
| "learning_rate": 0.00019483870967741936, | |
| "loss": 4.4583, | |
| "step": 122 | |
| }, | |
| { | |
| "epoch": 1.6421404682274248, | |
| "grad_norm": 0.6699966788291931, | |
| "learning_rate": 0.00019479421579532815, | |
| "loss": 4.3728, | |
| "step": 123 | |
| }, | |
| { | |
| "epoch": 1.6555183946488294, | |
| "grad_norm": 0.5904677510261536, | |
| "learning_rate": 0.00019474972191323696, | |
| "loss": 4.5452, | |
| "step": 124 | |
| }, | |
| { | |
| "epoch": 1.6688963210702341, | |
| "grad_norm": 0.6137760281562805, | |
| "learning_rate": 0.00019470522803114574, | |
| "loss": 4.2853, | |
| "step": 125 | |
| }, | |
| { | |
| "epoch": 1.6822742474916388, | |
| "grad_norm": 0.6396192908287048, | |
| "learning_rate": 0.00019466073414905452, | |
| "loss": 4.4258, | |
| "step": 126 | |
| }, | |
| { | |
| "epoch": 1.6956521739130435, | |
| "grad_norm": 0.6190487742424011, | |
| "learning_rate": 0.0001946162402669633, | |
| "loss": 4.9866, | |
| "step": 127 | |
| }, | |
| { | |
| "epoch": 1.7090301003344481, | |
| "grad_norm": 0.6971675157546997, | |
| "learning_rate": 0.0001945717463848721, | |
| "loss": 4.2126, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 1.7224080267558528, | |
| "grad_norm": 0.6245931386947632, | |
| "learning_rate": 0.00019452725250278087, | |
| "loss": 4.8477, | |
| "step": 129 | |
| }, | |
| { | |
| "epoch": 1.7357859531772575, | |
| "grad_norm": 0.5675052404403687, | |
| "learning_rate": 0.00019448275862068965, | |
| "loss": 4.4097, | |
| "step": 130 | |
| }, | |
| { | |
| "epoch": 1.7491638795986622, | |
| "grad_norm": 0.6594040393829346, | |
| "learning_rate": 0.00019443826473859846, | |
| "loss": 4.3747, | |
| "step": 131 | |
| }, | |
| { | |
| "epoch": 1.7625418060200668, | |
| "grad_norm": 0.6377655267715454, | |
| "learning_rate": 0.00019439377085650724, | |
| "loss": 4.2733, | |
| "step": 132 | |
| }, | |
| { | |
| "epoch": 1.7759197324414715, | |
| "grad_norm": 0.6167862415313721, | |
| "learning_rate": 0.00019434927697441602, | |
| "loss": 4.5694, | |
| "step": 133 | |
| }, | |
| { | |
| "epoch": 1.7892976588628762, | |
| "grad_norm": 0.577671468257904, | |
| "learning_rate": 0.0001943047830923248, | |
| "loss": 4.5006, | |
| "step": 134 | |
| }, | |
| { | |
| "epoch": 1.8026755852842808, | |
| "grad_norm": 0.6361016035079956, | |
| "learning_rate": 0.00019426028921023362, | |
| "loss": 4.9907, | |
| "step": 135 | |
| }, | |
| { | |
| "epoch": 1.8160535117056855, | |
| "grad_norm": 0.6445321440696716, | |
| "learning_rate": 0.00019421579532814237, | |
| "loss": 4.779, | |
| "step": 136 | |
| }, | |
| { | |
| "epoch": 1.8294314381270902, | |
| "grad_norm": 0.5955402851104736, | |
| "learning_rate": 0.00019417130144605118, | |
| "loss": 4.6026, | |
| "step": 137 | |
| }, | |
| { | |
| "epoch": 1.8428093645484949, | |
| "grad_norm": 0.6807080507278442, | |
| "learning_rate": 0.00019412680756395996, | |
| "loss": 4.7124, | |
| "step": 138 | |
| }, | |
| { | |
| "epoch": 1.8561872909698995, | |
| "grad_norm": 0.5799288153648376, | |
| "learning_rate": 0.00019408231368186875, | |
| "loss": 4.0701, | |
| "step": 139 | |
| }, | |
| { | |
| "epoch": 1.8695652173913042, | |
| "grad_norm": 0.6187757253646851, | |
| "learning_rate": 0.00019403781979977753, | |
| "loss": 4.705, | |
| "step": 140 | |
| }, | |
| { | |
| "epoch": 1.8829431438127089, | |
| "grad_norm": 0.6614826917648315, | |
| "learning_rate": 0.00019399332591768634, | |
| "loss": 4.8146, | |
| "step": 141 | |
| }, | |
| { | |
| "epoch": 1.8963210702341136, | |
| "grad_norm": 0.6204859614372253, | |
| "learning_rate": 0.00019394883203559512, | |
| "loss": 4.3041, | |
| "step": 142 | |
| }, | |
| { | |
| "epoch": 1.9096989966555182, | |
| "grad_norm": 0.6527450680732727, | |
| "learning_rate": 0.0001939043381535039, | |
| "loss": 4.4493, | |
| "step": 143 | |
| }, | |
| { | |
| "epoch": 1.9230769230769231, | |
| "grad_norm": 0.6470615267753601, | |
| "learning_rate": 0.00019385984427141268, | |
| "loss": 4.7771, | |
| "step": 144 | |
| }, | |
| { | |
| "epoch": 1.9364548494983278, | |
| "grad_norm": 0.5642555952072144, | |
| "learning_rate": 0.0001938153503893215, | |
| "loss": 4.3344, | |
| "step": 145 | |
| }, | |
| { | |
| "epoch": 1.9498327759197325, | |
| "grad_norm": 0.6206467151641846, | |
| "learning_rate": 0.00019377085650723025, | |
| "loss": 4.2191, | |
| "step": 146 | |
| }, | |
| { | |
| "epoch": 1.9632107023411371, | |
| "grad_norm": 0.6079016923904419, | |
| "learning_rate": 0.00019372636262513906, | |
| "loss": 4.7397, | |
| "step": 147 | |
| }, | |
| { | |
| "epoch": 1.9765886287625418, | |
| "grad_norm": 0.6197662353515625, | |
| "learning_rate": 0.00019368186874304784, | |
| "loss": 4.5342, | |
| "step": 148 | |
| }, | |
| { | |
| "epoch": 1.9899665551839465, | |
| "grad_norm": 0.6556297540664673, | |
| "learning_rate": 0.00019363737486095662, | |
| "loss": 4.6709, | |
| "step": 149 | |
| }, | |
| { | |
| "epoch": 2.0, | |
| "grad_norm": 0.7837930917739868, | |
| "learning_rate": 0.0001935928809788654, | |
| "loss": 4.6215, | |
| "step": 150 | |
| }, | |
| { | |
| "epoch": 2.0133779264214047, | |
| "grad_norm": 0.5267267227172852, | |
| "learning_rate": 0.00019354838709677422, | |
| "loss": 4.2695, | |
| "step": 151 | |
| }, | |
| { | |
| "epoch": 2.0267558528428093, | |
| "grad_norm": 0.5862157344818115, | |
| "learning_rate": 0.000193503893214683, | |
| "loss": 4.3702, | |
| "step": 152 | |
| }, | |
| { | |
| "epoch": 2.040133779264214, | |
| "grad_norm": 0.538254976272583, | |
| "learning_rate": 0.00019345939933259178, | |
| "loss": 4.3953, | |
| "step": 153 | |
| }, | |
| { | |
| "epoch": 2.0535117056856187, | |
| "grad_norm": 0.5977053642272949, | |
| "learning_rate": 0.00019341490545050056, | |
| "loss": 4.2156, | |
| "step": 154 | |
| }, | |
| { | |
| "epoch": 2.0668896321070234, | |
| "grad_norm": 0.606006383895874, | |
| "learning_rate": 0.00019337041156840937, | |
| "loss": 4.2802, | |
| "step": 155 | |
| }, | |
| { | |
| "epoch": 2.080267558528428, | |
| "grad_norm": 0.6071277856826782, | |
| "learning_rate": 0.00019332591768631813, | |
| "loss": 4.5545, | |
| "step": 156 | |
| }, | |
| { | |
| "epoch": 2.0936454849498327, | |
| "grad_norm": 0.6281546354293823, | |
| "learning_rate": 0.00019328142380422694, | |
| "loss": 4.6105, | |
| "step": 157 | |
| }, | |
| { | |
| "epoch": 2.1070234113712374, | |
| "grad_norm": 0.5703116655349731, | |
| "learning_rate": 0.00019323692992213572, | |
| "loss": 4.2751, | |
| "step": 158 | |
| }, | |
| { | |
| "epoch": 2.120401337792642, | |
| "grad_norm": 0.6587452292442322, | |
| "learning_rate": 0.0001931924360400445, | |
| "loss": 4.6342, | |
| "step": 159 | |
| }, | |
| { | |
| "epoch": 2.1337792642140467, | |
| "grad_norm": 0.6141905784606934, | |
| "learning_rate": 0.00019314794215795328, | |
| "loss": 4.4345, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 2.1471571906354514, | |
| "grad_norm": 0.6741939187049866, | |
| "learning_rate": 0.0001931034482758621, | |
| "loss": 4.0257, | |
| "step": 161 | |
| }, | |
| { | |
| "epoch": 2.160535117056856, | |
| "grad_norm": 0.6468759179115295, | |
| "learning_rate": 0.00019305895439377088, | |
| "loss": 4.2313, | |
| "step": 162 | |
| }, | |
| { | |
| "epoch": 2.1739130434782608, | |
| "grad_norm": 0.6703383326530457, | |
| "learning_rate": 0.00019301446051167966, | |
| "loss": 4.2164, | |
| "step": 163 | |
| }, | |
| { | |
| "epoch": 2.1872909698996654, | |
| "grad_norm": 0.710967481136322, | |
| "learning_rate": 0.00019296996662958844, | |
| "loss": 4.3398, | |
| "step": 164 | |
| }, | |
| { | |
| "epoch": 2.20066889632107, | |
| "grad_norm": 0.6862124800682068, | |
| "learning_rate": 0.00019292547274749725, | |
| "loss": 4.3379, | |
| "step": 165 | |
| }, | |
| { | |
| "epoch": 2.2140468227424748, | |
| "grad_norm": 0.6288430690765381, | |
| "learning_rate": 0.000192880978865406, | |
| "loss": 4.3487, | |
| "step": 166 | |
| }, | |
| { | |
| "epoch": 2.2274247491638794, | |
| "grad_norm": 0.6358796954154968, | |
| "learning_rate": 0.00019283648498331481, | |
| "loss": 4.1656, | |
| "step": 167 | |
| }, | |
| { | |
| "epoch": 2.240802675585284, | |
| "grad_norm": 0.6818917393684387, | |
| "learning_rate": 0.0001927919911012236, | |
| "loss": 4.5363, | |
| "step": 168 | |
| }, | |
| { | |
| "epoch": 2.254180602006689, | |
| "grad_norm": 0.6996105313301086, | |
| "learning_rate": 0.00019274749721913238, | |
| "loss": 4.3208, | |
| "step": 169 | |
| }, | |
| { | |
| "epoch": 2.2675585284280935, | |
| "grad_norm": 0.6730326414108276, | |
| "learning_rate": 0.00019270300333704116, | |
| "loss": 4.1401, | |
| "step": 170 | |
| }, | |
| { | |
| "epoch": 2.280936454849498, | |
| "grad_norm": 0.7022603750228882, | |
| "learning_rate": 0.00019265850945494994, | |
| "loss": 4.5761, | |
| "step": 171 | |
| }, | |
| { | |
| "epoch": 2.294314381270903, | |
| "grad_norm": 0.6525995135307312, | |
| "learning_rate": 0.00019261401557285875, | |
| "loss": 4.4017, | |
| "step": 172 | |
| }, | |
| { | |
| "epoch": 2.3076923076923075, | |
| "grad_norm": 0.7066033482551575, | |
| "learning_rate": 0.0001925695216907675, | |
| "loss": 4.0037, | |
| "step": 173 | |
| }, | |
| { | |
| "epoch": 2.321070234113712, | |
| "grad_norm": 0.6708059310913086, | |
| "learning_rate": 0.00019252502780867632, | |
| "loss": 4.1947, | |
| "step": 174 | |
| }, | |
| { | |
| "epoch": 2.334448160535117, | |
| "grad_norm": 0.8711172342300415, | |
| "learning_rate": 0.0001924805339265851, | |
| "loss": 3.9958, | |
| "step": 175 | |
| }, | |
| { | |
| "epoch": 2.3478260869565215, | |
| "grad_norm": 0.7258634567260742, | |
| "learning_rate": 0.00019243604004449388, | |
| "loss": 4.4682, | |
| "step": 176 | |
| }, | |
| { | |
| "epoch": 2.361204013377926, | |
| "grad_norm": 0.7693021893501282, | |
| "learning_rate": 0.00019239154616240267, | |
| "loss": 4.54, | |
| "step": 177 | |
| }, | |
| { | |
| "epoch": 2.374581939799331, | |
| "grad_norm": 0.7271276116371155, | |
| "learning_rate": 0.00019234705228031148, | |
| "loss": 4.2942, | |
| "step": 178 | |
| }, | |
| { | |
| "epoch": 2.387959866220736, | |
| "grad_norm": 0.6836609244346619, | |
| "learning_rate": 0.00019230255839822026, | |
| "loss": 4.3099, | |
| "step": 179 | |
| }, | |
| { | |
| "epoch": 2.4013377926421406, | |
| "grad_norm": 0.731164813041687, | |
| "learning_rate": 0.00019225806451612904, | |
| "loss": 4.4077, | |
| "step": 180 | |
| }, | |
| { | |
| "epoch": 2.4147157190635453, | |
| "grad_norm": 0.7575274109840393, | |
| "learning_rate": 0.00019221357063403782, | |
| "loss": 4.6572, | |
| "step": 181 | |
| }, | |
| { | |
| "epoch": 2.42809364548495, | |
| "grad_norm": 0.8461325168609619, | |
| "learning_rate": 0.00019216907675194663, | |
| "loss": 4.4922, | |
| "step": 182 | |
| }, | |
| { | |
| "epoch": 2.4414715719063547, | |
| "grad_norm": 0.7225251197814941, | |
| "learning_rate": 0.0001921245828698554, | |
| "loss": 4.0372, | |
| "step": 183 | |
| }, | |
| { | |
| "epoch": 2.4548494983277593, | |
| "grad_norm": 3.563720703125, | |
| "learning_rate": 0.0001920800889877642, | |
| "loss": 4.5412, | |
| "step": 184 | |
| }, | |
| { | |
| "epoch": 2.468227424749164, | |
| "grad_norm": 0.8452121019363403, | |
| "learning_rate": 0.00019203559510567298, | |
| "loss": 4.4961, | |
| "step": 185 | |
| }, | |
| { | |
| "epoch": 2.4816053511705687, | |
| "grad_norm": 0.8734024167060852, | |
| "learning_rate": 0.00019199110122358176, | |
| "loss": 4.0884, | |
| "step": 186 | |
| }, | |
| { | |
| "epoch": 2.4949832775919734, | |
| "grad_norm": 1.1765823364257812, | |
| "learning_rate": 0.00019194660734149054, | |
| "loss": 4.2228, | |
| "step": 187 | |
| }, | |
| { | |
| "epoch": 2.508361204013378, | |
| "grad_norm": 0.750206708908081, | |
| "learning_rate": 0.00019190211345939935, | |
| "loss": 4.4305, | |
| "step": 188 | |
| }, | |
| { | |
| "epoch": 2.5217391304347827, | |
| "grad_norm": 0.7574430704116821, | |
| "learning_rate": 0.00019185761957730814, | |
| "loss": 4.1511, | |
| "step": 189 | |
| }, | |
| { | |
| "epoch": 2.5351170568561874, | |
| "grad_norm": 0.7105517387390137, | |
| "learning_rate": 0.00019181312569521692, | |
| "loss": 4.4793, | |
| "step": 190 | |
| }, | |
| { | |
| "epoch": 2.548494983277592, | |
| "grad_norm": 0.7495557069778442, | |
| "learning_rate": 0.0001917686318131257, | |
| "loss": 4.1335, | |
| "step": 191 | |
| }, | |
| { | |
| "epoch": 2.5618729096989967, | |
| "grad_norm": 0.8001168966293335, | |
| "learning_rate": 0.0001917241379310345, | |
| "loss": 4.7898, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 2.5752508361204014, | |
| "grad_norm": 0.7402104735374451, | |
| "learning_rate": 0.00019167964404894327, | |
| "loss": 4.4482, | |
| "step": 193 | |
| }, | |
| { | |
| "epoch": 2.588628762541806, | |
| "grad_norm": 0.748267650604248, | |
| "learning_rate": 0.00019163515016685207, | |
| "loss": 4.3167, | |
| "step": 194 | |
| }, | |
| { | |
| "epoch": 2.6020066889632107, | |
| "grad_norm": 0.8291250467300415, | |
| "learning_rate": 0.00019159065628476086, | |
| "loss": 4.058, | |
| "step": 195 | |
| }, | |
| { | |
| "epoch": 2.6153846153846154, | |
| "grad_norm": 0.6945542693138123, | |
| "learning_rate": 0.00019154616240266964, | |
| "loss": 3.9751, | |
| "step": 196 | |
| }, | |
| { | |
| "epoch": 2.62876254180602, | |
| "grad_norm": 0.7307319045066833, | |
| "learning_rate": 0.00019150166852057842, | |
| "loss": 4.2736, | |
| "step": 197 | |
| }, | |
| { | |
| "epoch": 2.6421404682274248, | |
| "grad_norm": 0.7489168047904968, | |
| "learning_rate": 0.00019145717463848723, | |
| "loss": 4.3075, | |
| "step": 198 | |
| }, | |
| { | |
| "epoch": 2.6555183946488294, | |
| "grad_norm": 0.9727582931518555, | |
| "learning_rate": 0.00019141268075639601, | |
| "loss": 4.6474, | |
| "step": 199 | |
| }, | |
| { | |
| "epoch": 2.668896321070234, | |
| "grad_norm": 0.6776256561279297, | |
| "learning_rate": 0.0001913681868743048, | |
| "loss": 4.4217, | |
| "step": 200 | |
| }, | |
| { | |
| "epoch": 2.682274247491639, | |
| "grad_norm": 0.7305111885070801, | |
| "learning_rate": 0.00019132369299221358, | |
| "loss": 4.2804, | |
| "step": 201 | |
| }, | |
| { | |
| "epoch": 2.6956521739130435, | |
| "grad_norm": 0.7196978330612183, | |
| "learning_rate": 0.0001912791991101224, | |
| "loss": 4.3941, | |
| "step": 202 | |
| }, | |
| { | |
| "epoch": 2.709030100334448, | |
| "grad_norm": 0.7988458871841431, | |
| "learning_rate": 0.00019123470522803114, | |
| "loss": 4.437, | |
| "step": 203 | |
| }, | |
| { | |
| "epoch": 2.722408026755853, | |
| "grad_norm": 0.7004797458648682, | |
| "learning_rate": 0.00019119021134593995, | |
| "loss": 4.4986, | |
| "step": 204 | |
| }, | |
| { | |
| "epoch": 2.7357859531772575, | |
| "grad_norm": 0.677796483039856, | |
| "learning_rate": 0.00019114571746384874, | |
| "loss": 4.0851, | |
| "step": 205 | |
| }, | |
| { | |
| "epoch": 2.749163879598662, | |
| "grad_norm": 0.7527475357055664, | |
| "learning_rate": 0.00019110122358175752, | |
| "loss": 4.4469, | |
| "step": 206 | |
| }, | |
| { | |
| "epoch": 2.762541806020067, | |
| "grad_norm": 1.1659115552902222, | |
| "learning_rate": 0.0001910567296996663, | |
| "loss": 4.3284, | |
| "step": 207 | |
| }, | |
| { | |
| "epoch": 2.7759197324414715, | |
| "grad_norm": 0.7238364815711975, | |
| "learning_rate": 0.0001910122358175751, | |
| "loss": 4.2605, | |
| "step": 208 | |
| }, | |
| { | |
| "epoch": 2.789297658862876, | |
| "grad_norm": 0.7537760734558105, | |
| "learning_rate": 0.0001909677419354839, | |
| "loss": 4.3775, | |
| "step": 209 | |
| }, | |
| { | |
| "epoch": 2.802675585284281, | |
| "grad_norm": 0.6874127388000488, | |
| "learning_rate": 0.00019092324805339267, | |
| "loss": 4.3404, | |
| "step": 210 | |
| }, | |
| { | |
| "epoch": 2.8160535117056855, | |
| "grad_norm": 0.7045959830284119, | |
| "learning_rate": 0.00019087875417130146, | |
| "loss": 4.1568, | |
| "step": 211 | |
| }, | |
| { | |
| "epoch": 2.82943143812709, | |
| "grad_norm": 0.7249194383621216, | |
| "learning_rate": 0.00019083426028921027, | |
| "loss": 4.1969, | |
| "step": 212 | |
| }, | |
| { | |
| "epoch": 2.842809364548495, | |
| "grad_norm": 0.8331268429756165, | |
| "learning_rate": 0.00019078976640711902, | |
| "loss": 4.3169, | |
| "step": 213 | |
| }, | |
| { | |
| "epoch": 2.8561872909698995, | |
| "grad_norm": 0.7171936631202698, | |
| "learning_rate": 0.0001907452725250278, | |
| "loss": 4.5123, | |
| "step": 214 | |
| }, | |
| { | |
| "epoch": 2.869565217391304, | |
| "grad_norm": 0.759919285774231, | |
| "learning_rate": 0.0001907007786429366, | |
| "loss": 4.5412, | |
| "step": 215 | |
| }, | |
| { | |
| "epoch": 2.882943143812709, | |
| "grad_norm": 0.7451274991035461, | |
| "learning_rate": 0.0001906562847608454, | |
| "loss": 4.5253, | |
| "step": 216 | |
| }, | |
| { | |
| "epoch": 2.8963210702341136, | |
| "grad_norm": 0.6564481258392334, | |
| "learning_rate": 0.00019061179087875418, | |
| "loss": 4.1092, | |
| "step": 217 | |
| }, | |
| { | |
| "epoch": 2.9096989966555182, | |
| "grad_norm": 0.7339865565299988, | |
| "learning_rate": 0.00019056729699666296, | |
| "loss": 4.5092, | |
| "step": 218 | |
| }, | |
| { | |
| "epoch": 2.9230769230769234, | |
| "grad_norm": 0.7113937735557556, | |
| "learning_rate": 0.00019052280311457177, | |
| "loss": 4.3355, | |
| "step": 219 | |
| }, | |
| { | |
| "epoch": 2.936454849498328, | |
| "grad_norm": 0.7306456565856934, | |
| "learning_rate": 0.00019047830923248053, | |
| "loss": 4.5745, | |
| "step": 220 | |
| }, | |
| { | |
| "epoch": 2.9498327759197327, | |
| "grad_norm": 0.7971818447113037, | |
| "learning_rate": 0.00019043381535038933, | |
| "loss": 4.2903, | |
| "step": 221 | |
| }, | |
| { | |
| "epoch": 2.9632107023411374, | |
| "grad_norm": 0.7757331728935242, | |
| "learning_rate": 0.00019038932146829812, | |
| "loss": 4.2832, | |
| "step": 222 | |
| }, | |
| { | |
| "epoch": 2.976588628762542, | |
| "grad_norm": 0.7326288223266602, | |
| "learning_rate": 0.0001903448275862069, | |
| "loss": 4.2444, | |
| "step": 223 | |
| }, | |
| { | |
| "epoch": 2.9899665551839467, | |
| "grad_norm": 0.7363834381103516, | |
| "learning_rate": 0.00019030033370411568, | |
| "loss": 4.6744, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 3.0, | |
| "grad_norm": 0.8835271596908569, | |
| "learning_rate": 0.0001902558398220245, | |
| "loss": 4.432, | |
| "step": 225 | |
| }, | |
| { | |
| "epoch": 3.0133779264214047, | |
| "grad_norm": 0.6591921448707581, | |
| "learning_rate": 0.00019021134593993327, | |
| "loss": 4.1353, | |
| "step": 226 | |
| }, | |
| { | |
| "epoch": 3.0267558528428093, | |
| "grad_norm": 0.6895263195037842, | |
| "learning_rate": 0.00019016685205784206, | |
| "loss": 4.1253, | |
| "step": 227 | |
| }, | |
| { | |
| "epoch": 3.040133779264214, | |
| "grad_norm": 0.6476898789405823, | |
| "learning_rate": 0.00019012235817575084, | |
| "loss": 4.0354, | |
| "step": 228 | |
| }, | |
| { | |
| "epoch": 3.0535117056856187, | |
| "grad_norm": 0.6398957967758179, | |
| "learning_rate": 0.00019007786429365965, | |
| "loss": 4.024, | |
| "step": 229 | |
| }, | |
| { | |
| "epoch": 3.0668896321070234, | |
| "grad_norm": 0.7483389973640442, | |
| "learning_rate": 0.0001900333704115684, | |
| "loss": 4.1405, | |
| "step": 230 | |
| }, | |
| { | |
| "epoch": 3.080267558528428, | |
| "grad_norm": 0.7003724575042725, | |
| "learning_rate": 0.0001899888765294772, | |
| "loss": 4.3593, | |
| "step": 231 | |
| }, | |
| { | |
| "epoch": 3.0936454849498327, | |
| "grad_norm": 0.7426732182502747, | |
| "learning_rate": 0.000189944382647386, | |
| "loss": 4.485, | |
| "step": 232 | |
| }, | |
| { | |
| "epoch": 3.1070234113712374, | |
| "grad_norm": 0.6957541108131409, | |
| "learning_rate": 0.00018989988876529478, | |
| "loss": 4.1017, | |
| "step": 233 | |
| }, | |
| { | |
| "epoch": 3.120401337792642, | |
| "grad_norm": 0.8613067865371704, | |
| "learning_rate": 0.00018985539488320356, | |
| "loss": 4.3038, | |
| "step": 234 | |
| }, | |
| { | |
| "epoch": 3.1337792642140467, | |
| "grad_norm": 0.8375754952430725, | |
| "learning_rate": 0.00018981090100111237, | |
| "loss": 4.4356, | |
| "step": 235 | |
| }, | |
| { | |
| "epoch": 3.1471571906354514, | |
| "grad_norm": 0.7878522872924805, | |
| "learning_rate": 0.00018976640711902115, | |
| "loss": 3.9916, | |
| "step": 236 | |
| }, | |
| { | |
| "epoch": 3.160535117056856, | |
| "grad_norm": 0.7463901042938232, | |
| "learning_rate": 0.00018972191323692993, | |
| "loss": 3.6761, | |
| "step": 237 | |
| }, | |
| { | |
| "epoch": 3.1739130434782608, | |
| "grad_norm": 0.7360939979553223, | |
| "learning_rate": 0.00018967741935483872, | |
| "loss": 3.9573, | |
| "step": 238 | |
| }, | |
| { | |
| "epoch": 3.1872909698996654, | |
| "grad_norm": 0.891861081123352, | |
| "learning_rate": 0.00018963292547274753, | |
| "loss": 4.1853, | |
| "step": 239 | |
| }, | |
| { | |
| "epoch": 3.20066889632107, | |
| "grad_norm": 0.8589549660682678, | |
| "learning_rate": 0.00018958843159065628, | |
| "loss": 4.0679, | |
| "step": 240 | |
| }, | |
| { | |
| "epoch": 3.2140468227424748, | |
| "grad_norm": 0.9534163475036621, | |
| "learning_rate": 0.0001895439377085651, | |
| "loss": 3.732, | |
| "step": 241 | |
| }, | |
| { | |
| "epoch": 3.2274247491638794, | |
| "grad_norm": 0.8968185186386108, | |
| "learning_rate": 0.00018949944382647387, | |
| "loss": 4.2217, | |
| "step": 242 | |
| }, | |
| { | |
| "epoch": 3.240802675585284, | |
| "grad_norm": 0.81589275598526, | |
| "learning_rate": 0.00018945494994438266, | |
| "loss": 4.428, | |
| "step": 243 | |
| }, | |
| { | |
| "epoch": 3.254180602006689, | |
| "grad_norm": 0.929050862789154, | |
| "learning_rate": 0.00018941045606229144, | |
| "loss": 4.3468, | |
| "step": 244 | |
| }, | |
| { | |
| "epoch": 3.2675585284280935, | |
| "grad_norm": 0.8535035252571106, | |
| "learning_rate": 0.00018936596218020025, | |
| "loss": 3.8489, | |
| "step": 245 | |
| }, | |
| { | |
| "epoch": 3.280936454849498, | |
| "grad_norm": 0.9484681487083435, | |
| "learning_rate": 0.00018932146829810903, | |
| "loss": 4.0132, | |
| "step": 246 | |
| }, | |
| { | |
| "epoch": 3.294314381270903, | |
| "grad_norm": 0.8190047144889832, | |
| "learning_rate": 0.0001892769744160178, | |
| "loss": 4.3574, | |
| "step": 247 | |
| }, | |
| { | |
| "epoch": 3.3076923076923075, | |
| "grad_norm": 0.8764749765396118, | |
| "learning_rate": 0.0001892324805339266, | |
| "loss": 4.3103, | |
| "step": 248 | |
| }, | |
| { | |
| "epoch": 3.321070234113712, | |
| "grad_norm": 0.8929185271263123, | |
| "learning_rate": 0.0001891879866518354, | |
| "loss": 4.3606, | |
| "step": 249 | |
| }, | |
| { | |
| "epoch": 3.334448160535117, | |
| "grad_norm": 0.9096692204475403, | |
| "learning_rate": 0.00018914349276974416, | |
| "loss": 4.0047, | |
| "step": 250 | |
| }, | |
| { | |
| "epoch": 3.3478260869565215, | |
| "grad_norm": 0.885143518447876, | |
| "learning_rate": 0.00018909899888765297, | |
| "loss": 4.182, | |
| "step": 251 | |
| }, | |
| { | |
| "epoch": 3.361204013377926, | |
| "grad_norm": 0.7724215984344482, | |
| "learning_rate": 0.00018905450500556175, | |
| "loss": 3.9529, | |
| "step": 252 | |
| }, | |
| { | |
| "epoch": 3.374581939799331, | |
| "grad_norm": 0.8351865410804749, | |
| "learning_rate": 0.00018901001112347053, | |
| "loss": 3.9533, | |
| "step": 253 | |
| }, | |
| { | |
| "epoch": 3.387959866220736, | |
| "grad_norm": 0.8684999942779541, | |
| "learning_rate": 0.00018896551724137932, | |
| "loss": 3.8594, | |
| "step": 254 | |
| }, | |
| { | |
| "epoch": 3.4013377926421406, | |
| "grad_norm": 0.8903334736824036, | |
| "learning_rate": 0.0001889210233592881, | |
| "loss": 3.9248, | |
| "step": 255 | |
| }, | |
| { | |
| "epoch": 3.4147157190635453, | |
| "grad_norm": 0.826690137386322, | |
| "learning_rate": 0.0001888765294771969, | |
| "loss": 4.0389, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 3.42809364548495, | |
| "grad_norm": 0.8306142687797546, | |
| "learning_rate": 0.00018883203559510566, | |
| "loss": 3.8168, | |
| "step": 257 | |
| }, | |
| { | |
| "epoch": 3.4414715719063547, | |
| "grad_norm": 0.9032199382781982, | |
| "learning_rate": 0.00018878754171301447, | |
| "loss": 4.178, | |
| "step": 258 | |
| }, | |
| { | |
| "epoch": 3.4548494983277593, | |
| "grad_norm": 0.9081966280937195, | |
| "learning_rate": 0.00018874304783092326, | |
| "loss": 4.2583, | |
| "step": 259 | |
| }, | |
| { | |
| "epoch": 3.468227424749164, | |
| "grad_norm": 0.8424077033996582, | |
| "learning_rate": 0.00018869855394883204, | |
| "loss": 4.3285, | |
| "step": 260 | |
| }, | |
| { | |
| "epoch": 3.4816053511705687, | |
| "grad_norm": 0.8302170038223267, | |
| "learning_rate": 0.00018865406006674082, | |
| "loss": 4.1346, | |
| "step": 261 | |
| }, | |
| { | |
| "epoch": 3.4949832775919734, | |
| "grad_norm": 0.8747193217277527, | |
| "learning_rate": 0.00018860956618464963, | |
| "loss": 4.0747, | |
| "step": 262 | |
| }, | |
| { | |
| "epoch": 3.508361204013378, | |
| "grad_norm": 0.8613927364349365, | |
| "learning_rate": 0.0001885650723025584, | |
| "loss": 4.2346, | |
| "step": 263 | |
| }, | |
| { | |
| "epoch": 3.5217391304347827, | |
| "grad_norm": 0.8321558833122253, | |
| "learning_rate": 0.0001885205784204672, | |
| "loss": 3.9781, | |
| "step": 264 | |
| }, | |
| { | |
| "epoch": 3.5351170568561874, | |
| "grad_norm": 0.8961741328239441, | |
| "learning_rate": 0.00018847608453837598, | |
| "loss": 4.311, | |
| "step": 265 | |
| }, | |
| { | |
| "epoch": 3.548494983277592, | |
| "grad_norm": 0.7703898549079895, | |
| "learning_rate": 0.00018843159065628479, | |
| "loss": 4.1163, | |
| "step": 266 | |
| }, | |
| { | |
| "epoch": 3.5618729096989967, | |
| "grad_norm": 0.880051851272583, | |
| "learning_rate": 0.00018838709677419354, | |
| "loss": 3.8032, | |
| "step": 267 | |
| }, | |
| { | |
| "epoch": 3.5752508361204014, | |
| "grad_norm": 0.8287038207054138, | |
| "learning_rate": 0.00018834260289210235, | |
| "loss": 4.1627, | |
| "step": 268 | |
| }, | |
| { | |
| "epoch": 3.588628762541806, | |
| "grad_norm": 0.9726569652557373, | |
| "learning_rate": 0.00018829810901001113, | |
| "loss": 4.4055, | |
| "step": 269 | |
| }, | |
| { | |
| "epoch": 3.6020066889632107, | |
| "grad_norm": 0.8071132898330688, | |
| "learning_rate": 0.00018825361512791992, | |
| "loss": 4.1709, | |
| "step": 270 | |
| }, | |
| { | |
| "epoch": 3.6153846153846154, | |
| "grad_norm": 0.8310988545417786, | |
| "learning_rate": 0.0001882091212458287, | |
| "loss": 4.2359, | |
| "step": 271 | |
| }, | |
| { | |
| "epoch": 3.62876254180602, | |
| "grad_norm": 0.8713561296463013, | |
| "learning_rate": 0.0001881646273637375, | |
| "loss": 4.1247, | |
| "step": 272 | |
| }, | |
| { | |
| "epoch": 3.6421404682274248, | |
| "grad_norm": 0.8964342474937439, | |
| "learning_rate": 0.0001881201334816463, | |
| "loss": 4.0794, | |
| "step": 273 | |
| }, | |
| { | |
| "epoch": 3.6555183946488294, | |
| "grad_norm": 0.9901681542396545, | |
| "learning_rate": 0.00018807563959955507, | |
| "loss": 4.0217, | |
| "step": 274 | |
| }, | |
| { | |
| "epoch": 3.668896321070234, | |
| "grad_norm": 0.9279042482376099, | |
| "learning_rate": 0.00018803114571746385, | |
| "loss": 4.3244, | |
| "step": 275 | |
| }, | |
| { | |
| "epoch": 3.682274247491639, | |
| "grad_norm": 0.8105964660644531, | |
| "learning_rate": 0.00018798665183537266, | |
| "loss": 3.9041, | |
| "step": 276 | |
| }, | |
| { | |
| "epoch": 3.6956521739130435, | |
| "grad_norm": 0.8511622548103333, | |
| "learning_rate": 0.00018794215795328142, | |
| "loss": 3.8969, | |
| "step": 277 | |
| }, | |
| { | |
| "epoch": 3.709030100334448, | |
| "grad_norm": 0.9072037935256958, | |
| "learning_rate": 0.00018789766407119023, | |
| "loss": 4.2185, | |
| "step": 278 | |
| }, | |
| { | |
| "epoch": 3.722408026755853, | |
| "grad_norm": 0.9792962670326233, | |
| "learning_rate": 0.000187853170189099, | |
| "loss": 4.1915, | |
| "step": 279 | |
| }, | |
| { | |
| "epoch": 3.7357859531772575, | |
| "grad_norm": 0.8579828143119812, | |
| "learning_rate": 0.0001878086763070078, | |
| "loss": 3.8903, | |
| "step": 280 | |
| }, | |
| { | |
| "epoch": 3.749163879598662, | |
| "grad_norm": 0.9866719841957092, | |
| "learning_rate": 0.00018776418242491658, | |
| "loss": 4.2022, | |
| "step": 281 | |
| }, | |
| { | |
| "epoch": 3.762541806020067, | |
| "grad_norm": 0.9251964688301086, | |
| "learning_rate": 0.00018771968854282539, | |
| "loss": 3.9536, | |
| "step": 282 | |
| }, | |
| { | |
| "epoch": 3.7759197324414715, | |
| "grad_norm": 1.0300836563110352, | |
| "learning_rate": 0.00018767519466073417, | |
| "loss": 4.2908, | |
| "step": 283 | |
| }, | |
| { | |
| "epoch": 3.789297658862876, | |
| "grad_norm": 1.0194575786590576, | |
| "learning_rate": 0.00018763070077864295, | |
| "loss": 4.1851, | |
| "step": 284 | |
| }, | |
| { | |
| "epoch": 3.802675585284281, | |
| "grad_norm": 0.8165330290794373, | |
| "learning_rate": 0.00018758620689655173, | |
| "loss": 4.138, | |
| "step": 285 | |
| }, | |
| { | |
| "epoch": 3.8160535117056855, | |
| "grad_norm": 1.0104280710220337, | |
| "learning_rate": 0.00018754171301446054, | |
| "loss": 3.9481, | |
| "step": 286 | |
| }, | |
| { | |
| "epoch": 3.82943143812709, | |
| "grad_norm": 0.9972538352012634, | |
| "learning_rate": 0.0001874972191323693, | |
| "loss": 4.3932, | |
| "step": 287 | |
| }, | |
| { | |
| "epoch": 3.842809364548495, | |
| "grad_norm": 0.96323162317276, | |
| "learning_rate": 0.0001874527252502781, | |
| "loss": 4.1133, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 3.8561872909698995, | |
| "grad_norm": 0.8500615954399109, | |
| "learning_rate": 0.0001874082313681869, | |
| "loss": 4.2205, | |
| "step": 289 | |
| }, | |
| { | |
| "epoch": 3.869565217391304, | |
| "grad_norm": 0.8451250195503235, | |
| "learning_rate": 0.00018736373748609567, | |
| "loss": 4.1371, | |
| "step": 290 | |
| }, | |
| { | |
| "epoch": 3.882943143812709, | |
| "grad_norm": 0.9399815201759338, | |
| "learning_rate": 0.00018731924360400445, | |
| "loss": 4.5237, | |
| "step": 291 | |
| }, | |
| { | |
| "epoch": 3.8963210702341136, | |
| "grad_norm": 0.8061622977256775, | |
| "learning_rate": 0.00018727474972191326, | |
| "loss": 4.1033, | |
| "step": 292 | |
| }, | |
| { | |
| "epoch": 3.9096989966555182, | |
| "grad_norm": 0.7987121343612671, | |
| "learning_rate": 0.00018723025583982205, | |
| "loss": 3.9311, | |
| "step": 293 | |
| }, | |
| { | |
| "epoch": 3.9230769230769234, | |
| "grad_norm": 0.9041138291358948, | |
| "learning_rate": 0.00018718576195773083, | |
| "loss": 4.0252, | |
| "step": 294 | |
| }, | |
| { | |
| "epoch": 3.936454849498328, | |
| "grad_norm": 1.0002484321594238, | |
| "learning_rate": 0.0001871412680756396, | |
| "loss": 4.4605, | |
| "step": 295 | |
| }, | |
| { | |
| "epoch": 3.9498327759197327, | |
| "grad_norm": 0.9991098046302795, | |
| "learning_rate": 0.0001870967741935484, | |
| "loss": 4.1528, | |
| "step": 296 | |
| }, | |
| { | |
| "epoch": 3.9632107023411374, | |
| "grad_norm": 1.2179397344589233, | |
| "learning_rate": 0.00018705228031145718, | |
| "loss": 4.5224, | |
| "step": 297 | |
| }, | |
| { | |
| "epoch": 3.976588628762542, | |
| "grad_norm": 0.8279774785041809, | |
| "learning_rate": 0.00018700778642936596, | |
| "loss": 3.9464, | |
| "step": 298 | |
| }, | |
| { | |
| "epoch": 3.9899665551839467, | |
| "grad_norm": 0.8012803792953491, | |
| "learning_rate": 0.00018696329254727477, | |
| "loss": 4.0139, | |
| "step": 299 | |
| }, | |
| { | |
| "epoch": 4.0, | |
| "grad_norm": 0.9700272083282471, | |
| "learning_rate": 0.00018691879866518355, | |
| "loss": 3.8306, | |
| "step": 300 | |
| }, | |
| { | |
| "epoch": 4.013377926421405, | |
| "grad_norm": 0.7136749625205994, | |
| "learning_rate": 0.00018687430478309233, | |
| "loss": 3.9253, | |
| "step": 301 | |
| }, | |
| { | |
| "epoch": 4.026755852842809, | |
| "grad_norm": 0.7885096669197083, | |
| "learning_rate": 0.00018682981090100111, | |
| "loss": 3.927, | |
| "step": 302 | |
| }, | |
| { | |
| "epoch": 4.040133779264214, | |
| "grad_norm": 0.7801666855812073, | |
| "learning_rate": 0.00018678531701890992, | |
| "loss": 3.6482, | |
| "step": 303 | |
| }, | |
| { | |
| "epoch": 4.053511705685619, | |
| "grad_norm": 0.7857955098152161, | |
| "learning_rate": 0.00018674082313681868, | |
| "loss": 4.0665, | |
| "step": 304 | |
| }, | |
| { | |
| "epoch": 4.066889632107023, | |
| "grad_norm": 0.707421064376831, | |
| "learning_rate": 0.0001866963292547275, | |
| "loss": 3.9142, | |
| "step": 305 | |
| }, | |
| { | |
| "epoch": 4.080267558528428, | |
| "grad_norm": 0.7936912775039673, | |
| "learning_rate": 0.00018665183537263627, | |
| "loss": 4.1227, | |
| "step": 306 | |
| }, | |
| { | |
| "epoch": 4.093645484949833, | |
| "grad_norm": 0.8899754881858826, | |
| "learning_rate": 0.00018660734149054505, | |
| "loss": 3.7661, | |
| "step": 307 | |
| }, | |
| { | |
| "epoch": 4.107023411371237, | |
| "grad_norm": 0.7760347723960876, | |
| "learning_rate": 0.00018656284760845384, | |
| "loss": 3.8921, | |
| "step": 308 | |
| }, | |
| { | |
| "epoch": 4.120401337792642, | |
| "grad_norm": 0.8672968745231628, | |
| "learning_rate": 0.00018651835372636265, | |
| "loss": 3.6037, | |
| "step": 309 | |
| }, | |
| { | |
| "epoch": 4.133779264214047, | |
| "grad_norm": 0.8046863675117493, | |
| "learning_rate": 0.0001864738598442714, | |
| "loss": 3.9117, | |
| "step": 310 | |
| }, | |
| { | |
| "epoch": 4.147157190635451, | |
| "grad_norm": 0.9172897934913635, | |
| "learning_rate": 0.0001864293659621802, | |
| "loss": 3.7229, | |
| "step": 311 | |
| }, | |
| { | |
| "epoch": 4.160535117056856, | |
| "grad_norm": 0.9616653919219971, | |
| "learning_rate": 0.000186384872080089, | |
| "loss": 3.8851, | |
| "step": 312 | |
| }, | |
| { | |
| "epoch": 4.173913043478261, | |
| "grad_norm": 0.9659278988838196, | |
| "learning_rate": 0.0001863403781979978, | |
| "loss": 4.005, | |
| "step": 313 | |
| }, | |
| { | |
| "epoch": 4.187290969899665, | |
| "grad_norm": 0.9171205163002014, | |
| "learning_rate": 0.00018629588431590656, | |
| "loss": 3.8634, | |
| "step": 314 | |
| }, | |
| { | |
| "epoch": 4.20066889632107, | |
| "grad_norm": 0.9968683123588562, | |
| "learning_rate": 0.00018625139043381537, | |
| "loss": 3.7321, | |
| "step": 315 | |
| }, | |
| { | |
| "epoch": 4.214046822742475, | |
| "grad_norm": 0.8762083053588867, | |
| "learning_rate": 0.00018620689655172415, | |
| "loss": 3.931, | |
| "step": 316 | |
| }, | |
| { | |
| "epoch": 4.2274247491638794, | |
| "grad_norm": 0.9815887212753296, | |
| "learning_rate": 0.00018616240266963293, | |
| "loss": 3.9975, | |
| "step": 317 | |
| }, | |
| { | |
| "epoch": 4.240802675585284, | |
| "grad_norm": 1.0065505504608154, | |
| "learning_rate": 0.00018611790878754171, | |
| "loss": 3.8364, | |
| "step": 318 | |
| }, | |
| { | |
| "epoch": 4.254180602006689, | |
| "grad_norm": 0.9785431623458862, | |
| "learning_rate": 0.00018607341490545052, | |
| "loss": 3.8822, | |
| "step": 319 | |
| }, | |
| { | |
| "epoch": 4.2675585284280935, | |
| "grad_norm": 1.077799677848816, | |
| "learning_rate": 0.00018602892102335928, | |
| "loss": 3.8299, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 4.280936454849498, | |
| "grad_norm": 0.8109619617462158, | |
| "learning_rate": 0.0001859844271412681, | |
| "loss": 3.8096, | |
| "step": 321 | |
| }, | |
| { | |
| "epoch": 4.294314381270903, | |
| "grad_norm": 0.967856764793396, | |
| "learning_rate": 0.00018593993325917687, | |
| "loss": 3.8639, | |
| "step": 322 | |
| }, | |
| { | |
| "epoch": 4.3076923076923075, | |
| "grad_norm": 0.8657905459403992, | |
| "learning_rate": 0.00018589543937708568, | |
| "loss": 3.7556, | |
| "step": 323 | |
| }, | |
| { | |
| "epoch": 4.321070234113712, | |
| "grad_norm": 0.9641517400741577, | |
| "learning_rate": 0.00018585094549499444, | |
| "loss": 3.9702, | |
| "step": 324 | |
| }, | |
| { | |
| "epoch": 4.334448160535117, | |
| "grad_norm": 0.9664435982704163, | |
| "learning_rate": 0.00018580645161290325, | |
| "loss": 3.8754, | |
| "step": 325 | |
| }, | |
| { | |
| "epoch": 4.3478260869565215, | |
| "grad_norm": 0.8322617411613464, | |
| "learning_rate": 0.00018576195773081203, | |
| "loss": 3.83, | |
| "step": 326 | |
| }, | |
| { | |
| "epoch": 4.361204013377926, | |
| "grad_norm": 1.0363450050354004, | |
| "learning_rate": 0.0001857174638487208, | |
| "loss": 3.9825, | |
| "step": 327 | |
| }, | |
| { | |
| "epoch": 4.374581939799331, | |
| "grad_norm": 1.0125840902328491, | |
| "learning_rate": 0.0001856729699666296, | |
| "loss": 3.6525, | |
| "step": 328 | |
| }, | |
| { | |
| "epoch": 4.3879598662207355, | |
| "grad_norm": 0.9922601580619812, | |
| "learning_rate": 0.0001856284760845384, | |
| "loss": 4.2373, | |
| "step": 329 | |
| }, | |
| { | |
| "epoch": 4.40133779264214, | |
| "grad_norm": 0.9070426225662231, | |
| "learning_rate": 0.00018558398220244716, | |
| "loss": 3.9623, | |
| "step": 330 | |
| }, | |
| { | |
| "epoch": 4.414715719063545, | |
| "grad_norm": 0.9369637370109558, | |
| "learning_rate": 0.00018553948832035597, | |
| "loss": 3.9297, | |
| "step": 331 | |
| }, | |
| { | |
| "epoch": 4.4280936454849495, | |
| "grad_norm": 1.108876347541809, | |
| "learning_rate": 0.00018549499443826475, | |
| "loss": 3.7325, | |
| "step": 332 | |
| }, | |
| { | |
| "epoch": 4.441471571906354, | |
| "grad_norm": 0.9405660629272461, | |
| "learning_rate": 0.00018545050055617356, | |
| "loss": 3.8615, | |
| "step": 333 | |
| }, | |
| { | |
| "epoch": 4.454849498327759, | |
| "grad_norm": 0.9730128645896912, | |
| "learning_rate": 0.00018540600667408231, | |
| "loss": 4.1794, | |
| "step": 334 | |
| }, | |
| { | |
| "epoch": 4.468227424749164, | |
| "grad_norm": 0.9341335892677307, | |
| "learning_rate": 0.00018536151279199112, | |
| "loss": 3.9422, | |
| "step": 335 | |
| }, | |
| { | |
| "epoch": 4.481605351170568, | |
| "grad_norm": 0.9262625575065613, | |
| "learning_rate": 0.0001853170189098999, | |
| "loss": 3.9819, | |
| "step": 336 | |
| }, | |
| { | |
| "epoch": 4.494983277591973, | |
| "grad_norm": 1.0419141054153442, | |
| "learning_rate": 0.00018527252502780866, | |
| "loss": 3.7481, | |
| "step": 337 | |
| }, | |
| { | |
| "epoch": 4.508361204013378, | |
| "grad_norm": 0.8986826539039612, | |
| "learning_rate": 0.00018522803114571747, | |
| "loss": 4.1195, | |
| "step": 338 | |
| }, | |
| { | |
| "epoch": 4.521739130434782, | |
| "grad_norm": 0.9502431154251099, | |
| "learning_rate": 0.00018518353726362625, | |
| "loss": 3.8521, | |
| "step": 339 | |
| }, | |
| { | |
| "epoch": 4.535117056856187, | |
| "grad_norm": 0.8936267495155334, | |
| "learning_rate": 0.00018513904338153504, | |
| "loss": 3.647, | |
| "step": 340 | |
| }, | |
| { | |
| "epoch": 4.548494983277592, | |
| "grad_norm": 0.8870158195495605, | |
| "learning_rate": 0.00018509454949944382, | |
| "loss": 3.7624, | |
| "step": 341 | |
| }, | |
| { | |
| "epoch": 4.561872909698996, | |
| "grad_norm": 0.9030978679656982, | |
| "learning_rate": 0.00018505005561735263, | |
| "loss": 3.8018, | |
| "step": 342 | |
| }, | |
| { | |
| "epoch": 4.575250836120401, | |
| "grad_norm": 0.8690946698188782, | |
| "learning_rate": 0.0001850055617352614, | |
| "loss": 3.6183, | |
| "step": 343 | |
| }, | |
| { | |
| "epoch": 4.588628762541806, | |
| "grad_norm": 0.9812071323394775, | |
| "learning_rate": 0.0001849610678531702, | |
| "loss": 4.1231, | |
| "step": 344 | |
| }, | |
| { | |
| "epoch": 4.602006688963211, | |
| "grad_norm": 0.9404383301734924, | |
| "learning_rate": 0.00018491657397107897, | |
| "loss": 3.9645, | |
| "step": 345 | |
| }, | |
| { | |
| "epoch": 4.615384615384615, | |
| "grad_norm": 1.0422123670578003, | |
| "learning_rate": 0.00018487208008898778, | |
| "loss": 3.9031, | |
| "step": 346 | |
| }, | |
| { | |
| "epoch": 4.6287625418060205, | |
| "grad_norm": 0.9838129281997681, | |
| "learning_rate": 0.00018482758620689654, | |
| "loss": 3.9985, | |
| "step": 347 | |
| }, | |
| { | |
| "epoch": 4.642140468227424, | |
| "grad_norm": 0.9232532978057861, | |
| "learning_rate": 0.00018478309232480535, | |
| "loss": 4.0343, | |
| "step": 348 | |
| }, | |
| { | |
| "epoch": 4.65551839464883, | |
| "grad_norm": 0.9242956042289734, | |
| "learning_rate": 0.00018473859844271413, | |
| "loss": 4.0669, | |
| "step": 349 | |
| }, | |
| { | |
| "epoch": 4.668896321070234, | |
| "grad_norm": 0.919269859790802, | |
| "learning_rate": 0.0001846941045606229, | |
| "loss": 4.0549, | |
| "step": 350 | |
| }, | |
| { | |
| "epoch": 4.682274247491639, | |
| "grad_norm": 0.93565833568573, | |
| "learning_rate": 0.0001846496106785317, | |
| "loss": 4.1306, | |
| "step": 351 | |
| }, | |
| { | |
| "epoch": 4.695652173913043, | |
| "grad_norm": 0.9001899361610413, | |
| "learning_rate": 0.0001846051167964405, | |
| "loss": 3.8916, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 4.709030100334449, | |
| "grad_norm": 0.8896821737289429, | |
| "learning_rate": 0.0001845606229143493, | |
| "loss": 3.8377, | |
| "step": 353 | |
| }, | |
| { | |
| "epoch": 4.722408026755852, | |
| "grad_norm": 1.0137807130813599, | |
| "learning_rate": 0.00018451612903225807, | |
| "loss": 3.9923, | |
| "step": 354 | |
| }, | |
| { | |
| "epoch": 4.735785953177258, | |
| "grad_norm": 1.075823426246643, | |
| "learning_rate": 0.00018447163515016685, | |
| "loss": 4.0706, | |
| "step": 355 | |
| }, | |
| { | |
| "epoch": 4.749163879598662, | |
| "grad_norm": 1.0076895952224731, | |
| "learning_rate": 0.00018442714126807566, | |
| "loss": 4.0759, | |
| "step": 356 | |
| }, | |
| { | |
| "epoch": 4.762541806020067, | |
| "grad_norm": 0.9387428164482117, | |
| "learning_rate": 0.00018438264738598442, | |
| "loss": 3.6959, | |
| "step": 357 | |
| }, | |
| { | |
| "epoch": 4.775919732441472, | |
| "grad_norm": 0.8920648097991943, | |
| "learning_rate": 0.00018433815350389323, | |
| "loss": 3.9213, | |
| "step": 358 | |
| }, | |
| { | |
| "epoch": 4.789297658862877, | |
| "grad_norm": 1.0252491235733032, | |
| "learning_rate": 0.000184293659621802, | |
| "loss": 3.9118, | |
| "step": 359 | |
| }, | |
| { | |
| "epoch": 4.802675585284281, | |
| "grad_norm": 1.0382707118988037, | |
| "learning_rate": 0.0001842491657397108, | |
| "loss": 4.0172, | |
| "step": 360 | |
| }, | |
| { | |
| "epoch": 4.816053511705686, | |
| "grad_norm": 1.07838773727417, | |
| "learning_rate": 0.00018420467185761957, | |
| "loss": 3.8531, | |
| "step": 361 | |
| }, | |
| { | |
| "epoch": 4.829431438127091, | |
| "grad_norm": 0.9974546432495117, | |
| "learning_rate": 0.00018416017797552838, | |
| "loss": 4.0387, | |
| "step": 362 | |
| }, | |
| { | |
| "epoch": 4.842809364548495, | |
| "grad_norm": 1.024491548538208, | |
| "learning_rate": 0.00018411568409343717, | |
| "loss": 3.9504, | |
| "step": 363 | |
| }, | |
| { | |
| "epoch": 4.8561872909699, | |
| "grad_norm": 0.9236369132995605, | |
| "learning_rate": 0.00018407119021134595, | |
| "loss": 3.7119, | |
| "step": 364 | |
| }, | |
| { | |
| "epoch": 4.869565217391305, | |
| "grad_norm": 0.935644268989563, | |
| "learning_rate": 0.00018402669632925473, | |
| "loss": 4.0077, | |
| "step": 365 | |
| }, | |
| { | |
| "epoch": 4.882943143812709, | |
| "grad_norm": 0.9328681230545044, | |
| "learning_rate": 0.00018398220244716354, | |
| "loss": 3.9133, | |
| "step": 366 | |
| }, | |
| { | |
| "epoch": 4.896321070234114, | |
| "grad_norm": 0.9596607089042664, | |
| "learning_rate": 0.0001839377085650723, | |
| "loss": 3.8003, | |
| "step": 367 | |
| }, | |
| { | |
| "epoch": 4.909698996655519, | |
| "grad_norm": 0.9878052473068237, | |
| "learning_rate": 0.0001838932146829811, | |
| "loss": 3.8805, | |
| "step": 368 | |
| }, | |
| { | |
| "epoch": 4.923076923076923, | |
| "grad_norm": 1.00381600856781, | |
| "learning_rate": 0.0001838487208008899, | |
| "loss": 4.0264, | |
| "step": 369 | |
| }, | |
| { | |
| "epoch": 4.936454849498328, | |
| "grad_norm": 1.024754524230957, | |
| "learning_rate": 0.00018380422691879867, | |
| "loss": 3.7291, | |
| "step": 370 | |
| }, | |
| { | |
| "epoch": 4.949832775919733, | |
| "grad_norm": 0.9670823812484741, | |
| "learning_rate": 0.00018375973303670745, | |
| "loss": 3.9418, | |
| "step": 371 | |
| }, | |
| { | |
| "epoch": 4.963210702341137, | |
| "grad_norm": 0.9736581444740295, | |
| "learning_rate": 0.00018371523915461626, | |
| "loss": 3.8813, | |
| "step": 372 | |
| }, | |
| { | |
| "epoch": 4.976588628762542, | |
| "grad_norm": 0.9752672910690308, | |
| "learning_rate": 0.00018367074527252504, | |
| "loss": 3.6717, | |
| "step": 373 | |
| }, | |
| { | |
| "epoch": 4.989966555183947, | |
| "grad_norm": 1.1268304586410522, | |
| "learning_rate": 0.00018362625139043383, | |
| "loss": 3.9782, | |
| "step": 374 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "grad_norm": 1.7933701276779175, | |
| "learning_rate": 0.0001835817575083426, | |
| "loss": 3.001, | |
| "step": 375 | |
| }, | |
| { | |
| "epoch": 5.013377926421405, | |
| "grad_norm": 0.8035010099411011, | |
| "learning_rate": 0.00018353726362625142, | |
| "loss": 3.7943, | |
| "step": 376 | |
| }, | |
| { | |
| "epoch": 5.026755852842809, | |
| "grad_norm": 0.8016420006752014, | |
| "learning_rate": 0.00018349276974416017, | |
| "loss": 3.7454, | |
| "step": 377 | |
| }, | |
| { | |
| "epoch": 5.040133779264214, | |
| "grad_norm": 0.6844643354415894, | |
| "learning_rate": 0.00018344827586206896, | |
| "loss": 3.699, | |
| "step": 378 | |
| }, | |
| { | |
| "epoch": 5.053511705685619, | |
| "grad_norm": 0.8649943470954895, | |
| "learning_rate": 0.00018340378197997777, | |
| "loss": 3.7197, | |
| "step": 379 | |
| }, | |
| { | |
| "epoch": 5.066889632107023, | |
| "grad_norm": 0.9685015678405762, | |
| "learning_rate": 0.00018335928809788655, | |
| "loss": 3.6952, | |
| "step": 380 | |
| }, | |
| { | |
| "epoch": 5.080267558528428, | |
| "grad_norm": 0.8728330135345459, | |
| "learning_rate": 0.00018331479421579533, | |
| "loss": 3.7164, | |
| "step": 381 | |
| }, | |
| { | |
| "epoch": 5.093645484949833, | |
| "grad_norm": 0.962504506111145, | |
| "learning_rate": 0.0001832703003337041, | |
| "loss": 3.6123, | |
| "step": 382 | |
| }, | |
| { | |
| "epoch": 5.107023411371237, | |
| "grad_norm": 0.9194462895393372, | |
| "learning_rate": 0.00018322580645161292, | |
| "loss": 3.275, | |
| "step": 383 | |
| }, | |
| { | |
| "epoch": 5.120401337792642, | |
| "grad_norm": 0.9851329326629639, | |
| "learning_rate": 0.00018318131256952168, | |
| "loss": 3.6222, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 5.133779264214047, | |
| "grad_norm": 1.0702580213546753, | |
| "learning_rate": 0.0001831368186874305, | |
| "loss": 3.8728, | |
| "step": 385 | |
| }, | |
| { | |
| "epoch": 5.147157190635451, | |
| "grad_norm": 1.3237228393554688, | |
| "learning_rate": 0.00018309232480533927, | |
| "loss": 3.8948, | |
| "step": 386 | |
| }, | |
| { | |
| "epoch": 5.160535117056856, | |
| "grad_norm": 1.0076218843460083, | |
| "learning_rate": 0.00018304783092324805, | |
| "loss": 3.8894, | |
| "step": 387 | |
| }, | |
| { | |
| "epoch": 5.173913043478261, | |
| "grad_norm": 1.084722876548767, | |
| "learning_rate": 0.00018300333704115683, | |
| "loss": 3.7398, | |
| "step": 388 | |
| }, | |
| { | |
| "epoch": 5.187290969899665, | |
| "grad_norm": 0.9112711548805237, | |
| "learning_rate": 0.00018295884315906564, | |
| "loss": 3.5901, | |
| "step": 389 | |
| }, | |
| { | |
| "epoch": 5.20066889632107, | |
| "grad_norm": 0.9451406002044678, | |
| "learning_rate": 0.00018291434927697443, | |
| "loss": 3.6313, | |
| "step": 390 | |
| }, | |
| { | |
| "epoch": 5.214046822742475, | |
| "grad_norm": 0.8901047706604004, | |
| "learning_rate": 0.0001828698553948832, | |
| "loss": 3.3191, | |
| "step": 391 | |
| }, | |
| { | |
| "epoch": 5.2274247491638794, | |
| "grad_norm": 0.9838565587997437, | |
| "learning_rate": 0.000182825361512792, | |
| "loss": 3.882, | |
| "step": 392 | |
| }, | |
| { | |
| "epoch": 5.240802675585284, | |
| "grad_norm": 0.9839156866073608, | |
| "learning_rate": 0.0001827808676307008, | |
| "loss": 3.6068, | |
| "step": 393 | |
| }, | |
| { | |
| "epoch": 5.254180602006689, | |
| "grad_norm": 0.9328583478927612, | |
| "learning_rate": 0.00018273637374860956, | |
| "loss": 3.6856, | |
| "step": 394 | |
| }, | |
| { | |
| "epoch": 5.2675585284280935, | |
| "grad_norm": 0.8705796003341675, | |
| "learning_rate": 0.00018269187986651837, | |
| "loss": 3.7282, | |
| "step": 395 | |
| }, | |
| { | |
| "epoch": 5.280936454849498, | |
| "grad_norm": 0.9675374031066895, | |
| "learning_rate": 0.00018264738598442715, | |
| "loss": 3.6588, | |
| "step": 396 | |
| }, | |
| { | |
| "epoch": 5.294314381270903, | |
| "grad_norm": 1.145280361175537, | |
| "learning_rate": 0.00018260289210233593, | |
| "loss": 3.8843, | |
| "step": 397 | |
| }, | |
| { | |
| "epoch": 5.3076923076923075, | |
| "grad_norm": 0.9769694805145264, | |
| "learning_rate": 0.0001825583982202447, | |
| "loss": 3.7207, | |
| "step": 398 | |
| }, | |
| { | |
| "epoch": 5.321070234113712, | |
| "grad_norm": 0.9277816414833069, | |
| "learning_rate": 0.00018251390433815352, | |
| "loss": 3.712, | |
| "step": 399 | |
| }, | |
| { | |
| "epoch": 5.334448160535117, | |
| "grad_norm": 1.1015180349349976, | |
| "learning_rate": 0.0001824694104560623, | |
| "loss": 3.7941, | |
| "step": 400 | |
| }, | |
| { | |
| "epoch": 5.3478260869565215, | |
| "grad_norm": 1.2234200239181519, | |
| "learning_rate": 0.0001824249165739711, | |
| "loss": 3.6559, | |
| "step": 401 | |
| }, | |
| { | |
| "epoch": 5.361204013377926, | |
| "grad_norm": 0.9358471035957336, | |
| "learning_rate": 0.00018238042269187987, | |
| "loss": 3.7665, | |
| "step": 402 | |
| }, | |
| { | |
| "epoch": 5.374581939799331, | |
| "grad_norm": 0.8287034630775452, | |
| "learning_rate": 0.00018233592880978868, | |
| "loss": 3.8265, | |
| "step": 403 | |
| }, | |
| { | |
| "epoch": 5.3879598662207355, | |
| "grad_norm": 1.0219204425811768, | |
| "learning_rate": 0.00018229143492769743, | |
| "loss": 3.6829, | |
| "step": 404 | |
| }, | |
| { | |
| "epoch": 5.40133779264214, | |
| "grad_norm": 1.0601041316986084, | |
| "learning_rate": 0.00018224694104560624, | |
| "loss": 3.5879, | |
| "step": 405 | |
| }, | |
| { | |
| "epoch": 5.414715719063545, | |
| "grad_norm": 1.2221566438674927, | |
| "learning_rate": 0.00018220244716351503, | |
| "loss": 3.6023, | |
| "step": 406 | |
| }, | |
| { | |
| "epoch": 5.4280936454849495, | |
| "grad_norm": 0.9589087963104248, | |
| "learning_rate": 0.0001821579532814238, | |
| "loss": 3.9109, | |
| "step": 407 | |
| }, | |
| { | |
| "epoch": 5.441471571906354, | |
| "grad_norm": 1.088295340538025, | |
| "learning_rate": 0.0001821134593993326, | |
| "loss": 3.5695, | |
| "step": 408 | |
| }, | |
| { | |
| "epoch": 5.454849498327759, | |
| "grad_norm": 1.1284915208816528, | |
| "learning_rate": 0.0001820689655172414, | |
| "loss": 3.6004, | |
| "step": 409 | |
| }, | |
| { | |
| "epoch": 5.468227424749164, | |
| "grad_norm": 1.0108689069747925, | |
| "learning_rate": 0.00018202447163515018, | |
| "loss": 3.7797, | |
| "step": 410 | |
| }, | |
| { | |
| "epoch": 5.481605351170568, | |
| "grad_norm": 0.8550918102264404, | |
| "learning_rate": 0.00018197997775305896, | |
| "loss": 3.5758, | |
| "step": 411 | |
| }, | |
| { | |
| "epoch": 5.494983277591973, | |
| "grad_norm": 0.8765145540237427, | |
| "learning_rate": 0.00018193548387096775, | |
| "loss": 3.7625, | |
| "step": 412 | |
| }, | |
| { | |
| "epoch": 5.508361204013378, | |
| "grad_norm": 1.0253541469573975, | |
| "learning_rate": 0.00018189098998887656, | |
| "loss": 3.6033, | |
| "step": 413 | |
| }, | |
| { | |
| "epoch": 5.521739130434782, | |
| "grad_norm": 1.0475622415542603, | |
| "learning_rate": 0.0001818464961067853, | |
| "loss": 3.813, | |
| "step": 414 | |
| }, | |
| { | |
| "epoch": 5.535117056856187, | |
| "grad_norm": 1.053133249282837, | |
| "learning_rate": 0.00018180200222469412, | |
| "loss": 3.4779, | |
| "step": 415 | |
| }, | |
| { | |
| "epoch": 5.548494983277592, | |
| "grad_norm": 1.0151216983795166, | |
| "learning_rate": 0.0001817575083426029, | |
| "loss": 3.9038, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 5.561872909698996, | |
| "grad_norm": 1.4666434526443481, | |
| "learning_rate": 0.00018171301446051169, | |
| "loss": 3.4735, | |
| "step": 417 | |
| }, | |
| { | |
| "epoch": 5.575250836120401, | |
| "grad_norm": 1.1043344736099243, | |
| "learning_rate": 0.00018166852057842047, | |
| "loss": 3.7449, | |
| "step": 418 | |
| }, | |
| { | |
| "epoch": 5.588628762541806, | |
| "grad_norm": 0.900745153427124, | |
| "learning_rate": 0.00018162402669632925, | |
| "loss": 3.7401, | |
| "step": 419 | |
| }, | |
| { | |
| "epoch": 5.602006688963211, | |
| "grad_norm": 0.9771101474761963, | |
| "learning_rate": 0.00018157953281423806, | |
| "loss": 3.8328, | |
| "step": 420 | |
| }, | |
| { | |
| "epoch": 5.615384615384615, | |
| "grad_norm": 0.9099516272544861, | |
| "learning_rate": 0.00018153503893214682, | |
| "loss": 3.6245, | |
| "step": 421 | |
| }, | |
| { | |
| "epoch": 5.6287625418060205, | |
| "grad_norm": 0.9844585657119751, | |
| "learning_rate": 0.00018149054505005563, | |
| "loss": 3.5776, | |
| "step": 422 | |
| }, | |
| { | |
| "epoch": 5.642140468227424, | |
| "grad_norm": 1.0481154918670654, | |
| "learning_rate": 0.0001814460511679644, | |
| "loss": 3.6304, | |
| "step": 423 | |
| }, | |
| { | |
| "epoch": 5.65551839464883, | |
| "grad_norm": 0.9971081614494324, | |
| "learning_rate": 0.0001814015572858732, | |
| "loss": 3.7863, | |
| "step": 424 | |
| }, | |
| { | |
| "epoch": 5.668896321070234, | |
| "grad_norm": 0.9247872829437256, | |
| "learning_rate": 0.00018135706340378197, | |
| "loss": 3.7319, | |
| "step": 425 | |
| }, | |
| { | |
| "epoch": 5.682274247491639, | |
| "grad_norm": 0.9895725846290588, | |
| "learning_rate": 0.00018131256952169078, | |
| "loss": 3.78, | |
| "step": 426 | |
| }, | |
| { | |
| "epoch": 5.695652173913043, | |
| "grad_norm": 1.0847641229629517, | |
| "learning_rate": 0.00018126807563959956, | |
| "loss": 3.8662, | |
| "step": 427 | |
| }, | |
| { | |
| "epoch": 5.709030100334449, | |
| "grad_norm": 0.986259937286377, | |
| "learning_rate": 0.00018122358175750835, | |
| "loss": 3.5621, | |
| "step": 428 | |
| }, | |
| { | |
| "epoch": 5.722408026755852, | |
| "grad_norm": 0.9166681170463562, | |
| "learning_rate": 0.00018117908787541713, | |
| "loss": 3.6153, | |
| "step": 429 | |
| }, | |
| { | |
| "epoch": 5.735785953177258, | |
| "grad_norm": 1.1331177949905396, | |
| "learning_rate": 0.00018113459399332594, | |
| "loss": 3.5976, | |
| "step": 430 | |
| }, | |
| { | |
| "epoch": 5.749163879598662, | |
| "grad_norm": 0.8743540644645691, | |
| "learning_rate": 0.0001810901001112347, | |
| "loss": 3.2511, | |
| "step": 431 | |
| }, | |
| { | |
| "epoch": 5.762541806020067, | |
| "grad_norm": 1.0700207948684692, | |
| "learning_rate": 0.0001810456062291435, | |
| "loss": 3.7634, | |
| "step": 432 | |
| }, | |
| { | |
| "epoch": 5.775919732441472, | |
| "grad_norm": 0.9412694573402405, | |
| "learning_rate": 0.00018100111234705229, | |
| "loss": 3.6264, | |
| "step": 433 | |
| }, | |
| { | |
| "epoch": 5.789297658862877, | |
| "grad_norm": 1.0398496389389038, | |
| "learning_rate": 0.00018095661846496107, | |
| "loss": 3.744, | |
| "step": 434 | |
| }, | |
| { | |
| "epoch": 5.802675585284281, | |
| "grad_norm": 0.9605004787445068, | |
| "learning_rate": 0.00018091212458286985, | |
| "loss": 3.5532, | |
| "step": 435 | |
| }, | |
| { | |
| "epoch": 5.816053511705686, | |
| "grad_norm": 1.0449095964431763, | |
| "learning_rate": 0.00018086763070077866, | |
| "loss": 4.0611, | |
| "step": 436 | |
| }, | |
| { | |
| "epoch": 5.829431438127091, | |
| "grad_norm": 0.9342606663703918, | |
| "learning_rate": 0.00018082313681868744, | |
| "loss": 3.9957, | |
| "step": 437 | |
| }, | |
| { | |
| "epoch": 5.842809364548495, | |
| "grad_norm": 0.9687880873680115, | |
| "learning_rate": 0.00018077864293659622, | |
| "loss": 3.9299, | |
| "step": 438 | |
| }, | |
| { | |
| "epoch": 5.8561872909699, | |
| "grad_norm": 1.1390576362609863, | |
| "learning_rate": 0.000180734149054505, | |
| "loss": 3.6552, | |
| "step": 439 | |
| }, | |
| { | |
| "epoch": 5.869565217391305, | |
| "grad_norm": 0.9280988574028015, | |
| "learning_rate": 0.00018068965517241382, | |
| "loss": 3.7828, | |
| "step": 440 | |
| }, | |
| { | |
| "epoch": 5.882943143812709, | |
| "grad_norm": 1.2928193807601929, | |
| "learning_rate": 0.00018064516129032257, | |
| "loss": 3.6292, | |
| "step": 441 | |
| }, | |
| { | |
| "epoch": 5.896321070234114, | |
| "grad_norm": 1.0959875583648682, | |
| "learning_rate": 0.00018060066740823138, | |
| "loss": 3.4293, | |
| "step": 442 | |
| }, | |
| { | |
| "epoch": 5.909698996655519, | |
| "grad_norm": 1.0713289976119995, | |
| "learning_rate": 0.00018055617352614016, | |
| "loss": 3.7767, | |
| "step": 443 | |
| }, | |
| { | |
| "epoch": 5.923076923076923, | |
| "grad_norm": 0.9309440851211548, | |
| "learning_rate": 0.00018051167964404895, | |
| "loss": 3.5473, | |
| "step": 444 | |
| }, | |
| { | |
| "epoch": 5.936454849498328, | |
| "grad_norm": 1.0999056100845337, | |
| "learning_rate": 0.00018046718576195773, | |
| "loss": 3.9694, | |
| "step": 445 | |
| }, | |
| { | |
| "epoch": 5.949832775919733, | |
| "grad_norm": 1.1073781251907349, | |
| "learning_rate": 0.00018042269187986654, | |
| "loss": 3.6882, | |
| "step": 446 | |
| }, | |
| { | |
| "epoch": 5.963210702341137, | |
| "grad_norm": 1.0430257320404053, | |
| "learning_rate": 0.00018037819799777532, | |
| "loss": 3.4009, | |
| "step": 447 | |
| }, | |
| { | |
| "epoch": 5.976588628762542, | |
| "grad_norm": 1.1132690906524658, | |
| "learning_rate": 0.0001803337041156841, | |
| "loss": 3.8832, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 5.989966555183947, | |
| "grad_norm": 1.0147771835327148, | |
| "learning_rate": 0.00018028921023359289, | |
| "loss": 3.7117, | |
| "step": 449 | |
| }, | |
| { | |
| "epoch": 6.0, | |
| "grad_norm": 1.458959698677063, | |
| "learning_rate": 0.0001802447163515017, | |
| "loss": 3.5745, | |
| "step": 450 | |
| }, | |
| { | |
| "epoch": 6.013377926421405, | |
| "grad_norm": 0.8363592028617859, | |
| "learning_rate": 0.00018020022246941045, | |
| "loss": 3.5835, | |
| "step": 451 | |
| }, | |
| { | |
| "epoch": 6.026755852842809, | |
| "grad_norm": 0.8071937561035156, | |
| "learning_rate": 0.00018015572858731926, | |
| "loss": 3.5923, | |
| "step": 452 | |
| }, | |
| { | |
| "epoch": 6.040133779264214, | |
| "grad_norm": 0.7746313214302063, | |
| "learning_rate": 0.00018011123470522804, | |
| "loss": 3.5688, | |
| "step": 453 | |
| }, | |
| { | |
| "epoch": 6.053511705685619, | |
| "grad_norm": 0.689179539680481, | |
| "learning_rate": 0.00018006674082313682, | |
| "loss": 3.412, | |
| "step": 454 | |
| }, | |
| { | |
| "epoch": 6.066889632107023, | |
| "grad_norm": 0.8438050746917725, | |
| "learning_rate": 0.0001800222469410456, | |
| "loss": 3.403, | |
| "step": 455 | |
| }, | |
| { | |
| "epoch": 6.080267558528428, | |
| "grad_norm": 0.7670062780380249, | |
| "learning_rate": 0.00017997775305895442, | |
| "loss": 3.5029, | |
| "step": 456 | |
| }, | |
| { | |
| "epoch": 6.093645484949833, | |
| "grad_norm": 0.8185870051383972, | |
| "learning_rate": 0.0001799332591768632, | |
| "loss": 3.4584, | |
| "step": 457 | |
| }, | |
| { | |
| "epoch": 6.107023411371237, | |
| "grad_norm": 0.9618543386459351, | |
| "learning_rate": 0.00017988876529477198, | |
| "loss": 3.6538, | |
| "step": 458 | |
| }, | |
| { | |
| "epoch": 6.120401337792642, | |
| "grad_norm": 0.959724485874176, | |
| "learning_rate": 0.00017984427141268076, | |
| "loss": 3.5284, | |
| "step": 459 | |
| }, | |
| { | |
| "epoch": 6.133779264214047, | |
| "grad_norm": 0.8044765591621399, | |
| "learning_rate": 0.00017979977753058955, | |
| "loss": 3.2198, | |
| "step": 460 | |
| }, | |
| { | |
| "epoch": 6.147157190635451, | |
| "grad_norm": 0.8287092447280884, | |
| "learning_rate": 0.00017975528364849833, | |
| "loss": 3.4977, | |
| "step": 461 | |
| }, | |
| { | |
| "epoch": 6.160535117056856, | |
| "grad_norm": 0.8855329155921936, | |
| "learning_rate": 0.0001797107897664071, | |
| "loss": 3.5008, | |
| "step": 462 | |
| }, | |
| { | |
| "epoch": 6.173913043478261, | |
| "grad_norm": 0.8839483857154846, | |
| "learning_rate": 0.00017966629588431592, | |
| "loss": 3.6135, | |
| "step": 463 | |
| }, | |
| { | |
| "epoch": 6.187290969899665, | |
| "grad_norm": 0.963446319103241, | |
| "learning_rate": 0.0001796218020022247, | |
| "loss": 3.6156, | |
| "step": 464 | |
| }, | |
| { | |
| "epoch": 6.20066889632107, | |
| "grad_norm": 0.896743655204773, | |
| "learning_rate": 0.00017957730812013348, | |
| "loss": 3.6623, | |
| "step": 465 | |
| }, | |
| { | |
| "epoch": 6.214046822742475, | |
| "grad_norm": 0.9268617033958435, | |
| "learning_rate": 0.00017953281423804227, | |
| "loss": 3.4343, | |
| "step": 466 | |
| }, | |
| { | |
| "epoch": 6.2274247491638794, | |
| "grad_norm": 0.8335449695587158, | |
| "learning_rate": 0.00017948832035595108, | |
| "loss": 3.5716, | |
| "step": 467 | |
| }, | |
| { | |
| "epoch": 6.240802675585284, | |
| "grad_norm": 0.7771849036216736, | |
| "learning_rate": 0.00017944382647385983, | |
| "loss": 3.5191, | |
| "step": 468 | |
| }, | |
| { | |
| "epoch": 6.254180602006689, | |
| "grad_norm": 0.9157488346099854, | |
| "learning_rate": 0.00017939933259176864, | |
| "loss": 3.5583, | |
| "step": 469 | |
| }, | |
| { | |
| "epoch": 6.2675585284280935, | |
| "grad_norm": 0.9348477721214294, | |
| "learning_rate": 0.00017935483870967742, | |
| "loss": 3.3137, | |
| "step": 470 | |
| }, | |
| { | |
| "epoch": 6.280936454849498, | |
| "grad_norm": 0.8791135549545288, | |
| "learning_rate": 0.0001793103448275862, | |
| "loss": 3.5111, | |
| "step": 471 | |
| }, | |
| { | |
| "epoch": 6.294314381270903, | |
| "grad_norm": 0.9963672757148743, | |
| "learning_rate": 0.000179265850945495, | |
| "loss": 3.7518, | |
| "step": 472 | |
| }, | |
| { | |
| "epoch": 6.3076923076923075, | |
| "grad_norm": 0.9291539192199707, | |
| "learning_rate": 0.0001792213570634038, | |
| "loss": 3.4524, | |
| "step": 473 | |
| }, | |
| { | |
| "epoch": 6.321070234113712, | |
| "grad_norm": 0.9349279403686523, | |
| "learning_rate": 0.00017917686318131258, | |
| "loss": 3.4753, | |
| "step": 474 | |
| }, | |
| { | |
| "epoch": 6.334448160535117, | |
| "grad_norm": 0.8984476327896118, | |
| "learning_rate": 0.00017913236929922136, | |
| "loss": 3.7325, | |
| "step": 475 | |
| }, | |
| { | |
| "epoch": 6.3478260869565215, | |
| "grad_norm": 0.8452139496803284, | |
| "learning_rate": 0.00017908787541713015, | |
| "loss": 3.8021, | |
| "step": 476 | |
| }, | |
| { | |
| "epoch": 6.361204013377926, | |
| "grad_norm": 0.9418376088142395, | |
| "learning_rate": 0.00017904338153503895, | |
| "loss": 3.7426, | |
| "step": 477 | |
| }, | |
| { | |
| "epoch": 6.374581939799331, | |
| "grad_norm": 1.0661097764968872, | |
| "learning_rate": 0.0001789988876529477, | |
| "loss": 3.7556, | |
| "step": 478 | |
| }, | |
| { | |
| "epoch": 6.3879598662207355, | |
| "grad_norm": 0.9645984768867493, | |
| "learning_rate": 0.00017895439377085652, | |
| "loss": 3.3353, | |
| "step": 479 | |
| }, | |
| { | |
| "epoch": 6.40133779264214, | |
| "grad_norm": 0.9243470430374146, | |
| "learning_rate": 0.0001789098998887653, | |
| "loss": 3.5729, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 6.414715719063545, | |
| "grad_norm": 0.885061502456665, | |
| "learning_rate": 0.00017886540600667408, | |
| "loss": 3.5699, | |
| "step": 481 | |
| }, | |
| { | |
| "epoch": 6.4280936454849495, | |
| "grad_norm": 0.9025402069091797, | |
| "learning_rate": 0.00017882091212458287, | |
| "loss": 3.4532, | |
| "step": 482 | |
| }, | |
| { | |
| "epoch": 6.441471571906354, | |
| "grad_norm": 0.9760842323303223, | |
| "learning_rate": 0.00017877641824249168, | |
| "loss": 3.7222, | |
| "step": 483 | |
| }, | |
| { | |
| "epoch": 6.454849498327759, | |
| "grad_norm": 1.2709609270095825, | |
| "learning_rate": 0.00017873192436040046, | |
| "loss": 3.756, | |
| "step": 484 | |
| }, | |
| { | |
| "epoch": 6.468227424749164, | |
| "grad_norm": 0.904513955116272, | |
| "learning_rate": 0.00017868743047830924, | |
| "loss": 3.657, | |
| "step": 485 | |
| }, | |
| { | |
| "epoch": 6.481605351170568, | |
| "grad_norm": 1.158915400505066, | |
| "learning_rate": 0.00017864293659621802, | |
| "loss": 3.5897, | |
| "step": 486 | |
| }, | |
| { | |
| "epoch": 6.494983277591973, | |
| "grad_norm": 0.9457879066467285, | |
| "learning_rate": 0.00017859844271412683, | |
| "loss": 3.4394, | |
| "step": 487 | |
| }, | |
| { | |
| "epoch": 6.508361204013378, | |
| "grad_norm": 1.04762601852417, | |
| "learning_rate": 0.0001785539488320356, | |
| "loss": 3.39, | |
| "step": 488 | |
| }, | |
| { | |
| "epoch": 6.521739130434782, | |
| "grad_norm": 0.9370948076248169, | |
| "learning_rate": 0.0001785094549499444, | |
| "loss": 3.4816, | |
| "step": 489 | |
| }, | |
| { | |
| "epoch": 6.535117056856187, | |
| "grad_norm": 1.1307988166809082, | |
| "learning_rate": 0.00017846496106785318, | |
| "loss": 3.5079, | |
| "step": 490 | |
| }, | |
| { | |
| "epoch": 6.548494983277592, | |
| "grad_norm": 1.0025054216384888, | |
| "learning_rate": 0.00017842046718576196, | |
| "loss": 3.3447, | |
| "step": 491 | |
| }, | |
| { | |
| "epoch": 6.561872909698996, | |
| "grad_norm": 0.9893412590026855, | |
| "learning_rate": 0.00017837597330367074, | |
| "loss": 3.4698, | |
| "step": 492 | |
| }, | |
| { | |
| "epoch": 6.575250836120401, | |
| "grad_norm": 1.0476347208023071, | |
| "learning_rate": 0.00017833147942157955, | |
| "loss": 3.4955, | |
| "step": 493 | |
| }, | |
| { | |
| "epoch": 6.588628762541806, | |
| "grad_norm": 1.0508891344070435, | |
| "learning_rate": 0.00017828698553948834, | |
| "loss": 3.6568, | |
| "step": 494 | |
| }, | |
| { | |
| "epoch": 6.602006688963211, | |
| "grad_norm": 1.0397465229034424, | |
| "learning_rate": 0.00017824249165739712, | |
| "loss": 3.5087, | |
| "step": 495 | |
| }, | |
| { | |
| "epoch": 6.615384615384615, | |
| "grad_norm": 0.9884181022644043, | |
| "learning_rate": 0.0001781979977753059, | |
| "loss": 3.6778, | |
| "step": 496 | |
| }, | |
| { | |
| "epoch": 6.6287625418060205, | |
| "grad_norm": 1.1187562942504883, | |
| "learning_rate": 0.0001781535038932147, | |
| "loss": 3.4345, | |
| "step": 497 | |
| }, | |
| { | |
| "epoch": 6.642140468227424, | |
| "grad_norm": 1.1133880615234375, | |
| "learning_rate": 0.00017810901001112347, | |
| "loss": 3.4959, | |
| "step": 498 | |
| }, | |
| { | |
| "epoch": 6.65551839464883, | |
| "grad_norm": 0.9368589520454407, | |
| "learning_rate": 0.00017806451612903228, | |
| "loss": 3.628, | |
| "step": 499 | |
| }, | |
| { | |
| "epoch": 6.668896321070234, | |
| "grad_norm": 1.0427212715148926, | |
| "learning_rate": 0.00017802002224694106, | |
| "loss": 3.6073, | |
| "step": 500 | |
| } | |
| ], | |
| "logging_steps": 1, | |
| "max_steps": 4500, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 60, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": false | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 1.54743987923712e+16, | |
| "train_batch_size": 4, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |